Add files using upload-large-folder tool
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- env-llmeval/lib/python3.10/site-packages/aiohttp-3.9.4.dist-info/INSTALLER +1 -0
- env-llmeval/lib/python3.10/site-packages/aiohttp-3.9.4.dist-info/LICENSE.txt +13 -0
- env-llmeval/lib/python3.10/site-packages/attr/__init__.py +134 -0
- env-llmeval/lib/python3.10/site-packages/attr/_cmp.py +150 -0
- env-llmeval/lib/python3.10/site-packages/attr/_funcs.py +483 -0
- env-llmeval/lib/python3.10/site-packages/attr/_make.py +0 -0
- env-llmeval/lib/python3.10/site-packages/attr/_version_info.pyi +9 -0
- env-llmeval/lib/python3.10/site-packages/attr/converters.pyi +13 -0
- env-llmeval/lib/python3.10/site-packages/attr/exceptions.py +95 -0
- env-llmeval/lib/python3.10/site-packages/attr/exceptions.pyi +17 -0
- env-llmeval/lib/python3.10/site-packages/attr/filters.pyi +6 -0
- env-llmeval/lib/python3.10/site-packages/attr/py.typed +0 -0
- env-llmeval/lib/python3.10/site-packages/attr/setters.py +73 -0
- env-llmeval/lib/python3.10/site-packages/attr/validators.pyi +88 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/__init__.py +20 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/__pycache__/__init__.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/__pycache__/_base.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/__pycache__/_numpy.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/__pycache__/_tensorflow.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/__pycache__/_torch.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/_base.py +169 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/_numpy.py +68 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/_tensorflow.py +94 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/_torch.py +200 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_cache_manager.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_experimental.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_hf_folder.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_paths.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_runtime.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_safetensors.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_subprocess.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_telemetry.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_typing.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/sha.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/tqdm.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/__about__.py +26 -0
- env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/__init__.py +25 -0
- env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/__pycache__/_structures.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/__pycache__/requirements.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/__pycache__/specifiers.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/_manylinux.py +301 -0
- env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/_structures.py +67 -0
- env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/markers.py +304 -0
- env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/requirements.py +146 -0
- env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/specifiers.py +828 -0
- env-llmeval/lib/python3.10/site-packages/sacrebleu/dataset/__main__.py +45 -0
- env-llmeval/lib/python3.10/site-packages/sacrebleu/dataset/__pycache__/__main__.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/sacrebleu/dataset/__pycache__/base.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/sacrebleu/dataset/__pycache__/iwslt_xml.cpython-310.pyc +0 -0
- env-llmeval/lib/python3.10/site-packages/sacrebleu/dataset/__pycache__/plain_text.cpython-310.pyc +0 -0
env-llmeval/lib/python3.10/site-packages/aiohttp-3.9.4.dist-info/INSTALLER
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
pip
|
env-llmeval/lib/python3.10/site-packages/aiohttp-3.9.4.dist-info/LICENSE.txt
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Copyright aio-libs contributors.
|
| 2 |
+
|
| 3 |
+
Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
you may not use this file except in compliance with the License.
|
| 5 |
+
You may obtain a copy of the License at
|
| 6 |
+
|
| 7 |
+
http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
|
| 9 |
+
Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
See the License for the specific language governing permissions and
|
| 13 |
+
limitations under the License.
|
env-llmeval/lib/python3.10/site-packages/attr/__init__.py
ADDED
|
@@ -0,0 +1,134 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# SPDX-License-Identifier: MIT
|
| 2 |
+
|
| 3 |
+
"""
|
| 4 |
+
Classes Without Boilerplate
|
| 5 |
+
"""
|
| 6 |
+
|
| 7 |
+
from functools import partial
|
| 8 |
+
from typing import Callable
|
| 9 |
+
|
| 10 |
+
from . import converters, exceptions, filters, setters, validators
|
| 11 |
+
from ._cmp import cmp_using
|
| 12 |
+
from ._compat import Protocol
|
| 13 |
+
from ._config import get_run_validators, set_run_validators
|
| 14 |
+
from ._funcs import asdict, assoc, astuple, evolve, has, resolve_types
|
| 15 |
+
from ._make import (
|
| 16 |
+
NOTHING,
|
| 17 |
+
Attribute,
|
| 18 |
+
Factory,
|
| 19 |
+
attrib,
|
| 20 |
+
attrs,
|
| 21 |
+
fields,
|
| 22 |
+
fields_dict,
|
| 23 |
+
make_class,
|
| 24 |
+
validate,
|
| 25 |
+
)
|
| 26 |
+
from ._next_gen import define, field, frozen, mutable
|
| 27 |
+
from ._version_info import VersionInfo
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
s = attributes = attrs
|
| 31 |
+
ib = attr = attrib
|
| 32 |
+
dataclass = partial(attrs, auto_attribs=True) # happy Easter ;)
|
| 33 |
+
|
| 34 |
+
|
| 35 |
+
class AttrsInstance(Protocol):
|
| 36 |
+
pass
|
| 37 |
+
|
| 38 |
+
|
| 39 |
+
__all__ = [
|
| 40 |
+
"Attribute",
|
| 41 |
+
"AttrsInstance",
|
| 42 |
+
"Factory",
|
| 43 |
+
"NOTHING",
|
| 44 |
+
"asdict",
|
| 45 |
+
"assoc",
|
| 46 |
+
"astuple",
|
| 47 |
+
"attr",
|
| 48 |
+
"attrib",
|
| 49 |
+
"attributes",
|
| 50 |
+
"attrs",
|
| 51 |
+
"cmp_using",
|
| 52 |
+
"converters",
|
| 53 |
+
"define",
|
| 54 |
+
"evolve",
|
| 55 |
+
"exceptions",
|
| 56 |
+
"field",
|
| 57 |
+
"fields",
|
| 58 |
+
"fields_dict",
|
| 59 |
+
"filters",
|
| 60 |
+
"frozen",
|
| 61 |
+
"get_run_validators",
|
| 62 |
+
"has",
|
| 63 |
+
"ib",
|
| 64 |
+
"make_class",
|
| 65 |
+
"mutable",
|
| 66 |
+
"resolve_types",
|
| 67 |
+
"s",
|
| 68 |
+
"set_run_validators",
|
| 69 |
+
"setters",
|
| 70 |
+
"validate",
|
| 71 |
+
"validators",
|
| 72 |
+
]
|
| 73 |
+
|
| 74 |
+
|
| 75 |
+
def _make_getattr(mod_name: str) -> Callable:
|
| 76 |
+
"""
|
| 77 |
+
Create a metadata proxy for packaging information that uses *mod_name* in
|
| 78 |
+
its warnings and errors.
|
| 79 |
+
"""
|
| 80 |
+
|
| 81 |
+
def __getattr__(name: str) -> str:
|
| 82 |
+
dunder_to_metadata = {
|
| 83 |
+
"__title__": "Name",
|
| 84 |
+
"__copyright__": "",
|
| 85 |
+
"__version__": "version",
|
| 86 |
+
"__version_info__": "version",
|
| 87 |
+
"__description__": "summary",
|
| 88 |
+
"__uri__": "",
|
| 89 |
+
"__url__": "",
|
| 90 |
+
"__author__": "",
|
| 91 |
+
"__email__": "",
|
| 92 |
+
"__license__": "license",
|
| 93 |
+
}
|
| 94 |
+
if name not in dunder_to_metadata:
|
| 95 |
+
msg = f"module {mod_name} has no attribute {name}"
|
| 96 |
+
raise AttributeError(msg)
|
| 97 |
+
|
| 98 |
+
import sys
|
| 99 |
+
import warnings
|
| 100 |
+
|
| 101 |
+
if sys.version_info < (3, 8):
|
| 102 |
+
from importlib_metadata import metadata
|
| 103 |
+
else:
|
| 104 |
+
from importlib.metadata import metadata
|
| 105 |
+
|
| 106 |
+
if name not in ("__version__", "__version_info__"):
|
| 107 |
+
warnings.warn(
|
| 108 |
+
f"Accessing {mod_name}.{name} is deprecated and will be "
|
| 109 |
+
"removed in a future release. Use importlib.metadata directly "
|
| 110 |
+
"to query for attrs's packaging metadata.",
|
| 111 |
+
DeprecationWarning,
|
| 112 |
+
stacklevel=2,
|
| 113 |
+
)
|
| 114 |
+
|
| 115 |
+
meta = metadata("attrs")
|
| 116 |
+
if name == "__license__":
|
| 117 |
+
return "MIT"
|
| 118 |
+
if name == "__copyright__":
|
| 119 |
+
return "Copyright (c) 2015 Hynek Schlawack"
|
| 120 |
+
if name in ("__uri__", "__url__"):
|
| 121 |
+
return meta["Project-URL"].split(" ", 1)[-1]
|
| 122 |
+
if name == "__version_info__":
|
| 123 |
+
return VersionInfo._from_version_string(meta["version"])
|
| 124 |
+
if name == "__author__":
|
| 125 |
+
return meta["Author-email"].rsplit(" ", 1)[0]
|
| 126 |
+
if name == "__email__":
|
| 127 |
+
return meta["Author-email"].rsplit("<", 1)[1][:-1]
|
| 128 |
+
|
| 129 |
+
return meta[dunder_to_metadata[name]]
|
| 130 |
+
|
| 131 |
+
return __getattr__
|
| 132 |
+
|
| 133 |
+
|
| 134 |
+
__getattr__ = _make_getattr(__name__)
|
env-llmeval/lib/python3.10/site-packages/attr/_cmp.py
ADDED
|
@@ -0,0 +1,150 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# SPDX-License-Identifier: MIT
|
| 2 |
+
|
| 3 |
+
|
| 4 |
+
import functools
|
| 5 |
+
import types
|
| 6 |
+
|
| 7 |
+
from ._make import _make_ne
|
| 8 |
+
|
| 9 |
+
|
| 10 |
+
_operation_names = {"eq": "==", "lt": "<", "le": "<=", "gt": ">", "ge": ">="}
|
| 11 |
+
|
| 12 |
+
|
| 13 |
+
def cmp_using(
|
| 14 |
+
eq=None,
|
| 15 |
+
lt=None,
|
| 16 |
+
le=None,
|
| 17 |
+
gt=None,
|
| 18 |
+
ge=None,
|
| 19 |
+
require_same_type=True,
|
| 20 |
+
class_name="Comparable",
|
| 21 |
+
):
|
| 22 |
+
"""
|
| 23 |
+
Create a class that can be passed into `attrs.field`'s ``eq``, ``order``,
|
| 24 |
+
and ``cmp`` arguments to customize field comparison.
|
| 25 |
+
|
| 26 |
+
The resulting class will have a full set of ordering methods if at least
|
| 27 |
+
one of ``{lt, le, gt, ge}`` and ``eq`` are provided.
|
| 28 |
+
|
| 29 |
+
:param Optional[callable] eq: `callable` used to evaluate equality of two
|
| 30 |
+
objects.
|
| 31 |
+
:param Optional[callable] lt: `callable` used to evaluate whether one
|
| 32 |
+
object is less than another object.
|
| 33 |
+
:param Optional[callable] le: `callable` used to evaluate whether one
|
| 34 |
+
object is less than or equal to another object.
|
| 35 |
+
:param Optional[callable] gt: `callable` used to evaluate whether one
|
| 36 |
+
object is greater than another object.
|
| 37 |
+
:param Optional[callable] ge: `callable` used to evaluate whether one
|
| 38 |
+
object is greater than or equal to another object.
|
| 39 |
+
|
| 40 |
+
:param bool require_same_type: When `True`, equality and ordering methods
|
| 41 |
+
will return `NotImplemented` if objects are not of the same type.
|
| 42 |
+
|
| 43 |
+
:param Optional[str] class_name: Name of class. Defaults to 'Comparable'.
|
| 44 |
+
|
| 45 |
+
See `comparison` for more details.
|
| 46 |
+
|
| 47 |
+
.. versionadded:: 21.1.0
|
| 48 |
+
"""
|
| 49 |
+
|
| 50 |
+
body = {
|
| 51 |
+
"__slots__": ["value"],
|
| 52 |
+
"__init__": _make_init(),
|
| 53 |
+
"_requirements": [],
|
| 54 |
+
"_is_comparable_to": _is_comparable_to,
|
| 55 |
+
}
|
| 56 |
+
|
| 57 |
+
# Add operations.
|
| 58 |
+
num_order_functions = 0
|
| 59 |
+
has_eq_function = False
|
| 60 |
+
|
| 61 |
+
if eq is not None:
|
| 62 |
+
has_eq_function = True
|
| 63 |
+
body["__eq__"] = _make_operator("eq", eq)
|
| 64 |
+
body["__ne__"] = _make_ne()
|
| 65 |
+
|
| 66 |
+
if lt is not None:
|
| 67 |
+
num_order_functions += 1
|
| 68 |
+
body["__lt__"] = _make_operator("lt", lt)
|
| 69 |
+
|
| 70 |
+
if le is not None:
|
| 71 |
+
num_order_functions += 1
|
| 72 |
+
body["__le__"] = _make_operator("le", le)
|
| 73 |
+
|
| 74 |
+
if gt is not None:
|
| 75 |
+
num_order_functions += 1
|
| 76 |
+
body["__gt__"] = _make_operator("gt", gt)
|
| 77 |
+
|
| 78 |
+
if ge is not None:
|
| 79 |
+
num_order_functions += 1
|
| 80 |
+
body["__ge__"] = _make_operator("ge", ge)
|
| 81 |
+
|
| 82 |
+
type_ = types.new_class(
|
| 83 |
+
class_name, (object,), {}, lambda ns: ns.update(body)
|
| 84 |
+
)
|
| 85 |
+
|
| 86 |
+
# Add same type requirement.
|
| 87 |
+
if require_same_type:
|
| 88 |
+
type_._requirements.append(_check_same_type)
|
| 89 |
+
|
| 90 |
+
# Add total ordering if at least one operation was defined.
|
| 91 |
+
if 0 < num_order_functions < 4:
|
| 92 |
+
if not has_eq_function:
|
| 93 |
+
# functools.total_ordering requires __eq__ to be defined,
|
| 94 |
+
# so raise early error here to keep a nice stack.
|
| 95 |
+
msg = "eq must be define is order to complete ordering from lt, le, gt, ge."
|
| 96 |
+
raise ValueError(msg)
|
| 97 |
+
type_ = functools.total_ordering(type_)
|
| 98 |
+
|
| 99 |
+
return type_
|
| 100 |
+
|
| 101 |
+
|
| 102 |
+
def _make_init():
|
| 103 |
+
"""
|
| 104 |
+
Create __init__ method.
|
| 105 |
+
"""
|
| 106 |
+
|
| 107 |
+
def __init__(self, value):
|
| 108 |
+
"""
|
| 109 |
+
Initialize object with *value*.
|
| 110 |
+
"""
|
| 111 |
+
self.value = value
|
| 112 |
+
|
| 113 |
+
return __init__
|
| 114 |
+
|
| 115 |
+
|
| 116 |
+
def _make_operator(name, func):
|
| 117 |
+
"""
|
| 118 |
+
Create operator method.
|
| 119 |
+
"""
|
| 120 |
+
|
| 121 |
+
def method(self, other):
|
| 122 |
+
if not self._is_comparable_to(other):
|
| 123 |
+
return NotImplemented
|
| 124 |
+
|
| 125 |
+
result = func(self.value, other.value)
|
| 126 |
+
if result is NotImplemented:
|
| 127 |
+
return NotImplemented
|
| 128 |
+
|
| 129 |
+
return result
|
| 130 |
+
|
| 131 |
+
method.__name__ = f"__{name}__"
|
| 132 |
+
method.__doc__ = (
|
| 133 |
+
f"Return a {_operation_names[name]} b. Computed by attrs."
|
| 134 |
+
)
|
| 135 |
+
|
| 136 |
+
return method
|
| 137 |
+
|
| 138 |
+
|
| 139 |
+
def _is_comparable_to(self, other):
|
| 140 |
+
"""
|
| 141 |
+
Check whether `other` is comparable to `self`.
|
| 142 |
+
"""
|
| 143 |
+
return all(func(self, other) for func in self._requirements)
|
| 144 |
+
|
| 145 |
+
|
| 146 |
+
def _check_same_type(self, other):
|
| 147 |
+
"""
|
| 148 |
+
Return True if *self* and *other* are of the same type, False otherwise.
|
| 149 |
+
"""
|
| 150 |
+
return other.value.__class__ is self.value.__class__
|
env-llmeval/lib/python3.10/site-packages/attr/_funcs.py
ADDED
|
@@ -0,0 +1,483 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# SPDX-License-Identifier: MIT
|
| 2 |
+
|
| 3 |
+
|
| 4 |
+
import copy
|
| 5 |
+
|
| 6 |
+
from ._compat import PY_3_9_PLUS, get_generic_base
|
| 7 |
+
from ._make import NOTHING, _obj_setattr, fields
|
| 8 |
+
from .exceptions import AttrsAttributeNotFoundError
|
| 9 |
+
|
| 10 |
+
|
| 11 |
+
def asdict(
|
| 12 |
+
inst,
|
| 13 |
+
recurse=True,
|
| 14 |
+
filter=None,
|
| 15 |
+
dict_factory=dict,
|
| 16 |
+
retain_collection_types=False,
|
| 17 |
+
value_serializer=None,
|
| 18 |
+
):
|
| 19 |
+
"""
|
| 20 |
+
Return the *attrs* attribute values of *inst* as a dict.
|
| 21 |
+
|
| 22 |
+
Optionally recurse into other *attrs*-decorated classes.
|
| 23 |
+
|
| 24 |
+
:param inst: Instance of an *attrs*-decorated class.
|
| 25 |
+
:param bool recurse: Recurse into classes that are also
|
| 26 |
+
*attrs*-decorated.
|
| 27 |
+
:param callable filter: A callable whose return code determines whether an
|
| 28 |
+
attribute or element is included (``True``) or dropped (``False``). Is
|
| 29 |
+
called with the `attrs.Attribute` as the first argument and the
|
| 30 |
+
value as the second argument.
|
| 31 |
+
:param callable dict_factory: A callable to produce dictionaries from. For
|
| 32 |
+
example, to produce ordered dictionaries instead of normal Python
|
| 33 |
+
dictionaries, pass in ``collections.OrderedDict``.
|
| 34 |
+
:param bool retain_collection_types: Do not convert to ``list`` when
|
| 35 |
+
encountering an attribute whose type is ``tuple`` or ``set``. Only
|
| 36 |
+
meaningful if ``recurse`` is ``True``.
|
| 37 |
+
:param Optional[callable] value_serializer: A hook that is called for every
|
| 38 |
+
attribute or dict key/value. It receives the current instance, field
|
| 39 |
+
and value and must return the (updated) value. The hook is run *after*
|
| 40 |
+
the optional *filter* has been applied.
|
| 41 |
+
|
| 42 |
+
:rtype: return type of *dict_factory*
|
| 43 |
+
|
| 44 |
+
:raise attrs.exceptions.NotAnAttrsClassError: If *cls* is not an *attrs*
|
| 45 |
+
class.
|
| 46 |
+
|
| 47 |
+
.. versionadded:: 16.0.0 *dict_factory*
|
| 48 |
+
.. versionadded:: 16.1.0 *retain_collection_types*
|
| 49 |
+
.. versionadded:: 20.3.0 *value_serializer*
|
| 50 |
+
.. versionadded:: 21.3.0 If a dict has a collection for a key, it is
|
| 51 |
+
serialized as a tuple.
|
| 52 |
+
"""
|
| 53 |
+
attrs = fields(inst.__class__)
|
| 54 |
+
rv = dict_factory()
|
| 55 |
+
for a in attrs:
|
| 56 |
+
v = getattr(inst, a.name)
|
| 57 |
+
if filter is not None and not filter(a, v):
|
| 58 |
+
continue
|
| 59 |
+
|
| 60 |
+
if value_serializer is not None:
|
| 61 |
+
v = value_serializer(inst, a, v)
|
| 62 |
+
|
| 63 |
+
if recurse is True:
|
| 64 |
+
if has(v.__class__):
|
| 65 |
+
rv[a.name] = asdict(
|
| 66 |
+
v,
|
| 67 |
+
recurse=True,
|
| 68 |
+
filter=filter,
|
| 69 |
+
dict_factory=dict_factory,
|
| 70 |
+
retain_collection_types=retain_collection_types,
|
| 71 |
+
value_serializer=value_serializer,
|
| 72 |
+
)
|
| 73 |
+
elif isinstance(v, (tuple, list, set, frozenset)):
|
| 74 |
+
cf = v.__class__ if retain_collection_types is True else list
|
| 75 |
+
items = [
|
| 76 |
+
_asdict_anything(
|
| 77 |
+
i,
|
| 78 |
+
is_key=False,
|
| 79 |
+
filter=filter,
|
| 80 |
+
dict_factory=dict_factory,
|
| 81 |
+
retain_collection_types=retain_collection_types,
|
| 82 |
+
value_serializer=value_serializer,
|
| 83 |
+
)
|
| 84 |
+
for i in v
|
| 85 |
+
]
|
| 86 |
+
try:
|
| 87 |
+
rv[a.name] = cf(items)
|
| 88 |
+
except TypeError:
|
| 89 |
+
if not issubclass(cf, tuple):
|
| 90 |
+
raise
|
| 91 |
+
# Workaround for TypeError: cf.__new__() missing 1 required
|
| 92 |
+
# positional argument (which appears, for a namedturle)
|
| 93 |
+
rv[a.name] = cf(*items)
|
| 94 |
+
elif isinstance(v, dict):
|
| 95 |
+
df = dict_factory
|
| 96 |
+
rv[a.name] = df(
|
| 97 |
+
(
|
| 98 |
+
_asdict_anything(
|
| 99 |
+
kk,
|
| 100 |
+
is_key=True,
|
| 101 |
+
filter=filter,
|
| 102 |
+
dict_factory=df,
|
| 103 |
+
retain_collection_types=retain_collection_types,
|
| 104 |
+
value_serializer=value_serializer,
|
| 105 |
+
),
|
| 106 |
+
_asdict_anything(
|
| 107 |
+
vv,
|
| 108 |
+
is_key=False,
|
| 109 |
+
filter=filter,
|
| 110 |
+
dict_factory=df,
|
| 111 |
+
retain_collection_types=retain_collection_types,
|
| 112 |
+
value_serializer=value_serializer,
|
| 113 |
+
),
|
| 114 |
+
)
|
| 115 |
+
for kk, vv in v.items()
|
| 116 |
+
)
|
| 117 |
+
else:
|
| 118 |
+
rv[a.name] = v
|
| 119 |
+
else:
|
| 120 |
+
rv[a.name] = v
|
| 121 |
+
return rv
|
| 122 |
+
|
| 123 |
+
|
| 124 |
+
def _asdict_anything(
|
| 125 |
+
val,
|
| 126 |
+
is_key,
|
| 127 |
+
filter,
|
| 128 |
+
dict_factory,
|
| 129 |
+
retain_collection_types,
|
| 130 |
+
value_serializer,
|
| 131 |
+
):
|
| 132 |
+
"""
|
| 133 |
+
``asdict`` only works on attrs instances, this works on anything.
|
| 134 |
+
"""
|
| 135 |
+
if getattr(val.__class__, "__attrs_attrs__", None) is not None:
|
| 136 |
+
# Attrs class.
|
| 137 |
+
rv = asdict(
|
| 138 |
+
val,
|
| 139 |
+
recurse=True,
|
| 140 |
+
filter=filter,
|
| 141 |
+
dict_factory=dict_factory,
|
| 142 |
+
retain_collection_types=retain_collection_types,
|
| 143 |
+
value_serializer=value_serializer,
|
| 144 |
+
)
|
| 145 |
+
elif isinstance(val, (tuple, list, set, frozenset)):
|
| 146 |
+
if retain_collection_types is True:
|
| 147 |
+
cf = val.__class__
|
| 148 |
+
elif is_key:
|
| 149 |
+
cf = tuple
|
| 150 |
+
else:
|
| 151 |
+
cf = list
|
| 152 |
+
|
| 153 |
+
rv = cf(
|
| 154 |
+
[
|
| 155 |
+
_asdict_anything(
|
| 156 |
+
i,
|
| 157 |
+
is_key=False,
|
| 158 |
+
filter=filter,
|
| 159 |
+
dict_factory=dict_factory,
|
| 160 |
+
retain_collection_types=retain_collection_types,
|
| 161 |
+
value_serializer=value_serializer,
|
| 162 |
+
)
|
| 163 |
+
for i in val
|
| 164 |
+
]
|
| 165 |
+
)
|
| 166 |
+
elif isinstance(val, dict):
|
| 167 |
+
df = dict_factory
|
| 168 |
+
rv = df(
|
| 169 |
+
(
|
| 170 |
+
_asdict_anything(
|
| 171 |
+
kk,
|
| 172 |
+
is_key=True,
|
| 173 |
+
filter=filter,
|
| 174 |
+
dict_factory=df,
|
| 175 |
+
retain_collection_types=retain_collection_types,
|
| 176 |
+
value_serializer=value_serializer,
|
| 177 |
+
),
|
| 178 |
+
_asdict_anything(
|
| 179 |
+
vv,
|
| 180 |
+
is_key=False,
|
| 181 |
+
filter=filter,
|
| 182 |
+
dict_factory=df,
|
| 183 |
+
retain_collection_types=retain_collection_types,
|
| 184 |
+
value_serializer=value_serializer,
|
| 185 |
+
),
|
| 186 |
+
)
|
| 187 |
+
for kk, vv in val.items()
|
| 188 |
+
)
|
| 189 |
+
else:
|
| 190 |
+
rv = val
|
| 191 |
+
if value_serializer is not None:
|
| 192 |
+
rv = value_serializer(None, None, rv)
|
| 193 |
+
|
| 194 |
+
return rv
|
| 195 |
+
|
| 196 |
+
|
| 197 |
+
def astuple(
|
| 198 |
+
inst,
|
| 199 |
+
recurse=True,
|
| 200 |
+
filter=None,
|
| 201 |
+
tuple_factory=tuple,
|
| 202 |
+
retain_collection_types=False,
|
| 203 |
+
):
|
| 204 |
+
"""
|
| 205 |
+
Return the *attrs* attribute values of *inst* as a tuple.
|
| 206 |
+
|
| 207 |
+
Optionally recurse into other *attrs*-decorated classes.
|
| 208 |
+
|
| 209 |
+
:param inst: Instance of an *attrs*-decorated class.
|
| 210 |
+
:param bool recurse: Recurse into classes that are also
|
| 211 |
+
*attrs*-decorated.
|
| 212 |
+
:param callable filter: A callable whose return code determines whether an
|
| 213 |
+
attribute or element is included (``True``) or dropped (``False``). Is
|
| 214 |
+
called with the `attrs.Attribute` as the first argument and the
|
| 215 |
+
value as the second argument.
|
| 216 |
+
:param callable tuple_factory: A callable to produce tuples from. For
|
| 217 |
+
example, to produce lists instead of tuples.
|
| 218 |
+
:param bool retain_collection_types: Do not convert to ``list``
|
| 219 |
+
or ``dict`` when encountering an attribute which type is
|
| 220 |
+
``tuple``, ``dict`` or ``set``. Only meaningful if ``recurse`` is
|
| 221 |
+
``True``.
|
| 222 |
+
|
| 223 |
+
:rtype: return type of *tuple_factory*
|
| 224 |
+
|
| 225 |
+
:raise attrs.exceptions.NotAnAttrsClassError: If *cls* is not an *attrs*
|
| 226 |
+
class.
|
| 227 |
+
|
| 228 |
+
.. versionadded:: 16.2.0
|
| 229 |
+
"""
|
| 230 |
+
attrs = fields(inst.__class__)
|
| 231 |
+
rv = []
|
| 232 |
+
retain = retain_collection_types # Very long. :/
|
| 233 |
+
for a in attrs:
|
| 234 |
+
v = getattr(inst, a.name)
|
| 235 |
+
if filter is not None and not filter(a, v):
|
| 236 |
+
continue
|
| 237 |
+
if recurse is True:
|
| 238 |
+
if has(v.__class__):
|
| 239 |
+
rv.append(
|
| 240 |
+
astuple(
|
| 241 |
+
v,
|
| 242 |
+
recurse=True,
|
| 243 |
+
filter=filter,
|
| 244 |
+
tuple_factory=tuple_factory,
|
| 245 |
+
retain_collection_types=retain,
|
| 246 |
+
)
|
| 247 |
+
)
|
| 248 |
+
elif isinstance(v, (tuple, list, set, frozenset)):
|
| 249 |
+
cf = v.__class__ if retain is True else list
|
| 250 |
+
items = [
|
| 251 |
+
astuple(
|
| 252 |
+
j,
|
| 253 |
+
recurse=True,
|
| 254 |
+
filter=filter,
|
| 255 |
+
tuple_factory=tuple_factory,
|
| 256 |
+
retain_collection_types=retain,
|
| 257 |
+
)
|
| 258 |
+
if has(j.__class__)
|
| 259 |
+
else j
|
| 260 |
+
for j in v
|
| 261 |
+
]
|
| 262 |
+
try:
|
| 263 |
+
rv.append(cf(items))
|
| 264 |
+
except TypeError:
|
| 265 |
+
if not issubclass(cf, tuple):
|
| 266 |
+
raise
|
| 267 |
+
# Workaround for TypeError: cf.__new__() missing 1 required
|
| 268 |
+
# positional argument (which appears, for a namedturle)
|
| 269 |
+
rv.append(cf(*items))
|
| 270 |
+
elif isinstance(v, dict):
|
| 271 |
+
df = v.__class__ if retain is True else dict
|
| 272 |
+
rv.append(
|
| 273 |
+
df(
|
| 274 |
+
(
|
| 275 |
+
astuple(
|
| 276 |
+
kk,
|
| 277 |
+
tuple_factory=tuple_factory,
|
| 278 |
+
retain_collection_types=retain,
|
| 279 |
+
)
|
| 280 |
+
if has(kk.__class__)
|
| 281 |
+
else kk,
|
| 282 |
+
astuple(
|
| 283 |
+
vv,
|
| 284 |
+
tuple_factory=tuple_factory,
|
| 285 |
+
retain_collection_types=retain,
|
| 286 |
+
)
|
| 287 |
+
if has(vv.__class__)
|
| 288 |
+
else vv,
|
| 289 |
+
)
|
| 290 |
+
for kk, vv in v.items()
|
| 291 |
+
)
|
| 292 |
+
)
|
| 293 |
+
else:
|
| 294 |
+
rv.append(v)
|
| 295 |
+
else:
|
| 296 |
+
rv.append(v)
|
| 297 |
+
|
| 298 |
+
return rv if tuple_factory is list else tuple_factory(rv)
|
| 299 |
+
|
| 300 |
+
|
| 301 |
+
def has(cls):
|
| 302 |
+
"""
|
| 303 |
+
Check whether *cls* is a class with *attrs* attributes.
|
| 304 |
+
|
| 305 |
+
:param type cls: Class to introspect.
|
| 306 |
+
:raise TypeError: If *cls* is not a class.
|
| 307 |
+
|
| 308 |
+
:rtype: bool
|
| 309 |
+
"""
|
| 310 |
+
attrs = getattr(cls, "__attrs_attrs__", None)
|
| 311 |
+
if attrs is not None:
|
| 312 |
+
return True
|
| 313 |
+
|
| 314 |
+
# No attrs, maybe it's a specialized generic (A[str])?
|
| 315 |
+
generic_base = get_generic_base(cls)
|
| 316 |
+
if generic_base is not None:
|
| 317 |
+
generic_attrs = getattr(generic_base, "__attrs_attrs__", None)
|
| 318 |
+
if generic_attrs is not None:
|
| 319 |
+
# Stick it on here for speed next time.
|
| 320 |
+
cls.__attrs_attrs__ = generic_attrs
|
| 321 |
+
return generic_attrs is not None
|
| 322 |
+
return False
|
| 323 |
+
|
| 324 |
+
|
| 325 |
+
def assoc(inst, **changes):
|
| 326 |
+
"""
|
| 327 |
+
Copy *inst* and apply *changes*.
|
| 328 |
+
|
| 329 |
+
This is different from `evolve` that applies the changes to the arguments
|
| 330 |
+
that create the new instance.
|
| 331 |
+
|
| 332 |
+
`evolve`'s behavior is preferable, but there are `edge cases`_ where it
|
| 333 |
+
doesn't work. Therefore `assoc` is deprecated, but will not be removed.
|
| 334 |
+
|
| 335 |
+
.. _`edge cases`: https://github.com/python-attrs/attrs/issues/251
|
| 336 |
+
|
| 337 |
+
:param inst: Instance of a class with *attrs* attributes.
|
| 338 |
+
:param changes: Keyword changes in the new copy.
|
| 339 |
+
|
| 340 |
+
:return: A copy of inst with *changes* incorporated.
|
| 341 |
+
|
| 342 |
+
:raise attrs.exceptions.AttrsAttributeNotFoundError: If *attr_name*
|
| 343 |
+
couldn't be found on *cls*.
|
| 344 |
+
:raise attrs.exceptions.NotAnAttrsClassError: If *cls* is not an *attrs*
|
| 345 |
+
class.
|
| 346 |
+
|
| 347 |
+
.. deprecated:: 17.1.0
|
| 348 |
+
Use `attrs.evolve` instead if you can.
|
| 349 |
+
This function will not be removed du to the slightly different approach
|
| 350 |
+
compared to `attrs.evolve`.
|
| 351 |
+
"""
|
| 352 |
+
new = copy.copy(inst)
|
| 353 |
+
attrs = fields(inst.__class__)
|
| 354 |
+
for k, v in changes.items():
|
| 355 |
+
a = getattr(attrs, k, NOTHING)
|
| 356 |
+
if a is NOTHING:
|
| 357 |
+
msg = f"{k} is not an attrs attribute on {new.__class__}."
|
| 358 |
+
raise AttrsAttributeNotFoundError(msg)
|
| 359 |
+
_obj_setattr(new, k, v)
|
| 360 |
+
return new
|
| 361 |
+
|
| 362 |
+
|
| 363 |
+
def evolve(*args, **changes):
|
| 364 |
+
"""
|
| 365 |
+
Create a new instance, based on the first positional argument with
|
| 366 |
+
*changes* applied.
|
| 367 |
+
|
| 368 |
+
:param inst: Instance of a class with *attrs* attributes.
|
| 369 |
+
:param changes: Keyword changes in the new copy.
|
| 370 |
+
|
| 371 |
+
:return: A copy of inst with *changes* incorporated.
|
| 372 |
+
|
| 373 |
+
:raise TypeError: If *attr_name* couldn't be found in the class
|
| 374 |
+
``__init__``.
|
| 375 |
+
:raise attrs.exceptions.NotAnAttrsClassError: If *cls* is not an *attrs*
|
| 376 |
+
class.
|
| 377 |
+
|
| 378 |
+
.. versionadded:: 17.1.0
|
| 379 |
+
.. deprecated:: 23.1.0
|
| 380 |
+
It is now deprecated to pass the instance using the keyword argument
|
| 381 |
+
*inst*. It will raise a warning until at least April 2024, after which
|
| 382 |
+
it will become an error. Always pass the instance as a positional
|
| 383 |
+
argument.
|
| 384 |
+
"""
|
| 385 |
+
# Try to get instance by positional argument first.
|
| 386 |
+
# Use changes otherwise and warn it'll break.
|
| 387 |
+
if args:
|
| 388 |
+
try:
|
| 389 |
+
(inst,) = args
|
| 390 |
+
except ValueError:
|
| 391 |
+
msg = f"evolve() takes 1 positional argument, but {len(args)} were given"
|
| 392 |
+
raise TypeError(msg) from None
|
| 393 |
+
else:
|
| 394 |
+
try:
|
| 395 |
+
inst = changes.pop("inst")
|
| 396 |
+
except KeyError:
|
| 397 |
+
msg = "evolve() missing 1 required positional argument: 'inst'"
|
| 398 |
+
raise TypeError(msg) from None
|
| 399 |
+
|
| 400 |
+
import warnings
|
| 401 |
+
|
| 402 |
+
warnings.warn(
|
| 403 |
+
"Passing the instance per keyword argument is deprecated and "
|
| 404 |
+
"will stop working in, or after, April 2024.",
|
| 405 |
+
DeprecationWarning,
|
| 406 |
+
stacklevel=2,
|
| 407 |
+
)
|
| 408 |
+
|
| 409 |
+
cls = inst.__class__
|
| 410 |
+
attrs = fields(cls)
|
| 411 |
+
for a in attrs:
|
| 412 |
+
if not a.init:
|
| 413 |
+
continue
|
| 414 |
+
attr_name = a.name # To deal with private attributes.
|
| 415 |
+
init_name = a.alias
|
| 416 |
+
if init_name not in changes:
|
| 417 |
+
changes[init_name] = getattr(inst, attr_name)
|
| 418 |
+
|
| 419 |
+
return cls(**changes)
|
| 420 |
+
|
| 421 |
+
|
| 422 |
+
def resolve_types(
|
| 423 |
+
cls, globalns=None, localns=None, attribs=None, include_extras=True
|
| 424 |
+
):
|
| 425 |
+
"""
|
| 426 |
+
Resolve any strings and forward annotations in type annotations.
|
| 427 |
+
|
| 428 |
+
This is only required if you need concrete types in `Attribute`'s *type*
|
| 429 |
+
field. In other words, you don't need to resolve your types if you only
|
| 430 |
+
use them for static type checking.
|
| 431 |
+
|
| 432 |
+
With no arguments, names will be looked up in the module in which the class
|
| 433 |
+
was created. If this is not what you want, e.g. if the name only exists
|
| 434 |
+
inside a method, you may pass *globalns* or *localns* to specify other
|
| 435 |
+
dictionaries in which to look up these names. See the docs of
|
| 436 |
+
`typing.get_type_hints` for more details.
|
| 437 |
+
|
| 438 |
+
:param type cls: Class to resolve.
|
| 439 |
+
:param Optional[dict] globalns: Dictionary containing global variables.
|
| 440 |
+
:param Optional[dict] localns: Dictionary containing local variables.
|
| 441 |
+
:param Optional[list] attribs: List of attribs for the given class.
|
| 442 |
+
This is necessary when calling from inside a ``field_transformer``
|
| 443 |
+
since *cls* is not an *attrs* class yet.
|
| 444 |
+
:param bool include_extras: Resolve more accurately, if possible.
|
| 445 |
+
Pass ``include_extras`` to ``typing.get_hints``, if supported by the
|
| 446 |
+
typing module. On supported Python versions (3.9+), this resolves the
|
| 447 |
+
types more accurately.
|
| 448 |
+
|
| 449 |
+
:raise TypeError: If *cls* is not a class.
|
| 450 |
+
:raise attrs.exceptions.NotAnAttrsClassError: If *cls* is not an *attrs*
|
| 451 |
+
class and you didn't pass any attribs.
|
| 452 |
+
:raise NameError: If types cannot be resolved because of missing variables.
|
| 453 |
+
|
| 454 |
+
:returns: *cls* so you can use this function also as a class decorator.
|
| 455 |
+
Please note that you have to apply it **after** `attrs.define`. That
|
| 456 |
+
means the decorator has to come in the line **before** `attrs.define`.
|
| 457 |
+
|
| 458 |
+
.. versionadded:: 20.1.0
|
| 459 |
+
.. versionadded:: 21.1.0 *attribs*
|
| 460 |
+
.. versionadded:: 23.1.0 *include_extras*
|
| 461 |
+
|
| 462 |
+
"""
|
| 463 |
+
# Since calling get_type_hints is expensive we cache whether we've
|
| 464 |
+
# done it already.
|
| 465 |
+
if getattr(cls, "__attrs_types_resolved__", None) != cls:
|
| 466 |
+
import typing
|
| 467 |
+
|
| 468 |
+
kwargs = {"globalns": globalns, "localns": localns}
|
| 469 |
+
|
| 470 |
+
if PY_3_9_PLUS:
|
| 471 |
+
kwargs["include_extras"] = include_extras
|
| 472 |
+
|
| 473 |
+
hints = typing.get_type_hints(cls, **kwargs)
|
| 474 |
+
for field in fields(cls) if attribs is None else attribs:
|
| 475 |
+
if field.name in hints:
|
| 476 |
+
# Since fields have been frozen we must work around it.
|
| 477 |
+
_obj_setattr(field, "type", hints[field.name])
|
| 478 |
+
# We store the class we resolved so that subclasses know they haven't
|
| 479 |
+
# been resolved.
|
| 480 |
+
cls.__attrs_types_resolved__ = cls
|
| 481 |
+
|
| 482 |
+
# Return the class so you can use it as a decorator too.
|
| 483 |
+
return cls
|
env-llmeval/lib/python3.10/site-packages/attr/_make.py
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
env-llmeval/lib/python3.10/site-packages/attr/_version_info.pyi
ADDED
|
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
class VersionInfo:
|
| 2 |
+
@property
|
| 3 |
+
def year(self) -> int: ...
|
| 4 |
+
@property
|
| 5 |
+
def minor(self) -> int: ...
|
| 6 |
+
@property
|
| 7 |
+
def micro(self) -> int: ...
|
| 8 |
+
@property
|
| 9 |
+
def releaselevel(self) -> str: ...
|
env-llmeval/lib/python3.10/site-packages/attr/converters.pyi
ADDED
|
@@ -0,0 +1,13 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from typing import Callable, TypeVar, overload
|
| 2 |
+
|
| 3 |
+
from . import _ConverterType
|
| 4 |
+
|
| 5 |
+
_T = TypeVar("_T")
|
| 6 |
+
|
| 7 |
+
def pipe(*validators: _ConverterType) -> _ConverterType: ...
|
| 8 |
+
def optional(converter: _ConverterType) -> _ConverterType: ...
|
| 9 |
+
@overload
|
| 10 |
+
def default_if_none(default: _T) -> _ConverterType: ...
|
| 11 |
+
@overload
|
| 12 |
+
def default_if_none(*, factory: Callable[[], _T]) -> _ConverterType: ...
|
| 13 |
+
def to_bool(val: str) -> bool: ...
|
env-llmeval/lib/python3.10/site-packages/attr/exceptions.py
ADDED
|
@@ -0,0 +1,95 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# SPDX-License-Identifier: MIT
|
| 2 |
+
|
| 3 |
+
from __future__ import annotations
|
| 4 |
+
|
| 5 |
+
from typing import ClassVar
|
| 6 |
+
|
| 7 |
+
|
| 8 |
+
class FrozenError(AttributeError):
|
| 9 |
+
"""
|
| 10 |
+
A frozen/immutable instance or attribute have been attempted to be
|
| 11 |
+
modified.
|
| 12 |
+
|
| 13 |
+
It mirrors the behavior of ``namedtuples`` by using the same error message
|
| 14 |
+
and subclassing `AttributeError`.
|
| 15 |
+
|
| 16 |
+
.. versionadded:: 20.1.0
|
| 17 |
+
"""
|
| 18 |
+
|
| 19 |
+
msg = "can't set attribute"
|
| 20 |
+
args: ClassVar[tuple[str]] = [msg]
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
class FrozenInstanceError(FrozenError):
|
| 24 |
+
"""
|
| 25 |
+
A frozen instance has been attempted to be modified.
|
| 26 |
+
|
| 27 |
+
.. versionadded:: 16.1.0
|
| 28 |
+
"""
|
| 29 |
+
|
| 30 |
+
|
| 31 |
+
class FrozenAttributeError(FrozenError):
|
| 32 |
+
"""
|
| 33 |
+
A frozen attribute has been attempted to be modified.
|
| 34 |
+
|
| 35 |
+
.. versionadded:: 20.1.0
|
| 36 |
+
"""
|
| 37 |
+
|
| 38 |
+
|
| 39 |
+
class AttrsAttributeNotFoundError(ValueError):
|
| 40 |
+
"""
|
| 41 |
+
An *attrs* function couldn't find an attribute that the user asked for.
|
| 42 |
+
|
| 43 |
+
.. versionadded:: 16.2.0
|
| 44 |
+
"""
|
| 45 |
+
|
| 46 |
+
|
| 47 |
+
class NotAnAttrsClassError(ValueError):
|
| 48 |
+
"""
|
| 49 |
+
A non-*attrs* class has been passed into an *attrs* function.
|
| 50 |
+
|
| 51 |
+
.. versionadded:: 16.2.0
|
| 52 |
+
"""
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
class DefaultAlreadySetError(RuntimeError):
|
| 56 |
+
"""
|
| 57 |
+
A default has been set when defining the field and is attempted to be reset
|
| 58 |
+
using the decorator.
|
| 59 |
+
|
| 60 |
+
.. versionadded:: 17.1.0
|
| 61 |
+
"""
|
| 62 |
+
|
| 63 |
+
|
| 64 |
+
class UnannotatedAttributeError(RuntimeError):
|
| 65 |
+
"""
|
| 66 |
+
A class with ``auto_attribs=True`` has a field without a type annotation.
|
| 67 |
+
|
| 68 |
+
.. versionadded:: 17.3.0
|
| 69 |
+
"""
|
| 70 |
+
|
| 71 |
+
|
| 72 |
+
class PythonTooOldError(RuntimeError):
|
| 73 |
+
"""
|
| 74 |
+
It was attempted to use an *attrs* feature that requires a newer Python
|
| 75 |
+
version.
|
| 76 |
+
|
| 77 |
+
.. versionadded:: 18.2.0
|
| 78 |
+
"""
|
| 79 |
+
|
| 80 |
+
|
| 81 |
+
class NotCallableError(TypeError):
|
| 82 |
+
"""
|
| 83 |
+
A field requiring a callable has been set with a value that is not
|
| 84 |
+
callable.
|
| 85 |
+
|
| 86 |
+
.. versionadded:: 19.2.0
|
| 87 |
+
"""
|
| 88 |
+
|
| 89 |
+
def __init__(self, msg, value):
|
| 90 |
+
super(TypeError, self).__init__(msg, value)
|
| 91 |
+
self.msg = msg
|
| 92 |
+
self.value = value
|
| 93 |
+
|
| 94 |
+
def __str__(self):
|
| 95 |
+
return str(self.msg)
|
env-llmeval/lib/python3.10/site-packages/attr/exceptions.pyi
ADDED
|
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from typing import Any
|
| 2 |
+
|
| 3 |
+
class FrozenError(AttributeError):
|
| 4 |
+
msg: str = ...
|
| 5 |
+
|
| 6 |
+
class FrozenInstanceError(FrozenError): ...
|
| 7 |
+
class FrozenAttributeError(FrozenError): ...
|
| 8 |
+
class AttrsAttributeNotFoundError(ValueError): ...
|
| 9 |
+
class NotAnAttrsClassError(ValueError): ...
|
| 10 |
+
class DefaultAlreadySetError(RuntimeError): ...
|
| 11 |
+
class UnannotatedAttributeError(RuntimeError): ...
|
| 12 |
+
class PythonTooOldError(RuntimeError): ...
|
| 13 |
+
|
| 14 |
+
class NotCallableError(TypeError):
|
| 15 |
+
msg: str = ...
|
| 16 |
+
value: Any = ...
|
| 17 |
+
def __init__(self, msg: str, value: Any) -> None: ...
|
env-llmeval/lib/python3.10/site-packages/attr/filters.pyi
ADDED
|
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from typing import Any, Union
|
| 2 |
+
|
| 3 |
+
from . import Attribute, _FilterType
|
| 4 |
+
|
| 5 |
+
def include(*what: Union[type, str, Attribute[Any]]) -> _FilterType[Any]: ...
|
| 6 |
+
def exclude(*what: Union[type, str, Attribute[Any]]) -> _FilterType[Any]: ...
|
env-llmeval/lib/python3.10/site-packages/attr/py.typed
ADDED
|
File without changes
|
env-llmeval/lib/python3.10/site-packages/attr/setters.py
ADDED
|
@@ -0,0 +1,73 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# SPDX-License-Identifier: MIT
|
| 2 |
+
|
| 3 |
+
"""
|
| 4 |
+
Commonly used hooks for on_setattr.
|
| 5 |
+
"""
|
| 6 |
+
|
| 7 |
+
|
| 8 |
+
from . import _config
|
| 9 |
+
from .exceptions import FrozenAttributeError
|
| 10 |
+
|
| 11 |
+
|
| 12 |
+
def pipe(*setters):
|
| 13 |
+
"""
|
| 14 |
+
Run all *setters* and return the return value of the last one.
|
| 15 |
+
|
| 16 |
+
.. versionadded:: 20.1.0
|
| 17 |
+
"""
|
| 18 |
+
|
| 19 |
+
def wrapped_pipe(instance, attrib, new_value):
|
| 20 |
+
rv = new_value
|
| 21 |
+
|
| 22 |
+
for setter in setters:
|
| 23 |
+
rv = setter(instance, attrib, rv)
|
| 24 |
+
|
| 25 |
+
return rv
|
| 26 |
+
|
| 27 |
+
return wrapped_pipe
|
| 28 |
+
|
| 29 |
+
|
| 30 |
+
def frozen(_, __, ___):
|
| 31 |
+
"""
|
| 32 |
+
Prevent an attribute to be modified.
|
| 33 |
+
|
| 34 |
+
.. versionadded:: 20.1.0
|
| 35 |
+
"""
|
| 36 |
+
raise FrozenAttributeError()
|
| 37 |
+
|
| 38 |
+
|
| 39 |
+
def validate(instance, attrib, new_value):
|
| 40 |
+
"""
|
| 41 |
+
Run *attrib*'s validator on *new_value* if it has one.
|
| 42 |
+
|
| 43 |
+
.. versionadded:: 20.1.0
|
| 44 |
+
"""
|
| 45 |
+
if _config._run_validators is False:
|
| 46 |
+
return new_value
|
| 47 |
+
|
| 48 |
+
v = attrib.validator
|
| 49 |
+
if not v:
|
| 50 |
+
return new_value
|
| 51 |
+
|
| 52 |
+
v(instance, attrib, new_value)
|
| 53 |
+
|
| 54 |
+
return new_value
|
| 55 |
+
|
| 56 |
+
|
| 57 |
+
def convert(instance, attrib, new_value):
|
| 58 |
+
"""
|
| 59 |
+
Run *attrib*'s converter -- if it has one -- on *new_value* and return the
|
| 60 |
+
result.
|
| 61 |
+
|
| 62 |
+
.. versionadded:: 20.1.0
|
| 63 |
+
"""
|
| 64 |
+
c = attrib.converter
|
| 65 |
+
if c:
|
| 66 |
+
return c(new_value)
|
| 67 |
+
|
| 68 |
+
return new_value
|
| 69 |
+
|
| 70 |
+
|
| 71 |
+
# Sentinel for disabling class-wide *on_setattr* hooks for certain attributes.
|
| 72 |
+
# autodata stopped working, so the docstring is inlined in the API docs.
|
| 73 |
+
NO_OP = object()
|
env-llmeval/lib/python3.10/site-packages/attr/validators.pyi
ADDED
|
@@ -0,0 +1,88 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
from typing import (
|
| 2 |
+
Any,
|
| 3 |
+
AnyStr,
|
| 4 |
+
Callable,
|
| 5 |
+
Container,
|
| 6 |
+
ContextManager,
|
| 7 |
+
Iterable,
|
| 8 |
+
List,
|
| 9 |
+
Mapping,
|
| 10 |
+
Match,
|
| 11 |
+
Optional,
|
| 12 |
+
Pattern,
|
| 13 |
+
Tuple,
|
| 14 |
+
Type,
|
| 15 |
+
TypeVar,
|
| 16 |
+
Union,
|
| 17 |
+
overload,
|
| 18 |
+
)
|
| 19 |
+
|
| 20 |
+
from . import _ValidatorType
|
| 21 |
+
from . import _ValidatorArgType
|
| 22 |
+
|
| 23 |
+
_T = TypeVar("_T")
|
| 24 |
+
_T1 = TypeVar("_T1")
|
| 25 |
+
_T2 = TypeVar("_T2")
|
| 26 |
+
_T3 = TypeVar("_T3")
|
| 27 |
+
_I = TypeVar("_I", bound=Iterable)
|
| 28 |
+
_K = TypeVar("_K")
|
| 29 |
+
_V = TypeVar("_V")
|
| 30 |
+
_M = TypeVar("_M", bound=Mapping)
|
| 31 |
+
|
| 32 |
+
def set_disabled(run: bool) -> None: ...
|
| 33 |
+
def get_disabled() -> bool: ...
|
| 34 |
+
def disabled() -> ContextManager[None]: ...
|
| 35 |
+
|
| 36 |
+
# To be more precise on instance_of use some overloads.
|
| 37 |
+
# If there are more than 3 items in the tuple then we fall back to Any
|
| 38 |
+
@overload
|
| 39 |
+
def instance_of(type: Type[_T]) -> _ValidatorType[_T]: ...
|
| 40 |
+
@overload
|
| 41 |
+
def instance_of(type: Tuple[Type[_T]]) -> _ValidatorType[_T]: ...
|
| 42 |
+
@overload
|
| 43 |
+
def instance_of(
|
| 44 |
+
type: Tuple[Type[_T1], Type[_T2]]
|
| 45 |
+
) -> _ValidatorType[Union[_T1, _T2]]: ...
|
| 46 |
+
@overload
|
| 47 |
+
def instance_of(
|
| 48 |
+
type: Tuple[Type[_T1], Type[_T2], Type[_T3]]
|
| 49 |
+
) -> _ValidatorType[Union[_T1, _T2, _T3]]: ...
|
| 50 |
+
@overload
|
| 51 |
+
def instance_of(type: Tuple[type, ...]) -> _ValidatorType[Any]: ...
|
| 52 |
+
def provides(interface: Any) -> _ValidatorType[Any]: ...
|
| 53 |
+
def optional(
|
| 54 |
+
validator: Union[
|
| 55 |
+
_ValidatorType[_T], List[_ValidatorType[_T]], Tuple[_ValidatorType[_T]]
|
| 56 |
+
]
|
| 57 |
+
) -> _ValidatorType[Optional[_T]]: ...
|
| 58 |
+
def in_(options: Container[_T]) -> _ValidatorType[_T]: ...
|
| 59 |
+
def and_(*validators: _ValidatorType[_T]) -> _ValidatorType[_T]: ...
|
| 60 |
+
def matches_re(
|
| 61 |
+
regex: Union[Pattern[AnyStr], AnyStr],
|
| 62 |
+
flags: int = ...,
|
| 63 |
+
func: Optional[
|
| 64 |
+
Callable[[AnyStr, AnyStr, int], Optional[Match[AnyStr]]]
|
| 65 |
+
] = ...,
|
| 66 |
+
) -> _ValidatorType[AnyStr]: ...
|
| 67 |
+
def deep_iterable(
|
| 68 |
+
member_validator: _ValidatorArgType[_T],
|
| 69 |
+
iterable_validator: Optional[_ValidatorType[_I]] = ...,
|
| 70 |
+
) -> _ValidatorType[_I]: ...
|
| 71 |
+
def deep_mapping(
|
| 72 |
+
key_validator: _ValidatorType[_K],
|
| 73 |
+
value_validator: _ValidatorType[_V],
|
| 74 |
+
mapping_validator: Optional[_ValidatorType[_M]] = ...,
|
| 75 |
+
) -> _ValidatorType[_M]: ...
|
| 76 |
+
def is_callable() -> _ValidatorType[_T]: ...
|
| 77 |
+
def lt(val: _T) -> _ValidatorType[_T]: ...
|
| 78 |
+
def le(val: _T) -> _ValidatorType[_T]: ...
|
| 79 |
+
def ge(val: _T) -> _ValidatorType[_T]: ...
|
| 80 |
+
def gt(val: _T) -> _ValidatorType[_T]: ...
|
| 81 |
+
def max_len(length: int) -> _ValidatorType[_T]: ...
|
| 82 |
+
def min_len(length: int) -> _ValidatorType[_T]: ...
|
| 83 |
+
def not_(
|
| 84 |
+
validator: _ValidatorType[_T],
|
| 85 |
+
*,
|
| 86 |
+
msg: Optional[str] = None,
|
| 87 |
+
exc_types: Union[Type[Exception], Iterable[Type[Exception]]] = ...,
|
| 88 |
+
) -> _ValidatorType[_T]: ...
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/__init__.py
ADDED
|
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2024 The HuggingFace Team. All rights reserved.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
# ruff: noqa: F401
|
| 15 |
+
"""Contains helpers to serialize tensors."""
|
| 16 |
+
|
| 17 |
+
from ._base import StateDictSplit, split_state_dict_into_shards_factory
|
| 18 |
+
from ._numpy import split_numpy_state_dict_into_shards
|
| 19 |
+
from ._tensorflow import split_tf_state_dict_into_shards
|
| 20 |
+
from ._torch import split_torch_state_dict_into_shards
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/__pycache__/__init__.cpython-310.pyc
ADDED
|
Binary file (537 Bytes). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/__pycache__/_base.cpython-310.pyc
ADDED
|
Binary file (4.69 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/__pycache__/_numpy.cpython-310.pyc
ADDED
|
Binary file (2.3 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/__pycache__/_tensorflow.cpython-310.pyc
ADDED
|
Binary file (3.09 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/__pycache__/_torch.cpython-310.pyc
ADDED
|
Binary file (6.5 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/_base.py
ADDED
|
@@ -0,0 +1,169 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2024 The HuggingFace Team. All rights reserved.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
"""Contains helpers to split tensors into shards."""
|
| 15 |
+
|
| 16 |
+
from dataclasses import dataclass, field
|
| 17 |
+
from typing import Any, Callable, Dict, List, Optional, TypeVar
|
| 18 |
+
|
| 19 |
+
from .. import logging
|
| 20 |
+
|
| 21 |
+
|
| 22 |
+
TensorT = TypeVar("TensorT")
|
| 23 |
+
TensorSizeFn_T = Callable[[TensorT], int]
|
| 24 |
+
StorageIDFn_T = Callable[[TensorT], Optional[Any]]
|
| 25 |
+
|
| 26 |
+
MAX_SHARD_SIZE = 5_000_000_000 # 5GB
|
| 27 |
+
FILENAME_PATTERN = "model{suffix}.safetensors"
|
| 28 |
+
|
| 29 |
+
logger = logging.get_logger(__file__)
|
| 30 |
+
|
| 31 |
+
|
| 32 |
+
@dataclass
|
| 33 |
+
class StateDictSplit:
|
| 34 |
+
is_sharded: bool = field(init=False)
|
| 35 |
+
metadata: Dict[str, Any]
|
| 36 |
+
filename_to_tensors: Dict[str, List[str]]
|
| 37 |
+
tensor_to_filename: Dict[str, str]
|
| 38 |
+
|
| 39 |
+
def __post_init__(self):
|
| 40 |
+
self.is_sharded = len(self.filename_to_tensors) > 1
|
| 41 |
+
|
| 42 |
+
|
| 43 |
+
def split_state_dict_into_shards_factory(
|
| 44 |
+
state_dict: Dict[str, TensorT],
|
| 45 |
+
*,
|
| 46 |
+
get_tensor_size: TensorSizeFn_T,
|
| 47 |
+
get_storage_id: StorageIDFn_T = lambda tensor: None,
|
| 48 |
+
filename_pattern: str = FILENAME_PATTERN,
|
| 49 |
+
max_shard_size: int = MAX_SHARD_SIZE,
|
| 50 |
+
) -> StateDictSplit:
|
| 51 |
+
"""
|
| 52 |
+
Split a model state dictionary in shards so that each shard is smaller than a given size.
|
| 53 |
+
|
| 54 |
+
The shards are determined by iterating through the `state_dict` in the order of its keys. There is no optimization
|
| 55 |
+
made to make each shard as close as possible to the maximum size passed. For example, if the limit is 10GB and we
|
| 56 |
+
have tensors of sizes [6GB, 6GB, 2GB, 6GB, 2GB, 2GB] they will get sharded as [6GB], [6+2GB], [6+2+2GB] and not
|
| 57 |
+
[6+2+2GB], [6+2GB], [6GB].
|
| 58 |
+
|
| 59 |
+
<Tip warning={true}>
|
| 60 |
+
|
| 61 |
+
If one of the model's tensor is bigger than `max_shard_size`, it will end up in its own shard which will have a
|
| 62 |
+
size greater than `max_shard_size`.
|
| 63 |
+
|
| 64 |
+
</Tip>
|
| 65 |
+
|
| 66 |
+
Args:
|
| 67 |
+
state_dict (`Dict[str, Tensor]`):
|
| 68 |
+
The state dictionary to save.
|
| 69 |
+
get_tensor_size (`Callable[[Tensor], int]`):
|
| 70 |
+
A function that returns the size of a tensor in bytes.
|
| 71 |
+
get_storage_id (`Callable[[Tensor], Optional[Any]]`, *optional*):
|
| 72 |
+
A function that returns a unique identifier to a tensor storage. Multiple different tensors can share the
|
| 73 |
+
same underlying storage. This identifier is guaranteed to be unique and constant for this tensor's storage
|
| 74 |
+
during its lifetime. Two tensor storages with non-overlapping lifetimes may have the same id.
|
| 75 |
+
filename_pattern (`str`, *optional*):
|
| 76 |
+
The pattern to generate the files names in which the model will be saved. Pattern must be a string that
|
| 77 |
+
can be formatted with `filename_pattern.format(suffix=...)` and must contain the keyword `suffix`
|
| 78 |
+
Defaults to `"model{suffix}.safetensors"`.
|
| 79 |
+
max_shard_size (`int` or `str`, *optional*):
|
| 80 |
+
The maximum size of each shard, in bytes. Defaults to 5GB.
|
| 81 |
+
|
| 82 |
+
Returns:
|
| 83 |
+
[`StateDictSplit`]: A `StateDictSplit` object containing the shards and the index to retrieve them.
|
| 84 |
+
"""
|
| 85 |
+
storage_id_to_tensors: Dict[Any, List[str]] = {}
|
| 86 |
+
|
| 87 |
+
shard_list: List[Dict[str, TensorT]] = []
|
| 88 |
+
current_shard: Dict[str, TensorT] = {}
|
| 89 |
+
current_shard_size = 0
|
| 90 |
+
total_size = 0
|
| 91 |
+
|
| 92 |
+
for key, tensor in state_dict.items():
|
| 93 |
+
# when bnb serialization is used the weights in the state dict can be strings
|
| 94 |
+
# check: https://github.com/huggingface/transformers/pull/24416 for more details
|
| 95 |
+
if isinstance(tensor, str):
|
| 96 |
+
logger.info("Skipping tensor %s as it is a string (bnb serialization)", key)
|
| 97 |
+
continue
|
| 98 |
+
|
| 99 |
+
# If a `tensor` shares the same underlying storage as another tensor, we put `tensor` in the same `block`
|
| 100 |
+
storage_id = get_storage_id(tensor)
|
| 101 |
+
if storage_id is not None:
|
| 102 |
+
if storage_id in storage_id_to_tensors:
|
| 103 |
+
# We skip this tensor for now and will reassign to correct shard later
|
| 104 |
+
storage_id_to_tensors[storage_id].append(key)
|
| 105 |
+
continue
|
| 106 |
+
else:
|
| 107 |
+
# This is the first tensor with this storage_id, we create a new entry
|
| 108 |
+
# in the storage_id_to_tensors dict => we will assign the shard id later
|
| 109 |
+
storage_id_to_tensors[storage_id] = [key]
|
| 110 |
+
|
| 111 |
+
# Compute tensor size
|
| 112 |
+
tensor_size = get_tensor_size(tensor)
|
| 113 |
+
|
| 114 |
+
# If this tensor is bigger than the maximal size, we put it in its own shard
|
| 115 |
+
if tensor_size > max_shard_size:
|
| 116 |
+
total_size += tensor_size
|
| 117 |
+
shard_list.append({key: tensor})
|
| 118 |
+
continue
|
| 119 |
+
|
| 120 |
+
# If this tensor is going to tip up over the maximal size, we split.
|
| 121 |
+
# Current shard already has some tensors, we add it to the list of shards and create a new one.
|
| 122 |
+
if current_shard_size + tensor_size > max_shard_size:
|
| 123 |
+
shard_list.append(current_shard)
|
| 124 |
+
current_shard = {}
|
| 125 |
+
current_shard_size = 0
|
| 126 |
+
|
| 127 |
+
# Add the tensor to the current shard
|
| 128 |
+
current_shard[key] = tensor
|
| 129 |
+
current_shard_size += tensor_size
|
| 130 |
+
total_size += tensor_size
|
| 131 |
+
|
| 132 |
+
# Add the last shard
|
| 133 |
+
if len(current_shard) > 0:
|
| 134 |
+
shard_list.append(current_shard)
|
| 135 |
+
nb_shards = len(shard_list)
|
| 136 |
+
|
| 137 |
+
# Loop over the tensors that share the same storage and assign them together
|
| 138 |
+
for storage_id, keys in storage_id_to_tensors.items():
|
| 139 |
+
# Let's try to find the shard where the first tensor of this storage is and put all tensors in the same shard
|
| 140 |
+
for shard in shard_list:
|
| 141 |
+
if keys[0] in shard:
|
| 142 |
+
for key in keys:
|
| 143 |
+
shard[key] = state_dict[key]
|
| 144 |
+
break
|
| 145 |
+
|
| 146 |
+
# If we only have one shard, we return it => no need to build the index
|
| 147 |
+
if nb_shards == 1:
|
| 148 |
+
filename = filename_pattern.format(suffix="")
|
| 149 |
+
return StateDictSplit(
|
| 150 |
+
metadata={"total_size": total_size},
|
| 151 |
+
filename_to_tensors={filename: list(state_dict.keys())},
|
| 152 |
+
tensor_to_filename={key: filename for key in state_dict.keys()},
|
| 153 |
+
)
|
| 154 |
+
|
| 155 |
+
# Now that each tensor is assigned to a shard, let's assign a filename to each shard
|
| 156 |
+
tensor_name_to_filename = {}
|
| 157 |
+
filename_to_tensors = {}
|
| 158 |
+
for idx, shard in enumerate(shard_list):
|
| 159 |
+
filename = filename_pattern.format(suffix=f"-{idx+1:05d}-of-{nb_shards:05d}")
|
| 160 |
+
for key in shard:
|
| 161 |
+
tensor_name_to_filename[key] = filename
|
| 162 |
+
filename_to_tensors[filename] = list(shard.keys())
|
| 163 |
+
|
| 164 |
+
# Build the index and return
|
| 165 |
+
return StateDictSplit(
|
| 166 |
+
metadata={"total_size": total_size},
|
| 167 |
+
filename_to_tensors=filename_to_tensors,
|
| 168 |
+
tensor_to_filename=tensor_name_to_filename,
|
| 169 |
+
)
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/_numpy.py
ADDED
|
@@ -0,0 +1,68 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2024 The HuggingFace Team. All rights reserved.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
"""Contains numpy-specific helpers."""
|
| 15 |
+
|
| 16 |
+
from typing import TYPE_CHECKING, Dict
|
| 17 |
+
|
| 18 |
+
from ._base import FILENAME_PATTERN, MAX_SHARD_SIZE, StateDictSplit, split_state_dict_into_shards_factory
|
| 19 |
+
|
| 20 |
+
|
| 21 |
+
if TYPE_CHECKING:
|
| 22 |
+
import numpy as np
|
| 23 |
+
|
| 24 |
+
|
| 25 |
+
def split_numpy_state_dict_into_shards(
|
| 26 |
+
state_dict: Dict[str, "np.ndarray"],
|
| 27 |
+
*,
|
| 28 |
+
filename_pattern: str = FILENAME_PATTERN,
|
| 29 |
+
max_shard_size: int = MAX_SHARD_SIZE,
|
| 30 |
+
) -> StateDictSplit:
|
| 31 |
+
"""
|
| 32 |
+
Split a model state dictionary in shards so that each shard is smaller than a given size.
|
| 33 |
+
|
| 34 |
+
The shards are determined by iterating through the `state_dict` in the order of its keys. There is no optimization
|
| 35 |
+
made to make each shard as close as possible to the maximum size passed. For example, if the limit is 10GB and we
|
| 36 |
+
have tensors of sizes [6GB, 6GB, 2GB, 6GB, 2GB, 2GB] they will get sharded as [6GB], [6+2GB], [6+2+2GB] and not
|
| 37 |
+
[6+2+2GB], [6+2GB], [6GB].
|
| 38 |
+
|
| 39 |
+
<Tip warning={true}>
|
| 40 |
+
|
| 41 |
+
If one of the model's tensor is bigger than `max_shard_size`, it will end up in its own shard which will have a
|
| 42 |
+
size greater than `max_shard_size`.
|
| 43 |
+
|
| 44 |
+
</Tip>
|
| 45 |
+
|
| 46 |
+
Args:
|
| 47 |
+
state_dict (`Dict[str, np.ndarray]`):
|
| 48 |
+
The state dictionary to save.
|
| 49 |
+
filename_pattern (`str`, *optional*):
|
| 50 |
+
The pattern to generate the files names in which the model will be saved. Pattern must be a string that
|
| 51 |
+
can be formatted with `filename_pattern.format(suffix=...)` and must contain the keyword `suffix`
|
| 52 |
+
Defaults to `"model{suffix}.safetensors"`.
|
| 53 |
+
max_shard_size (`int` or `str`, *optional*):
|
| 54 |
+
The maximum size of each shard, in bytes. Defaults to 5GB.
|
| 55 |
+
|
| 56 |
+
Returns:
|
| 57 |
+
[`StateDictSplit`]: A `StateDictSplit` object containing the shards and the index to retrieve them.
|
| 58 |
+
"""
|
| 59 |
+
return split_state_dict_into_shards_factory(
|
| 60 |
+
state_dict,
|
| 61 |
+
max_shard_size=max_shard_size,
|
| 62 |
+
filename_pattern=filename_pattern,
|
| 63 |
+
get_tensor_size=get_tensor_size,
|
| 64 |
+
)
|
| 65 |
+
|
| 66 |
+
|
| 67 |
+
def get_tensor_size(tensor: "np.ndarray") -> int:
|
| 68 |
+
return tensor.nbytes
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/_tensorflow.py
ADDED
|
@@ -0,0 +1,94 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2024 The HuggingFace Team. All rights reserved.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
"""Contains tensorflow-specific helpers."""
|
| 15 |
+
|
| 16 |
+
import math
|
| 17 |
+
import re
|
| 18 |
+
from typing import TYPE_CHECKING, Dict
|
| 19 |
+
|
| 20 |
+
from ._base import MAX_SHARD_SIZE, StateDictSplit, split_state_dict_into_shards_factory
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
if TYPE_CHECKING:
|
| 24 |
+
import tensorflow as tf
|
| 25 |
+
|
| 26 |
+
|
| 27 |
+
def split_tf_state_dict_into_shards(
|
| 28 |
+
state_dict: Dict[str, "tf.Tensor"],
|
| 29 |
+
*,
|
| 30 |
+
filename_pattern: str = "tf_model{suffix}.h5",
|
| 31 |
+
max_shard_size: int = MAX_SHARD_SIZE,
|
| 32 |
+
) -> StateDictSplit:
|
| 33 |
+
"""
|
| 34 |
+
Split a model state dictionary in shards so that each shard is smaller than a given size.
|
| 35 |
+
|
| 36 |
+
The shards are determined by iterating through the `state_dict` in the order of its keys. There is no optimization
|
| 37 |
+
made to make each shard as close as possible to the maximum size passed. For example, if the limit is 10GB and we
|
| 38 |
+
have tensors of sizes [6GB, 6GB, 2GB, 6GB, 2GB, 2GB] they will get sharded as [6GB], [6+2GB], [6+2+2GB] and not
|
| 39 |
+
[6+2+2GB], [6+2GB], [6GB].
|
| 40 |
+
|
| 41 |
+
<Tip warning={true}>
|
| 42 |
+
|
| 43 |
+
If one of the model's tensor is bigger than `max_shard_size`, it will end up in its own shard which will have a
|
| 44 |
+
size greater than `max_shard_size`.
|
| 45 |
+
|
| 46 |
+
</Tip>
|
| 47 |
+
|
| 48 |
+
Args:
|
| 49 |
+
state_dict (`Dict[str, Tensor]`):
|
| 50 |
+
The state dictionary to save.
|
| 51 |
+
filename_pattern (`str`, *optional*):
|
| 52 |
+
The pattern to generate the files names in which the model will be saved. Pattern must be a string that
|
| 53 |
+
can be formatted with `filename_pattern.format(suffix=...)` and must contain the keyword `suffix`
|
| 54 |
+
Defaults to `"tf_model{suffix}.h5"`.
|
| 55 |
+
max_shard_size (`int` or `str`, *optional*):
|
| 56 |
+
The maximum size of each shard, in bytes. Defaults to 5GB.
|
| 57 |
+
|
| 58 |
+
Returns:
|
| 59 |
+
[`StateDictSplit`]: A `StateDictSplit` object containing the shards and the index to retrieve them.
|
| 60 |
+
"""
|
| 61 |
+
return split_state_dict_into_shards_factory(
|
| 62 |
+
state_dict,
|
| 63 |
+
max_shard_size=max_shard_size,
|
| 64 |
+
filename_pattern=filename_pattern,
|
| 65 |
+
get_tensor_size=get_tensor_size,
|
| 66 |
+
)
|
| 67 |
+
|
| 68 |
+
|
| 69 |
+
def get_tensor_size(tensor: "tf.Tensor") -> int:
|
| 70 |
+
# Return `math.ceil` since dtype byte size can be a float (e.g., 0.125 for tf.bool).
|
| 71 |
+
# Better to overestimate than underestimate.
|
| 72 |
+
return math.ceil(tensor.numpy().size * _dtype_byte_size_tf(tensor.dtype))
|
| 73 |
+
|
| 74 |
+
|
| 75 |
+
def _dtype_byte_size_tf(dtype) -> float:
|
| 76 |
+
"""
|
| 77 |
+
Returns the size (in bytes) occupied by one parameter of type `dtype`.
|
| 78 |
+
Taken from https://github.com/huggingface/transformers/blob/74d9d0cebb0263a3f8ab9c280569170cc74651d0/src/transformers/modeling_tf_utils.py#L608.
|
| 79 |
+
NOTE: why not `tensor.numpy().nbytes`?
|
| 80 |
+
Example:
|
| 81 |
+
```py
|
| 82 |
+
>>> _dtype_byte_size(tf.float32)
|
| 83 |
+
4
|
| 84 |
+
```
|
| 85 |
+
"""
|
| 86 |
+
import tensorflow as tf
|
| 87 |
+
|
| 88 |
+
if dtype == tf.bool:
|
| 89 |
+
return 1 / 8
|
| 90 |
+
bit_search = re.search(r"[^\d](\d+)$", dtype.name)
|
| 91 |
+
if bit_search is None:
|
| 92 |
+
raise ValueError(f"`dtype` is not a valid dtype: {dtype}.")
|
| 93 |
+
bit_size = int(bit_search.groups()[0])
|
| 94 |
+
return bit_size // 8
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/serialization/_torch.py
ADDED
|
@@ -0,0 +1,200 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Copyright 2024 The HuggingFace Team. All rights reserved.
|
| 2 |
+
#
|
| 3 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 4 |
+
# you may not use this file except in compliance with the License.
|
| 5 |
+
# You may obtain a copy of the License at
|
| 6 |
+
#
|
| 7 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 8 |
+
#
|
| 9 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 10 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 11 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 12 |
+
# See the License for the specific language governing permissions and
|
| 13 |
+
# limitations under the License.
|
| 14 |
+
"""Contains pytorch-specific helpers."""
|
| 15 |
+
|
| 16 |
+
import importlib
|
| 17 |
+
from functools import lru_cache
|
| 18 |
+
from typing import TYPE_CHECKING, Dict, Tuple
|
| 19 |
+
|
| 20 |
+
from ._base import FILENAME_PATTERN, MAX_SHARD_SIZE, StateDictSplit, split_state_dict_into_shards_factory
|
| 21 |
+
|
| 22 |
+
|
| 23 |
+
if TYPE_CHECKING:
|
| 24 |
+
import torch
|
| 25 |
+
|
| 26 |
+
|
| 27 |
+
def split_torch_state_dict_into_shards(
|
| 28 |
+
state_dict: Dict[str, "torch.Tensor"],
|
| 29 |
+
*,
|
| 30 |
+
filename_pattern: str = FILENAME_PATTERN,
|
| 31 |
+
max_shard_size: int = MAX_SHARD_SIZE,
|
| 32 |
+
) -> StateDictSplit:
|
| 33 |
+
"""
|
| 34 |
+
Split a model state dictionary in shards so that each shard is smaller than a given size.
|
| 35 |
+
|
| 36 |
+
The shards are determined by iterating through the `state_dict` in the order of its keys. There is no optimization
|
| 37 |
+
made to make each shard as close as possible to the maximum size passed. For example, if the limit is 10GB and we
|
| 38 |
+
have tensors of sizes [6GB, 6GB, 2GB, 6GB, 2GB, 2GB] they will get sharded as [6GB], [6+2GB], [6+2+2GB] and not
|
| 39 |
+
[6+2+2GB], [6+2GB], [6GB].
|
| 40 |
+
|
| 41 |
+
<Tip warning={true}>
|
| 42 |
+
|
| 43 |
+
If one of the model's tensor is bigger than `max_shard_size`, it will end up in its own shard which will have a
|
| 44 |
+
size greater than `max_shard_size`.
|
| 45 |
+
|
| 46 |
+
</Tip>
|
| 47 |
+
|
| 48 |
+
Args:
|
| 49 |
+
state_dict (`Dict[str, torch.Tensor]`):
|
| 50 |
+
The state dictionary to save.
|
| 51 |
+
filename_pattern (`str`, *optional*):
|
| 52 |
+
The pattern to generate the files names in which the model will be saved. Pattern must be a string that
|
| 53 |
+
can be formatted with `filename_pattern.format(suffix=...)` and must contain the keyword `suffix`
|
| 54 |
+
Defaults to `"model{suffix}.safetensors"`.
|
| 55 |
+
max_shard_size (`int` or `str`, *optional*):
|
| 56 |
+
The maximum size of each shard, in bytes. Defaults to 5GB.
|
| 57 |
+
|
| 58 |
+
Returns:
|
| 59 |
+
[`StateDictSplit`]: A `StateDictSplit` object containing the shards and the index to retrieve them.
|
| 60 |
+
|
| 61 |
+
Example:
|
| 62 |
+
```py
|
| 63 |
+
>>> import json
|
| 64 |
+
>>> import os
|
| 65 |
+
>>> from safetensors.torch import save_file as safe_save_file
|
| 66 |
+
>>> from huggingface_hub import split_torch_state_dict_into_shards
|
| 67 |
+
|
| 68 |
+
>>> def save_state_dict(state_dict: Dict[str, torch.Tensor], save_directory: str):
|
| 69 |
+
... state_dict_split = split_torch_state_dict_into_shards(state_dict)
|
| 70 |
+
... for filename, tensors in state_dict_split.filename_to_tensors.values():
|
| 71 |
+
... shard = {tensor: state_dict[tensor] for tensor in tensors}
|
| 72 |
+
... safe_save_file(
|
| 73 |
+
... shard,
|
| 74 |
+
... os.path.join(save_directory, filename),
|
| 75 |
+
... metadata={"format": "pt"},
|
| 76 |
+
... )
|
| 77 |
+
... if state_dict_split.is_sharded:
|
| 78 |
+
... index = {
|
| 79 |
+
... "metadata": state_dict_split.metadata,
|
| 80 |
+
... "weight_map": state_dict_split.tensor_to_filename,
|
| 81 |
+
... }
|
| 82 |
+
... with open(os.path.join(save_directory, "model.safetensors.index.json"), "w") as f:
|
| 83 |
+
... f.write(json.dumps(index, indent=2))
|
| 84 |
+
```
|
| 85 |
+
"""
|
| 86 |
+
return split_state_dict_into_shards_factory(
|
| 87 |
+
state_dict,
|
| 88 |
+
max_shard_size=max_shard_size,
|
| 89 |
+
filename_pattern=filename_pattern,
|
| 90 |
+
get_tensor_size=get_tensor_size,
|
| 91 |
+
get_storage_id=get_storage_id,
|
| 92 |
+
)
|
| 93 |
+
|
| 94 |
+
|
| 95 |
+
def get_storage_id(tensor: "torch.Tensor") -> Tuple["torch.device", int, int]:
|
| 96 |
+
"""
|
| 97 |
+
Return unique identifier to a tensor storage.
|
| 98 |
+
|
| 99 |
+
Multiple different tensors can share the same underlying storage. For
|
| 100 |
+
example, "meta" tensors all share the same storage, and thus their identifier will all be equal. This identifier is
|
| 101 |
+
guaranteed to be unique and constant for this tensor's storage during its lifetime. Two tensor storages with
|
| 102 |
+
non-overlapping lifetimes may have the same id.
|
| 103 |
+
|
| 104 |
+
Taken from https://github.com/huggingface/transformers/blob/1ecf5f7c982d761b4daaa96719d162c324187c64/src/transformers/pytorch_utils.py#L278.
|
| 105 |
+
"""
|
| 106 |
+
if tensor.device.type == "xla" and is_torch_tpu_available():
|
| 107 |
+
# NOTE: xla tensors dont have storage
|
| 108 |
+
# use some other unique id to distinguish.
|
| 109 |
+
# this is a XLA tensor, it must be created using torch_xla's
|
| 110 |
+
# device. So the following import is safe:
|
| 111 |
+
import torch_xla
|
| 112 |
+
|
| 113 |
+
unique_id = torch_xla._XLAC._xla_get_tensor_id(tensor)
|
| 114 |
+
else:
|
| 115 |
+
unique_id = storage_ptr(tensor)
|
| 116 |
+
|
| 117 |
+
return tensor.device, unique_id, get_storage_size(tensor)
|
| 118 |
+
|
| 119 |
+
|
| 120 |
+
def get_tensor_size(tensor: "torch.Tensor") -> int:
|
| 121 |
+
return tensor.numel() * tensor.element_size()
|
| 122 |
+
|
| 123 |
+
|
| 124 |
+
@lru_cache()
|
| 125 |
+
def is_torch_tpu_available(check_device=True):
|
| 126 |
+
"""
|
| 127 |
+
Checks if `torch_xla` is installed and potentially if a TPU is in the environment
|
| 128 |
+
|
| 129 |
+
Taken from https://github.com/huggingface/transformers/blob/1ecf5f7c982d761b4daaa96719d162c324187c64/src/transformers/utils/import_utils.py#L463.
|
| 130 |
+
"""
|
| 131 |
+
if importlib.util.find_spec("torch_xla") is not None:
|
| 132 |
+
if check_device:
|
| 133 |
+
# We need to check if `xla_device` can be found, will raise a RuntimeError if not
|
| 134 |
+
try:
|
| 135 |
+
import torch_xla.core.xla_model as xm
|
| 136 |
+
|
| 137 |
+
_ = xm.xla_device()
|
| 138 |
+
return True
|
| 139 |
+
except RuntimeError:
|
| 140 |
+
return False
|
| 141 |
+
return True
|
| 142 |
+
return False
|
| 143 |
+
|
| 144 |
+
|
| 145 |
+
def storage_ptr(tensor: "torch.Tensor") -> int:
|
| 146 |
+
"""
|
| 147 |
+
Taken from https://github.com/huggingface/safetensors/blob/08db34094e9e59e2f9218f2df133b7b4aaff5a99/bindings/python/py_src/safetensors/torch.py#L11C1-L20C21.
|
| 148 |
+
"""
|
| 149 |
+
try:
|
| 150 |
+
return tensor.untyped_storage().data_ptr()
|
| 151 |
+
except Exception:
|
| 152 |
+
# Fallback for torch==1.10
|
| 153 |
+
try:
|
| 154 |
+
return tensor.storage().data_ptr()
|
| 155 |
+
except NotImplementedError:
|
| 156 |
+
# Fallback for meta storage
|
| 157 |
+
return 0
|
| 158 |
+
|
| 159 |
+
|
| 160 |
+
def get_storage_size(tensor: "torch.Tensor") -> int:
|
| 161 |
+
"""
|
| 162 |
+
Taken from https://github.com/huggingface/safetensors/blob/08db34094e9e59e2f9218f2df133b7b4aaff5a99/bindings/python/py_src/safetensors/torch.py#L31C1-L41C59
|
| 163 |
+
"""
|
| 164 |
+
try:
|
| 165 |
+
return tensor.untyped_storage().nbytes()
|
| 166 |
+
except AttributeError:
|
| 167 |
+
# Fallback for torch==1.10
|
| 168 |
+
try:
|
| 169 |
+
return tensor.storage().size() * _get_dtype_size(tensor.dtype)
|
| 170 |
+
except NotImplementedError:
|
| 171 |
+
# Fallback for meta storage
|
| 172 |
+
# On torch >=2.0 this is the tensor size
|
| 173 |
+
return tensor.nelement() * _get_dtype_size(tensor.dtype)
|
| 174 |
+
|
| 175 |
+
|
| 176 |
+
@lru_cache()
|
| 177 |
+
def _get_dtype_size(dtype: "torch.dtype") -> int:
|
| 178 |
+
"""
|
| 179 |
+
Taken from https://github.com/huggingface/safetensors/blob/08db34094e9e59e2f9218f2df133b7b4aaff5a99/bindings/python/py_src/safetensors/torch.py#L344
|
| 180 |
+
"""
|
| 181 |
+
import torch
|
| 182 |
+
|
| 183 |
+
# torch.float8 formats require 2.1; we do not support these dtypes on earlier versions
|
| 184 |
+
_float8_e4m3fn = getattr(torch, "float8_e4m3fn", None)
|
| 185 |
+
_float8_e5m2 = getattr(torch, "float8_e5m2", None)
|
| 186 |
+
_SIZE = {
|
| 187 |
+
torch.int64: 8,
|
| 188 |
+
torch.float32: 4,
|
| 189 |
+
torch.int32: 4,
|
| 190 |
+
torch.bfloat16: 2,
|
| 191 |
+
torch.float16: 2,
|
| 192 |
+
torch.int16: 2,
|
| 193 |
+
torch.uint8: 1,
|
| 194 |
+
torch.int8: 1,
|
| 195 |
+
torch.bool: 1,
|
| 196 |
+
torch.float64: 8,
|
| 197 |
+
_float8_e4m3fn: 1,
|
| 198 |
+
_float8_e5m2: 1,
|
| 199 |
+
}
|
| 200 |
+
return _SIZE[dtype]
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_cache_manager.cpython-310.pyc
ADDED
|
Binary file (25 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_experimental.cpython-310.pyc
ADDED
|
Binary file (1.93 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_hf_folder.cpython-310.pyc
ADDED
|
Binary file (2.72 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_paths.cpython-310.pyc
ADDED
|
Binary file (3.87 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_runtime.cpython-310.pyc
ADDED
|
Binary file (9.31 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_safetensors.cpython-310.pyc
ADDED
|
Binary file (5.6 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_subprocess.cpython-310.pyc
ADDED
|
Binary file (3.93 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_telemetry.cpython-310.pyc
ADDED
|
Binary file (4.4 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/_typing.cpython-310.pyc
ADDED
|
Binary file (1.72 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/sha.cpython-310.pyc
ADDED
|
Binary file (1.09 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/huggingface_hub/utils/__pycache__/tqdm.cpython-310.pyc
ADDED
|
Binary file (5.61 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/__about__.py
ADDED
|
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# This file is dual licensed under the terms of the Apache License, Version
|
| 2 |
+
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
|
| 3 |
+
# for complete details.
|
| 4 |
+
|
| 5 |
+
__all__ = [
|
| 6 |
+
"__title__",
|
| 7 |
+
"__summary__",
|
| 8 |
+
"__uri__",
|
| 9 |
+
"__version__",
|
| 10 |
+
"__author__",
|
| 11 |
+
"__email__",
|
| 12 |
+
"__license__",
|
| 13 |
+
"__copyright__",
|
| 14 |
+
]
|
| 15 |
+
|
| 16 |
+
__title__ = "packaging"
|
| 17 |
+
__summary__ = "Core utilities for Python packages"
|
| 18 |
+
__uri__ = "https://github.com/pypa/packaging"
|
| 19 |
+
|
| 20 |
+
__version__ = "21.2"
|
| 21 |
+
|
| 22 |
+
__author__ = "Donald Stufft and individual contributors"
|
| 23 |
+
__email__ = "[email protected]"
|
| 24 |
+
|
| 25 |
+
__license__ = "BSD-2-Clause or Apache-2.0"
|
| 26 |
+
__copyright__ = "2014-2019 %s" % __author__
|
env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/__init__.py
ADDED
|
@@ -0,0 +1,25 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# This file is dual licensed under the terms of the Apache License, Version
|
| 2 |
+
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
|
| 3 |
+
# for complete details.
|
| 4 |
+
|
| 5 |
+
from .__about__ import (
|
| 6 |
+
__author__,
|
| 7 |
+
__copyright__,
|
| 8 |
+
__email__,
|
| 9 |
+
__license__,
|
| 10 |
+
__summary__,
|
| 11 |
+
__title__,
|
| 12 |
+
__uri__,
|
| 13 |
+
__version__,
|
| 14 |
+
)
|
| 15 |
+
|
| 16 |
+
__all__ = [
|
| 17 |
+
"__title__",
|
| 18 |
+
"__summary__",
|
| 19 |
+
"__uri__",
|
| 20 |
+
"__version__",
|
| 21 |
+
"__author__",
|
| 22 |
+
"__email__",
|
| 23 |
+
"__license__",
|
| 24 |
+
"__copyright__",
|
| 25 |
+
]
|
env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/__pycache__/_structures.cpython-310.pyc
ADDED
|
Binary file (2.97 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/__pycache__/requirements.cpython-310.pyc
ADDED
|
Binary file (3.99 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/__pycache__/specifiers.cpython-310.pyc
ADDED
|
Binary file (22.2 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/_manylinux.py
ADDED
|
@@ -0,0 +1,301 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import collections
|
| 2 |
+
import functools
|
| 3 |
+
import os
|
| 4 |
+
import re
|
| 5 |
+
import struct
|
| 6 |
+
import sys
|
| 7 |
+
import warnings
|
| 8 |
+
from typing import IO, Dict, Iterator, NamedTuple, Optional, Tuple
|
| 9 |
+
|
| 10 |
+
|
| 11 |
+
# Python does not provide platform information at sufficient granularity to
|
| 12 |
+
# identify the architecture of the running executable in some cases, so we
|
| 13 |
+
# determine it dynamically by reading the information from the running
|
| 14 |
+
# process. This only applies on Linux, which uses the ELF format.
|
| 15 |
+
class _ELFFileHeader:
|
| 16 |
+
# https://en.wikipedia.org/wiki/Executable_and_Linkable_Format#File_header
|
| 17 |
+
class _InvalidELFFileHeader(ValueError):
|
| 18 |
+
"""
|
| 19 |
+
An invalid ELF file header was found.
|
| 20 |
+
"""
|
| 21 |
+
|
| 22 |
+
ELF_MAGIC_NUMBER = 0x7F454C46
|
| 23 |
+
ELFCLASS32 = 1
|
| 24 |
+
ELFCLASS64 = 2
|
| 25 |
+
ELFDATA2LSB = 1
|
| 26 |
+
ELFDATA2MSB = 2
|
| 27 |
+
EM_386 = 3
|
| 28 |
+
EM_S390 = 22
|
| 29 |
+
EM_ARM = 40
|
| 30 |
+
EM_X86_64 = 62
|
| 31 |
+
EF_ARM_ABIMASK = 0xFF000000
|
| 32 |
+
EF_ARM_ABI_VER5 = 0x05000000
|
| 33 |
+
EF_ARM_ABI_FLOAT_HARD = 0x00000400
|
| 34 |
+
|
| 35 |
+
def __init__(self, file: IO[bytes]) -> None:
|
| 36 |
+
def unpack(fmt: str) -> int:
|
| 37 |
+
try:
|
| 38 |
+
data = file.read(struct.calcsize(fmt))
|
| 39 |
+
result: Tuple[int, ...] = struct.unpack(fmt, data)
|
| 40 |
+
except struct.error:
|
| 41 |
+
raise _ELFFileHeader._InvalidELFFileHeader()
|
| 42 |
+
return result[0]
|
| 43 |
+
|
| 44 |
+
self.e_ident_magic = unpack(">I")
|
| 45 |
+
if self.e_ident_magic != self.ELF_MAGIC_NUMBER:
|
| 46 |
+
raise _ELFFileHeader._InvalidELFFileHeader()
|
| 47 |
+
self.e_ident_class = unpack("B")
|
| 48 |
+
if self.e_ident_class not in {self.ELFCLASS32, self.ELFCLASS64}:
|
| 49 |
+
raise _ELFFileHeader._InvalidELFFileHeader()
|
| 50 |
+
self.e_ident_data = unpack("B")
|
| 51 |
+
if self.e_ident_data not in {self.ELFDATA2LSB, self.ELFDATA2MSB}:
|
| 52 |
+
raise _ELFFileHeader._InvalidELFFileHeader()
|
| 53 |
+
self.e_ident_version = unpack("B")
|
| 54 |
+
self.e_ident_osabi = unpack("B")
|
| 55 |
+
self.e_ident_abiversion = unpack("B")
|
| 56 |
+
self.e_ident_pad = file.read(7)
|
| 57 |
+
format_h = "<H" if self.e_ident_data == self.ELFDATA2LSB else ">H"
|
| 58 |
+
format_i = "<I" if self.e_ident_data == self.ELFDATA2LSB else ">I"
|
| 59 |
+
format_q = "<Q" if self.e_ident_data == self.ELFDATA2LSB else ">Q"
|
| 60 |
+
format_p = format_i if self.e_ident_class == self.ELFCLASS32 else format_q
|
| 61 |
+
self.e_type = unpack(format_h)
|
| 62 |
+
self.e_machine = unpack(format_h)
|
| 63 |
+
self.e_version = unpack(format_i)
|
| 64 |
+
self.e_entry = unpack(format_p)
|
| 65 |
+
self.e_phoff = unpack(format_p)
|
| 66 |
+
self.e_shoff = unpack(format_p)
|
| 67 |
+
self.e_flags = unpack(format_i)
|
| 68 |
+
self.e_ehsize = unpack(format_h)
|
| 69 |
+
self.e_phentsize = unpack(format_h)
|
| 70 |
+
self.e_phnum = unpack(format_h)
|
| 71 |
+
self.e_shentsize = unpack(format_h)
|
| 72 |
+
self.e_shnum = unpack(format_h)
|
| 73 |
+
self.e_shstrndx = unpack(format_h)
|
| 74 |
+
|
| 75 |
+
|
| 76 |
+
def _get_elf_header() -> Optional[_ELFFileHeader]:
|
| 77 |
+
try:
|
| 78 |
+
with open(sys.executable, "rb") as f:
|
| 79 |
+
elf_header = _ELFFileHeader(f)
|
| 80 |
+
except (OSError, TypeError, _ELFFileHeader._InvalidELFFileHeader):
|
| 81 |
+
return None
|
| 82 |
+
return elf_header
|
| 83 |
+
|
| 84 |
+
|
| 85 |
+
def _is_linux_armhf() -> bool:
|
| 86 |
+
# hard-float ABI can be detected from the ELF header of the running
|
| 87 |
+
# process
|
| 88 |
+
# https://static.docs.arm.com/ihi0044/g/aaelf32.pdf
|
| 89 |
+
elf_header = _get_elf_header()
|
| 90 |
+
if elf_header is None:
|
| 91 |
+
return False
|
| 92 |
+
result = elf_header.e_ident_class == elf_header.ELFCLASS32
|
| 93 |
+
result &= elf_header.e_ident_data == elf_header.ELFDATA2LSB
|
| 94 |
+
result &= elf_header.e_machine == elf_header.EM_ARM
|
| 95 |
+
result &= (
|
| 96 |
+
elf_header.e_flags & elf_header.EF_ARM_ABIMASK
|
| 97 |
+
) == elf_header.EF_ARM_ABI_VER5
|
| 98 |
+
result &= (
|
| 99 |
+
elf_header.e_flags & elf_header.EF_ARM_ABI_FLOAT_HARD
|
| 100 |
+
) == elf_header.EF_ARM_ABI_FLOAT_HARD
|
| 101 |
+
return result
|
| 102 |
+
|
| 103 |
+
|
| 104 |
+
def _is_linux_i686() -> bool:
|
| 105 |
+
elf_header = _get_elf_header()
|
| 106 |
+
if elf_header is None:
|
| 107 |
+
return False
|
| 108 |
+
result = elf_header.e_ident_class == elf_header.ELFCLASS32
|
| 109 |
+
result &= elf_header.e_ident_data == elf_header.ELFDATA2LSB
|
| 110 |
+
result &= elf_header.e_machine == elf_header.EM_386
|
| 111 |
+
return result
|
| 112 |
+
|
| 113 |
+
|
| 114 |
+
def _have_compatible_abi(arch: str) -> bool:
|
| 115 |
+
if arch == "armv7l":
|
| 116 |
+
return _is_linux_armhf()
|
| 117 |
+
if arch == "i686":
|
| 118 |
+
return _is_linux_i686()
|
| 119 |
+
return arch in {"x86_64", "aarch64", "ppc64", "ppc64le", "s390x"}
|
| 120 |
+
|
| 121 |
+
|
| 122 |
+
# If glibc ever changes its major version, we need to know what the last
|
| 123 |
+
# minor version was, so we can build the complete list of all versions.
|
| 124 |
+
# For now, guess what the highest minor version might be, assume it will
|
| 125 |
+
# be 50 for testing. Once this actually happens, update the dictionary
|
| 126 |
+
# with the actual value.
|
| 127 |
+
_LAST_GLIBC_MINOR: Dict[int, int] = collections.defaultdict(lambda: 50)
|
| 128 |
+
|
| 129 |
+
|
| 130 |
+
class _GLibCVersion(NamedTuple):
|
| 131 |
+
major: int
|
| 132 |
+
minor: int
|
| 133 |
+
|
| 134 |
+
|
| 135 |
+
def _glibc_version_string_confstr() -> Optional[str]:
|
| 136 |
+
"""
|
| 137 |
+
Primary implementation of glibc_version_string using os.confstr.
|
| 138 |
+
"""
|
| 139 |
+
# os.confstr is quite a bit faster than ctypes.DLL. It's also less likely
|
| 140 |
+
# to be broken or missing. This strategy is used in the standard library
|
| 141 |
+
# platform module.
|
| 142 |
+
# https://github.com/python/cpython/blob/fcf1d003bf4f0100c/Lib/platform.py#L175-L183
|
| 143 |
+
try:
|
| 144 |
+
# os.confstr("CS_GNU_LIBC_VERSION") returns a string like "glibc 2.17".
|
| 145 |
+
version_string = os.confstr("CS_GNU_LIBC_VERSION")
|
| 146 |
+
assert version_string is not None
|
| 147 |
+
_, version = version_string.split()
|
| 148 |
+
except (AssertionError, AttributeError, OSError, ValueError):
|
| 149 |
+
# os.confstr() or CS_GNU_LIBC_VERSION not available (or a bad value)...
|
| 150 |
+
return None
|
| 151 |
+
return version
|
| 152 |
+
|
| 153 |
+
|
| 154 |
+
def _glibc_version_string_ctypes() -> Optional[str]:
|
| 155 |
+
"""
|
| 156 |
+
Fallback implementation of glibc_version_string using ctypes.
|
| 157 |
+
"""
|
| 158 |
+
try:
|
| 159 |
+
import ctypes
|
| 160 |
+
except ImportError:
|
| 161 |
+
return None
|
| 162 |
+
|
| 163 |
+
# ctypes.CDLL(None) internally calls dlopen(NULL), and as the dlopen
|
| 164 |
+
# manpage says, "If filename is NULL, then the returned handle is for the
|
| 165 |
+
# main program". This way we can let the linker do the work to figure out
|
| 166 |
+
# which libc our process is actually using.
|
| 167 |
+
#
|
| 168 |
+
# We must also handle the special case where the executable is not a
|
| 169 |
+
# dynamically linked executable. This can occur when using musl libc,
|
| 170 |
+
# for example. In this situation, dlopen() will error, leading to an
|
| 171 |
+
# OSError. Interestingly, at least in the case of musl, there is no
|
| 172 |
+
# errno set on the OSError. The single string argument used to construct
|
| 173 |
+
# OSError comes from libc itself and is therefore not portable to
|
| 174 |
+
# hard code here. In any case, failure to call dlopen() means we
|
| 175 |
+
# can proceed, so we bail on our attempt.
|
| 176 |
+
try:
|
| 177 |
+
process_namespace = ctypes.CDLL(None)
|
| 178 |
+
except OSError:
|
| 179 |
+
return None
|
| 180 |
+
|
| 181 |
+
try:
|
| 182 |
+
gnu_get_libc_version = process_namespace.gnu_get_libc_version
|
| 183 |
+
except AttributeError:
|
| 184 |
+
# Symbol doesn't exist -> therefore, we are not linked to
|
| 185 |
+
# glibc.
|
| 186 |
+
return None
|
| 187 |
+
|
| 188 |
+
# Call gnu_get_libc_version, which returns a string like "2.5"
|
| 189 |
+
gnu_get_libc_version.restype = ctypes.c_char_p
|
| 190 |
+
version_str: str = gnu_get_libc_version()
|
| 191 |
+
# py2 / py3 compatibility:
|
| 192 |
+
if not isinstance(version_str, str):
|
| 193 |
+
version_str = version_str.decode("ascii")
|
| 194 |
+
|
| 195 |
+
return version_str
|
| 196 |
+
|
| 197 |
+
|
| 198 |
+
def _glibc_version_string() -> Optional[str]:
|
| 199 |
+
"""Returns glibc version string, or None if not using glibc."""
|
| 200 |
+
return _glibc_version_string_confstr() or _glibc_version_string_ctypes()
|
| 201 |
+
|
| 202 |
+
|
| 203 |
+
def _parse_glibc_version(version_str: str) -> Tuple[int, int]:
|
| 204 |
+
"""Parse glibc version.
|
| 205 |
+
|
| 206 |
+
We use a regexp instead of str.split because we want to discard any
|
| 207 |
+
random junk that might come after the minor version -- this might happen
|
| 208 |
+
in patched/forked versions of glibc (e.g. Linaro's version of glibc
|
| 209 |
+
uses version strings like "2.20-2014.11"). See gh-3588.
|
| 210 |
+
"""
|
| 211 |
+
m = re.match(r"(?P<major>[0-9]+)\.(?P<minor>[0-9]+)", version_str)
|
| 212 |
+
if not m:
|
| 213 |
+
warnings.warn(
|
| 214 |
+
"Expected glibc version with 2 components major.minor,"
|
| 215 |
+
" got: %s" % version_str,
|
| 216 |
+
RuntimeWarning,
|
| 217 |
+
)
|
| 218 |
+
return -1, -1
|
| 219 |
+
return int(m.group("major")), int(m.group("minor"))
|
| 220 |
+
|
| 221 |
+
|
| 222 |
+
@functools.lru_cache()
|
| 223 |
+
def _get_glibc_version() -> Tuple[int, int]:
|
| 224 |
+
version_str = _glibc_version_string()
|
| 225 |
+
if version_str is None:
|
| 226 |
+
return (-1, -1)
|
| 227 |
+
return _parse_glibc_version(version_str)
|
| 228 |
+
|
| 229 |
+
|
| 230 |
+
# From PEP 513, PEP 600
|
| 231 |
+
def _is_compatible(name: str, arch: str, version: _GLibCVersion) -> bool:
|
| 232 |
+
sys_glibc = _get_glibc_version()
|
| 233 |
+
if sys_glibc < version:
|
| 234 |
+
return False
|
| 235 |
+
# Check for presence of _manylinux module.
|
| 236 |
+
try:
|
| 237 |
+
import _manylinux # noqa
|
| 238 |
+
except ImportError:
|
| 239 |
+
return True
|
| 240 |
+
if hasattr(_manylinux, "manylinux_compatible"):
|
| 241 |
+
result = _manylinux.manylinux_compatible(version[0], version[1], arch)
|
| 242 |
+
if result is not None:
|
| 243 |
+
return bool(result)
|
| 244 |
+
return True
|
| 245 |
+
if version == _GLibCVersion(2, 5):
|
| 246 |
+
if hasattr(_manylinux, "manylinux1_compatible"):
|
| 247 |
+
return bool(_manylinux.manylinux1_compatible)
|
| 248 |
+
if version == _GLibCVersion(2, 12):
|
| 249 |
+
if hasattr(_manylinux, "manylinux2010_compatible"):
|
| 250 |
+
return bool(_manylinux.manylinux2010_compatible)
|
| 251 |
+
if version == _GLibCVersion(2, 17):
|
| 252 |
+
if hasattr(_manylinux, "manylinux2014_compatible"):
|
| 253 |
+
return bool(_manylinux.manylinux2014_compatible)
|
| 254 |
+
return True
|
| 255 |
+
|
| 256 |
+
|
| 257 |
+
_LEGACY_MANYLINUX_MAP = {
|
| 258 |
+
# CentOS 7 w/ glibc 2.17 (PEP 599)
|
| 259 |
+
(2, 17): "manylinux2014",
|
| 260 |
+
# CentOS 6 w/ glibc 2.12 (PEP 571)
|
| 261 |
+
(2, 12): "manylinux2010",
|
| 262 |
+
# CentOS 5 w/ glibc 2.5 (PEP 513)
|
| 263 |
+
(2, 5): "manylinux1",
|
| 264 |
+
}
|
| 265 |
+
|
| 266 |
+
|
| 267 |
+
def platform_tags(linux: str, arch: str) -> Iterator[str]:
|
| 268 |
+
if not _have_compatible_abi(arch):
|
| 269 |
+
return
|
| 270 |
+
# Oldest glibc to be supported regardless of architecture is (2, 17).
|
| 271 |
+
too_old_glibc2 = _GLibCVersion(2, 16)
|
| 272 |
+
if arch in {"x86_64", "i686"}:
|
| 273 |
+
# On x86/i686 also oldest glibc to be supported is (2, 5).
|
| 274 |
+
too_old_glibc2 = _GLibCVersion(2, 4)
|
| 275 |
+
current_glibc = _GLibCVersion(*_get_glibc_version())
|
| 276 |
+
glibc_max_list = [current_glibc]
|
| 277 |
+
# We can assume compatibility across glibc major versions.
|
| 278 |
+
# https://sourceware.org/bugzilla/show_bug.cgi?id=24636
|
| 279 |
+
#
|
| 280 |
+
# Build a list of maximum glibc versions so that we can
|
| 281 |
+
# output the canonical list of all glibc from current_glibc
|
| 282 |
+
# down to too_old_glibc2, including all intermediary versions.
|
| 283 |
+
for glibc_major in range(current_glibc.major - 1, 1, -1):
|
| 284 |
+
glibc_minor = _LAST_GLIBC_MINOR[glibc_major]
|
| 285 |
+
glibc_max_list.append(_GLibCVersion(glibc_major, glibc_minor))
|
| 286 |
+
for glibc_max in glibc_max_list:
|
| 287 |
+
if glibc_max.major == too_old_glibc2.major:
|
| 288 |
+
min_minor = too_old_glibc2.minor
|
| 289 |
+
else:
|
| 290 |
+
# For other glibc major versions oldest supported is (x, 0).
|
| 291 |
+
min_minor = -1
|
| 292 |
+
for glibc_minor in range(glibc_max.minor, min_minor, -1):
|
| 293 |
+
glibc_version = _GLibCVersion(glibc_max.major, glibc_minor)
|
| 294 |
+
tag = "manylinux_{}_{}".format(*glibc_version)
|
| 295 |
+
if _is_compatible(tag, arch, glibc_version):
|
| 296 |
+
yield linux.replace("linux", tag)
|
| 297 |
+
# Handle the legacy manylinux1, manylinux2010, manylinux2014 tags.
|
| 298 |
+
if glibc_version in _LEGACY_MANYLINUX_MAP:
|
| 299 |
+
legacy_tag = _LEGACY_MANYLINUX_MAP[glibc_version]
|
| 300 |
+
if _is_compatible(legacy_tag, arch, glibc_version):
|
| 301 |
+
yield linux.replace("linux", legacy_tag)
|
env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/_structures.py
ADDED
|
@@ -0,0 +1,67 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# This file is dual licensed under the terms of the Apache License, Version
|
| 2 |
+
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
|
| 3 |
+
# for complete details.
|
| 4 |
+
|
| 5 |
+
|
| 6 |
+
class InfinityType:
|
| 7 |
+
def __repr__(self) -> str:
|
| 8 |
+
return "Infinity"
|
| 9 |
+
|
| 10 |
+
def __hash__(self) -> int:
|
| 11 |
+
return hash(repr(self))
|
| 12 |
+
|
| 13 |
+
def __lt__(self, other: object) -> bool:
|
| 14 |
+
return False
|
| 15 |
+
|
| 16 |
+
def __le__(self, other: object) -> bool:
|
| 17 |
+
return False
|
| 18 |
+
|
| 19 |
+
def __eq__(self, other: object) -> bool:
|
| 20 |
+
return isinstance(other, self.__class__)
|
| 21 |
+
|
| 22 |
+
def __ne__(self, other: object) -> bool:
|
| 23 |
+
return not isinstance(other, self.__class__)
|
| 24 |
+
|
| 25 |
+
def __gt__(self, other: object) -> bool:
|
| 26 |
+
return True
|
| 27 |
+
|
| 28 |
+
def __ge__(self, other: object) -> bool:
|
| 29 |
+
return True
|
| 30 |
+
|
| 31 |
+
def __neg__(self: object) -> "NegativeInfinityType":
|
| 32 |
+
return NegativeInfinity
|
| 33 |
+
|
| 34 |
+
|
| 35 |
+
Infinity = InfinityType()
|
| 36 |
+
|
| 37 |
+
|
| 38 |
+
class NegativeInfinityType:
|
| 39 |
+
def __repr__(self) -> str:
|
| 40 |
+
return "-Infinity"
|
| 41 |
+
|
| 42 |
+
def __hash__(self) -> int:
|
| 43 |
+
return hash(repr(self))
|
| 44 |
+
|
| 45 |
+
def __lt__(self, other: object) -> bool:
|
| 46 |
+
return True
|
| 47 |
+
|
| 48 |
+
def __le__(self, other: object) -> bool:
|
| 49 |
+
return True
|
| 50 |
+
|
| 51 |
+
def __eq__(self, other: object) -> bool:
|
| 52 |
+
return isinstance(other, self.__class__)
|
| 53 |
+
|
| 54 |
+
def __ne__(self, other: object) -> bool:
|
| 55 |
+
return not isinstance(other, self.__class__)
|
| 56 |
+
|
| 57 |
+
def __gt__(self, other: object) -> bool:
|
| 58 |
+
return False
|
| 59 |
+
|
| 60 |
+
def __ge__(self, other: object) -> bool:
|
| 61 |
+
return False
|
| 62 |
+
|
| 63 |
+
def __neg__(self: object) -> InfinityType:
|
| 64 |
+
return Infinity
|
| 65 |
+
|
| 66 |
+
|
| 67 |
+
NegativeInfinity = NegativeInfinityType()
|
env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/markers.py
ADDED
|
@@ -0,0 +1,304 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# This file is dual licensed under the terms of the Apache License, Version
|
| 2 |
+
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
|
| 3 |
+
# for complete details.
|
| 4 |
+
|
| 5 |
+
import operator
|
| 6 |
+
import os
|
| 7 |
+
import platform
|
| 8 |
+
import sys
|
| 9 |
+
from typing import Any, Callable, Dict, List, Optional, Tuple, Union
|
| 10 |
+
|
| 11 |
+
from pkg_resources.extern.pyparsing import ( # noqa: N817
|
| 12 |
+
Forward,
|
| 13 |
+
Group,
|
| 14 |
+
Literal as L,
|
| 15 |
+
ParseException,
|
| 16 |
+
ParseResults,
|
| 17 |
+
QuotedString,
|
| 18 |
+
ZeroOrMore,
|
| 19 |
+
stringEnd,
|
| 20 |
+
stringStart,
|
| 21 |
+
)
|
| 22 |
+
|
| 23 |
+
from .specifiers import InvalidSpecifier, Specifier
|
| 24 |
+
|
| 25 |
+
__all__ = [
|
| 26 |
+
"InvalidMarker",
|
| 27 |
+
"UndefinedComparison",
|
| 28 |
+
"UndefinedEnvironmentName",
|
| 29 |
+
"Marker",
|
| 30 |
+
"default_environment",
|
| 31 |
+
]
|
| 32 |
+
|
| 33 |
+
Operator = Callable[[str, str], bool]
|
| 34 |
+
|
| 35 |
+
|
| 36 |
+
class InvalidMarker(ValueError):
|
| 37 |
+
"""
|
| 38 |
+
An invalid marker was found, users should refer to PEP 508.
|
| 39 |
+
"""
|
| 40 |
+
|
| 41 |
+
|
| 42 |
+
class UndefinedComparison(ValueError):
|
| 43 |
+
"""
|
| 44 |
+
An invalid operation was attempted on a value that doesn't support it.
|
| 45 |
+
"""
|
| 46 |
+
|
| 47 |
+
|
| 48 |
+
class UndefinedEnvironmentName(ValueError):
|
| 49 |
+
"""
|
| 50 |
+
A name was attempted to be used that does not exist inside of the
|
| 51 |
+
environment.
|
| 52 |
+
"""
|
| 53 |
+
|
| 54 |
+
|
| 55 |
+
class Node:
|
| 56 |
+
def __init__(self, value: Any) -> None:
|
| 57 |
+
self.value = value
|
| 58 |
+
|
| 59 |
+
def __str__(self) -> str:
|
| 60 |
+
return str(self.value)
|
| 61 |
+
|
| 62 |
+
def __repr__(self) -> str:
|
| 63 |
+
return f"<{self.__class__.__name__}('{self}')>"
|
| 64 |
+
|
| 65 |
+
def serialize(self) -> str:
|
| 66 |
+
raise NotImplementedError
|
| 67 |
+
|
| 68 |
+
|
| 69 |
+
class Variable(Node):
|
| 70 |
+
def serialize(self) -> str:
|
| 71 |
+
return str(self)
|
| 72 |
+
|
| 73 |
+
|
| 74 |
+
class Value(Node):
|
| 75 |
+
def serialize(self) -> str:
|
| 76 |
+
return f'"{self}"'
|
| 77 |
+
|
| 78 |
+
|
| 79 |
+
class Op(Node):
|
| 80 |
+
def serialize(self) -> str:
|
| 81 |
+
return str(self)
|
| 82 |
+
|
| 83 |
+
|
| 84 |
+
VARIABLE = (
|
| 85 |
+
L("implementation_version")
|
| 86 |
+
| L("platform_python_implementation")
|
| 87 |
+
| L("implementation_name")
|
| 88 |
+
| L("python_full_version")
|
| 89 |
+
| L("platform_release")
|
| 90 |
+
| L("platform_version")
|
| 91 |
+
| L("platform_machine")
|
| 92 |
+
| L("platform_system")
|
| 93 |
+
| L("python_version")
|
| 94 |
+
| L("sys_platform")
|
| 95 |
+
| L("os_name")
|
| 96 |
+
| L("os.name") # PEP-345
|
| 97 |
+
| L("sys.platform") # PEP-345
|
| 98 |
+
| L("platform.version") # PEP-345
|
| 99 |
+
| L("platform.machine") # PEP-345
|
| 100 |
+
| L("platform.python_implementation") # PEP-345
|
| 101 |
+
| L("python_implementation") # undocumented setuptools legacy
|
| 102 |
+
| L("extra") # PEP-508
|
| 103 |
+
)
|
| 104 |
+
ALIASES = {
|
| 105 |
+
"os.name": "os_name",
|
| 106 |
+
"sys.platform": "sys_platform",
|
| 107 |
+
"platform.version": "platform_version",
|
| 108 |
+
"platform.machine": "platform_machine",
|
| 109 |
+
"platform.python_implementation": "platform_python_implementation",
|
| 110 |
+
"python_implementation": "platform_python_implementation",
|
| 111 |
+
}
|
| 112 |
+
VARIABLE.setParseAction(lambda s, l, t: Variable(ALIASES.get(t[0], t[0])))
|
| 113 |
+
|
| 114 |
+
VERSION_CMP = (
|
| 115 |
+
L("===") | L("==") | L(">=") | L("<=") | L("!=") | L("~=") | L(">") | L("<")
|
| 116 |
+
)
|
| 117 |
+
|
| 118 |
+
MARKER_OP = VERSION_CMP | L("not in") | L("in")
|
| 119 |
+
MARKER_OP.setParseAction(lambda s, l, t: Op(t[0]))
|
| 120 |
+
|
| 121 |
+
MARKER_VALUE = QuotedString("'") | QuotedString('"')
|
| 122 |
+
MARKER_VALUE.setParseAction(lambda s, l, t: Value(t[0]))
|
| 123 |
+
|
| 124 |
+
BOOLOP = L("and") | L("or")
|
| 125 |
+
|
| 126 |
+
MARKER_VAR = VARIABLE | MARKER_VALUE
|
| 127 |
+
|
| 128 |
+
MARKER_ITEM = Group(MARKER_VAR + MARKER_OP + MARKER_VAR)
|
| 129 |
+
MARKER_ITEM.setParseAction(lambda s, l, t: tuple(t[0]))
|
| 130 |
+
|
| 131 |
+
LPAREN = L("(").suppress()
|
| 132 |
+
RPAREN = L(")").suppress()
|
| 133 |
+
|
| 134 |
+
MARKER_EXPR = Forward()
|
| 135 |
+
MARKER_ATOM = MARKER_ITEM | Group(LPAREN + MARKER_EXPR + RPAREN)
|
| 136 |
+
MARKER_EXPR << MARKER_ATOM + ZeroOrMore(BOOLOP + MARKER_EXPR)
|
| 137 |
+
|
| 138 |
+
MARKER = stringStart + MARKER_EXPR + stringEnd
|
| 139 |
+
|
| 140 |
+
|
| 141 |
+
def _coerce_parse_result(results: Union[ParseResults, List[Any]]) -> List[Any]:
|
| 142 |
+
if isinstance(results, ParseResults):
|
| 143 |
+
return [_coerce_parse_result(i) for i in results]
|
| 144 |
+
else:
|
| 145 |
+
return results
|
| 146 |
+
|
| 147 |
+
|
| 148 |
+
def _format_marker(
|
| 149 |
+
marker: Union[List[str], Tuple[Node, ...], str], first: Optional[bool] = True
|
| 150 |
+
) -> str:
|
| 151 |
+
|
| 152 |
+
assert isinstance(marker, (list, tuple, str))
|
| 153 |
+
|
| 154 |
+
# Sometimes we have a structure like [[...]] which is a single item list
|
| 155 |
+
# where the single item is itself it's own list. In that case we want skip
|
| 156 |
+
# the rest of this function so that we don't get extraneous () on the
|
| 157 |
+
# outside.
|
| 158 |
+
if (
|
| 159 |
+
isinstance(marker, list)
|
| 160 |
+
and len(marker) == 1
|
| 161 |
+
and isinstance(marker[0], (list, tuple))
|
| 162 |
+
):
|
| 163 |
+
return _format_marker(marker[0])
|
| 164 |
+
|
| 165 |
+
if isinstance(marker, list):
|
| 166 |
+
inner = (_format_marker(m, first=False) for m in marker)
|
| 167 |
+
if first:
|
| 168 |
+
return " ".join(inner)
|
| 169 |
+
else:
|
| 170 |
+
return "(" + " ".join(inner) + ")"
|
| 171 |
+
elif isinstance(marker, tuple):
|
| 172 |
+
return " ".join([m.serialize() for m in marker])
|
| 173 |
+
else:
|
| 174 |
+
return marker
|
| 175 |
+
|
| 176 |
+
|
| 177 |
+
_operators: Dict[str, Operator] = {
|
| 178 |
+
"in": lambda lhs, rhs: lhs in rhs,
|
| 179 |
+
"not in": lambda lhs, rhs: lhs not in rhs,
|
| 180 |
+
"<": operator.lt,
|
| 181 |
+
"<=": operator.le,
|
| 182 |
+
"==": operator.eq,
|
| 183 |
+
"!=": operator.ne,
|
| 184 |
+
">=": operator.ge,
|
| 185 |
+
">": operator.gt,
|
| 186 |
+
}
|
| 187 |
+
|
| 188 |
+
|
| 189 |
+
def _eval_op(lhs: str, op: Op, rhs: str) -> bool:
|
| 190 |
+
try:
|
| 191 |
+
spec = Specifier("".join([op.serialize(), rhs]))
|
| 192 |
+
except InvalidSpecifier:
|
| 193 |
+
pass
|
| 194 |
+
else:
|
| 195 |
+
return spec.contains(lhs)
|
| 196 |
+
|
| 197 |
+
oper: Optional[Operator] = _operators.get(op.serialize())
|
| 198 |
+
if oper is None:
|
| 199 |
+
raise UndefinedComparison(f"Undefined {op!r} on {lhs!r} and {rhs!r}.")
|
| 200 |
+
|
| 201 |
+
return oper(lhs, rhs)
|
| 202 |
+
|
| 203 |
+
|
| 204 |
+
class Undefined:
|
| 205 |
+
pass
|
| 206 |
+
|
| 207 |
+
|
| 208 |
+
_undefined = Undefined()
|
| 209 |
+
|
| 210 |
+
|
| 211 |
+
def _get_env(environment: Dict[str, str], name: str) -> str:
|
| 212 |
+
value: Union[str, Undefined] = environment.get(name, _undefined)
|
| 213 |
+
|
| 214 |
+
if isinstance(value, Undefined):
|
| 215 |
+
raise UndefinedEnvironmentName(
|
| 216 |
+
f"{name!r} does not exist in evaluation environment."
|
| 217 |
+
)
|
| 218 |
+
|
| 219 |
+
return value
|
| 220 |
+
|
| 221 |
+
|
| 222 |
+
def _evaluate_markers(markers: List[Any], environment: Dict[str, str]) -> bool:
|
| 223 |
+
groups: List[List[bool]] = [[]]
|
| 224 |
+
|
| 225 |
+
for marker in markers:
|
| 226 |
+
assert isinstance(marker, (list, tuple, str))
|
| 227 |
+
|
| 228 |
+
if isinstance(marker, list):
|
| 229 |
+
groups[-1].append(_evaluate_markers(marker, environment))
|
| 230 |
+
elif isinstance(marker, tuple):
|
| 231 |
+
lhs, op, rhs = marker
|
| 232 |
+
|
| 233 |
+
if isinstance(lhs, Variable):
|
| 234 |
+
lhs_value = _get_env(environment, lhs.value)
|
| 235 |
+
rhs_value = rhs.value
|
| 236 |
+
else:
|
| 237 |
+
lhs_value = lhs.value
|
| 238 |
+
rhs_value = _get_env(environment, rhs.value)
|
| 239 |
+
|
| 240 |
+
groups[-1].append(_eval_op(lhs_value, op, rhs_value))
|
| 241 |
+
else:
|
| 242 |
+
assert marker in ["and", "or"]
|
| 243 |
+
if marker == "or":
|
| 244 |
+
groups.append([])
|
| 245 |
+
|
| 246 |
+
return any(all(item) for item in groups)
|
| 247 |
+
|
| 248 |
+
|
| 249 |
+
def format_full_version(info: "sys._version_info") -> str:
|
| 250 |
+
version = "{0.major}.{0.minor}.{0.micro}".format(info)
|
| 251 |
+
kind = info.releaselevel
|
| 252 |
+
if kind != "final":
|
| 253 |
+
version += kind[0] + str(info.serial)
|
| 254 |
+
return version
|
| 255 |
+
|
| 256 |
+
|
| 257 |
+
def default_environment() -> Dict[str, str]:
|
| 258 |
+
iver = format_full_version(sys.implementation.version)
|
| 259 |
+
implementation_name = sys.implementation.name
|
| 260 |
+
return {
|
| 261 |
+
"implementation_name": implementation_name,
|
| 262 |
+
"implementation_version": iver,
|
| 263 |
+
"os_name": os.name,
|
| 264 |
+
"platform_machine": platform.machine(),
|
| 265 |
+
"platform_release": platform.release(),
|
| 266 |
+
"platform_system": platform.system(),
|
| 267 |
+
"platform_version": platform.version(),
|
| 268 |
+
"python_full_version": platform.python_version(),
|
| 269 |
+
"platform_python_implementation": platform.python_implementation(),
|
| 270 |
+
"python_version": ".".join(platform.python_version_tuple()[:2]),
|
| 271 |
+
"sys_platform": sys.platform,
|
| 272 |
+
}
|
| 273 |
+
|
| 274 |
+
|
| 275 |
+
class Marker:
|
| 276 |
+
def __init__(self, marker: str) -> None:
|
| 277 |
+
try:
|
| 278 |
+
self._markers = _coerce_parse_result(MARKER.parseString(marker))
|
| 279 |
+
except ParseException as e:
|
| 280 |
+
raise InvalidMarker(
|
| 281 |
+
f"Invalid marker: {marker!r}, parse error at "
|
| 282 |
+
f"{marker[e.loc : e.loc + 8]!r}"
|
| 283 |
+
)
|
| 284 |
+
|
| 285 |
+
def __str__(self) -> str:
|
| 286 |
+
return _format_marker(self._markers)
|
| 287 |
+
|
| 288 |
+
def __repr__(self) -> str:
|
| 289 |
+
return f"<Marker('{self}')>"
|
| 290 |
+
|
| 291 |
+
def evaluate(self, environment: Optional[Dict[str, str]] = None) -> bool:
|
| 292 |
+
"""Evaluate a marker.
|
| 293 |
+
|
| 294 |
+
Return the boolean from evaluating the given marker against the
|
| 295 |
+
environment. environment is an optional argument to override all or
|
| 296 |
+
part of the determined environment.
|
| 297 |
+
|
| 298 |
+
The environment is determined from the current Python process.
|
| 299 |
+
"""
|
| 300 |
+
current_environment = default_environment()
|
| 301 |
+
if environment is not None:
|
| 302 |
+
current_environment.update(environment)
|
| 303 |
+
|
| 304 |
+
return _evaluate_markers(self._markers, current_environment)
|
env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/requirements.py
ADDED
|
@@ -0,0 +1,146 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# This file is dual licensed under the terms of the Apache License, Version
|
| 2 |
+
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
|
| 3 |
+
# for complete details.
|
| 4 |
+
|
| 5 |
+
import re
|
| 6 |
+
import string
|
| 7 |
+
import urllib.parse
|
| 8 |
+
from typing import List, Optional as TOptional, Set
|
| 9 |
+
|
| 10 |
+
from pkg_resources.extern.pyparsing import ( # noqa
|
| 11 |
+
Combine,
|
| 12 |
+
Literal as L,
|
| 13 |
+
Optional,
|
| 14 |
+
ParseException,
|
| 15 |
+
Regex,
|
| 16 |
+
Word,
|
| 17 |
+
ZeroOrMore,
|
| 18 |
+
originalTextFor,
|
| 19 |
+
stringEnd,
|
| 20 |
+
stringStart,
|
| 21 |
+
)
|
| 22 |
+
|
| 23 |
+
from .markers import MARKER_EXPR, Marker
|
| 24 |
+
from .specifiers import LegacySpecifier, Specifier, SpecifierSet
|
| 25 |
+
|
| 26 |
+
|
| 27 |
+
class InvalidRequirement(ValueError):
|
| 28 |
+
"""
|
| 29 |
+
An invalid requirement was found, users should refer to PEP 508.
|
| 30 |
+
"""
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
ALPHANUM = Word(string.ascii_letters + string.digits)
|
| 34 |
+
|
| 35 |
+
LBRACKET = L("[").suppress()
|
| 36 |
+
RBRACKET = L("]").suppress()
|
| 37 |
+
LPAREN = L("(").suppress()
|
| 38 |
+
RPAREN = L(")").suppress()
|
| 39 |
+
COMMA = L(",").suppress()
|
| 40 |
+
SEMICOLON = L(";").suppress()
|
| 41 |
+
AT = L("@").suppress()
|
| 42 |
+
|
| 43 |
+
PUNCTUATION = Word("-_.")
|
| 44 |
+
IDENTIFIER_END = ALPHANUM | (ZeroOrMore(PUNCTUATION) + ALPHANUM)
|
| 45 |
+
IDENTIFIER = Combine(ALPHANUM + ZeroOrMore(IDENTIFIER_END))
|
| 46 |
+
|
| 47 |
+
NAME = IDENTIFIER("name")
|
| 48 |
+
EXTRA = IDENTIFIER
|
| 49 |
+
|
| 50 |
+
URI = Regex(r"[^ ]+")("url")
|
| 51 |
+
URL = AT + URI
|
| 52 |
+
|
| 53 |
+
EXTRAS_LIST = EXTRA + ZeroOrMore(COMMA + EXTRA)
|
| 54 |
+
EXTRAS = (LBRACKET + Optional(EXTRAS_LIST) + RBRACKET)("extras")
|
| 55 |
+
|
| 56 |
+
VERSION_PEP440 = Regex(Specifier._regex_str, re.VERBOSE | re.IGNORECASE)
|
| 57 |
+
VERSION_LEGACY = Regex(LegacySpecifier._regex_str, re.VERBOSE | re.IGNORECASE)
|
| 58 |
+
|
| 59 |
+
VERSION_ONE = VERSION_PEP440 ^ VERSION_LEGACY
|
| 60 |
+
VERSION_MANY = Combine(
|
| 61 |
+
VERSION_ONE + ZeroOrMore(COMMA + VERSION_ONE), joinString=",", adjacent=False
|
| 62 |
+
)("_raw_spec")
|
| 63 |
+
_VERSION_SPEC = Optional((LPAREN + VERSION_MANY + RPAREN) | VERSION_MANY)
|
| 64 |
+
_VERSION_SPEC.setParseAction(lambda s, l, t: t._raw_spec or "")
|
| 65 |
+
|
| 66 |
+
VERSION_SPEC = originalTextFor(_VERSION_SPEC)("specifier")
|
| 67 |
+
VERSION_SPEC.setParseAction(lambda s, l, t: t[1])
|
| 68 |
+
|
| 69 |
+
MARKER_EXPR = originalTextFor(MARKER_EXPR())("marker")
|
| 70 |
+
MARKER_EXPR.setParseAction(
|
| 71 |
+
lambda s, l, t: Marker(s[t._original_start : t._original_end])
|
| 72 |
+
)
|
| 73 |
+
MARKER_SEPARATOR = SEMICOLON
|
| 74 |
+
MARKER = MARKER_SEPARATOR + MARKER_EXPR
|
| 75 |
+
|
| 76 |
+
VERSION_AND_MARKER = VERSION_SPEC + Optional(MARKER)
|
| 77 |
+
URL_AND_MARKER = URL + Optional(MARKER)
|
| 78 |
+
|
| 79 |
+
NAMED_REQUIREMENT = NAME + Optional(EXTRAS) + (URL_AND_MARKER | VERSION_AND_MARKER)
|
| 80 |
+
|
| 81 |
+
REQUIREMENT = stringStart + NAMED_REQUIREMENT + stringEnd
|
| 82 |
+
# pkg_resources.extern.pyparsing isn't thread safe during initialization, so we do it eagerly, see
|
| 83 |
+
# issue #104
|
| 84 |
+
REQUIREMENT.parseString("x[]")
|
| 85 |
+
|
| 86 |
+
|
| 87 |
+
class Requirement:
|
| 88 |
+
"""Parse a requirement.
|
| 89 |
+
|
| 90 |
+
Parse a given requirement string into its parts, such as name, specifier,
|
| 91 |
+
URL, and extras. Raises InvalidRequirement on a badly-formed requirement
|
| 92 |
+
string.
|
| 93 |
+
"""
|
| 94 |
+
|
| 95 |
+
# TODO: Can we test whether something is contained within a requirement?
|
| 96 |
+
# If so how do we do that? Do we need to test against the _name_ of
|
| 97 |
+
# the thing as well as the version? What about the markers?
|
| 98 |
+
# TODO: Can we normalize the name and extra name?
|
| 99 |
+
|
| 100 |
+
def __init__(self, requirement_string: str) -> None:
|
| 101 |
+
try:
|
| 102 |
+
req = REQUIREMENT.parseString(requirement_string)
|
| 103 |
+
except ParseException as e:
|
| 104 |
+
raise InvalidRequirement(
|
| 105 |
+
f'Parse error at "{ requirement_string[e.loc : e.loc + 8]!r}": {e.msg}'
|
| 106 |
+
)
|
| 107 |
+
|
| 108 |
+
self.name: str = req.name
|
| 109 |
+
if req.url:
|
| 110 |
+
parsed_url = urllib.parse.urlparse(req.url)
|
| 111 |
+
if parsed_url.scheme == "file":
|
| 112 |
+
if urllib.parse.urlunparse(parsed_url) != req.url:
|
| 113 |
+
raise InvalidRequirement("Invalid URL given")
|
| 114 |
+
elif not (parsed_url.scheme and parsed_url.netloc) or (
|
| 115 |
+
not parsed_url.scheme and not parsed_url.netloc
|
| 116 |
+
):
|
| 117 |
+
raise InvalidRequirement(f"Invalid URL: {req.url}")
|
| 118 |
+
self.url: TOptional[str] = req.url
|
| 119 |
+
else:
|
| 120 |
+
self.url = None
|
| 121 |
+
self.extras: Set[str] = set(req.extras.asList() if req.extras else [])
|
| 122 |
+
self.specifier: SpecifierSet = SpecifierSet(req.specifier)
|
| 123 |
+
self.marker: TOptional[Marker] = req.marker if req.marker else None
|
| 124 |
+
|
| 125 |
+
def __str__(self) -> str:
|
| 126 |
+
parts: List[str] = [self.name]
|
| 127 |
+
|
| 128 |
+
if self.extras:
|
| 129 |
+
formatted_extras = ",".join(sorted(self.extras))
|
| 130 |
+
parts.append(f"[{formatted_extras}]")
|
| 131 |
+
|
| 132 |
+
if self.specifier:
|
| 133 |
+
parts.append(str(self.specifier))
|
| 134 |
+
|
| 135 |
+
if self.url:
|
| 136 |
+
parts.append(f"@ {self.url}")
|
| 137 |
+
if self.marker:
|
| 138 |
+
parts.append(" ")
|
| 139 |
+
|
| 140 |
+
if self.marker:
|
| 141 |
+
parts.append(f"; {self.marker}")
|
| 142 |
+
|
| 143 |
+
return "".join(parts)
|
| 144 |
+
|
| 145 |
+
def __repr__(self) -> str:
|
| 146 |
+
return f"<Requirement('{self}')>"
|
env-llmeval/lib/python3.10/site-packages/pkg_resources/_vendor/packaging/specifiers.py
ADDED
|
@@ -0,0 +1,828 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# This file is dual licensed under the terms of the Apache License, Version
|
| 2 |
+
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
|
| 3 |
+
# for complete details.
|
| 4 |
+
|
| 5 |
+
import abc
|
| 6 |
+
import functools
|
| 7 |
+
import itertools
|
| 8 |
+
import re
|
| 9 |
+
import warnings
|
| 10 |
+
from typing import (
|
| 11 |
+
Callable,
|
| 12 |
+
Dict,
|
| 13 |
+
Iterable,
|
| 14 |
+
Iterator,
|
| 15 |
+
List,
|
| 16 |
+
Optional,
|
| 17 |
+
Pattern,
|
| 18 |
+
Set,
|
| 19 |
+
Tuple,
|
| 20 |
+
TypeVar,
|
| 21 |
+
Union,
|
| 22 |
+
)
|
| 23 |
+
|
| 24 |
+
from .utils import canonicalize_version
|
| 25 |
+
from .version import LegacyVersion, Version, parse
|
| 26 |
+
|
| 27 |
+
ParsedVersion = Union[Version, LegacyVersion]
|
| 28 |
+
UnparsedVersion = Union[Version, LegacyVersion, str]
|
| 29 |
+
VersionTypeVar = TypeVar("VersionTypeVar", bound=UnparsedVersion)
|
| 30 |
+
CallableOperator = Callable[[ParsedVersion, str], bool]
|
| 31 |
+
|
| 32 |
+
|
| 33 |
+
class InvalidSpecifier(ValueError):
|
| 34 |
+
"""
|
| 35 |
+
An invalid specifier was found, users should refer to PEP 440.
|
| 36 |
+
"""
|
| 37 |
+
|
| 38 |
+
|
| 39 |
+
class BaseSpecifier(metaclass=abc.ABCMeta):
|
| 40 |
+
@abc.abstractmethod
|
| 41 |
+
def __str__(self) -> str:
|
| 42 |
+
"""
|
| 43 |
+
Returns the str representation of this Specifier like object. This
|
| 44 |
+
should be representative of the Specifier itself.
|
| 45 |
+
"""
|
| 46 |
+
|
| 47 |
+
@abc.abstractmethod
|
| 48 |
+
def __hash__(self) -> int:
|
| 49 |
+
"""
|
| 50 |
+
Returns a hash value for this Specifier like object.
|
| 51 |
+
"""
|
| 52 |
+
|
| 53 |
+
@abc.abstractmethod
|
| 54 |
+
def __eq__(self, other: object) -> bool:
|
| 55 |
+
"""
|
| 56 |
+
Returns a boolean representing whether or not the two Specifier like
|
| 57 |
+
objects are equal.
|
| 58 |
+
"""
|
| 59 |
+
|
| 60 |
+
@abc.abstractmethod
|
| 61 |
+
def __ne__(self, other: object) -> bool:
|
| 62 |
+
"""
|
| 63 |
+
Returns a boolean representing whether or not the two Specifier like
|
| 64 |
+
objects are not equal.
|
| 65 |
+
"""
|
| 66 |
+
|
| 67 |
+
@abc.abstractproperty
|
| 68 |
+
def prereleases(self) -> Optional[bool]:
|
| 69 |
+
"""
|
| 70 |
+
Returns whether or not pre-releases as a whole are allowed by this
|
| 71 |
+
specifier.
|
| 72 |
+
"""
|
| 73 |
+
|
| 74 |
+
@prereleases.setter
|
| 75 |
+
def prereleases(self, value: bool) -> None:
|
| 76 |
+
"""
|
| 77 |
+
Sets whether or not pre-releases as a whole are allowed by this
|
| 78 |
+
specifier.
|
| 79 |
+
"""
|
| 80 |
+
|
| 81 |
+
@abc.abstractmethod
|
| 82 |
+
def contains(self, item: str, prereleases: Optional[bool] = None) -> bool:
|
| 83 |
+
"""
|
| 84 |
+
Determines if the given item is contained within this specifier.
|
| 85 |
+
"""
|
| 86 |
+
|
| 87 |
+
@abc.abstractmethod
|
| 88 |
+
def filter(
|
| 89 |
+
self, iterable: Iterable[VersionTypeVar], prereleases: Optional[bool] = None
|
| 90 |
+
) -> Iterable[VersionTypeVar]:
|
| 91 |
+
"""
|
| 92 |
+
Takes an iterable of items and filters them so that only items which
|
| 93 |
+
are contained within this specifier are allowed in it.
|
| 94 |
+
"""
|
| 95 |
+
|
| 96 |
+
|
| 97 |
+
class _IndividualSpecifier(BaseSpecifier):
|
| 98 |
+
|
| 99 |
+
_operators: Dict[str, str] = {}
|
| 100 |
+
_regex: Pattern[str]
|
| 101 |
+
|
| 102 |
+
def __init__(self, spec: str = "", prereleases: Optional[bool] = None) -> None:
|
| 103 |
+
match = self._regex.search(spec)
|
| 104 |
+
if not match:
|
| 105 |
+
raise InvalidSpecifier(f"Invalid specifier: '{spec}'")
|
| 106 |
+
|
| 107 |
+
self._spec: Tuple[str, str] = (
|
| 108 |
+
match.group("operator").strip(),
|
| 109 |
+
match.group("version").strip(),
|
| 110 |
+
)
|
| 111 |
+
|
| 112 |
+
# Store whether or not this Specifier should accept prereleases
|
| 113 |
+
self._prereleases = prereleases
|
| 114 |
+
|
| 115 |
+
def __repr__(self) -> str:
|
| 116 |
+
pre = (
|
| 117 |
+
f", prereleases={self.prereleases!r}"
|
| 118 |
+
if self._prereleases is not None
|
| 119 |
+
else ""
|
| 120 |
+
)
|
| 121 |
+
|
| 122 |
+
return "<{}({!r}{})>".format(self.__class__.__name__, str(self), pre)
|
| 123 |
+
|
| 124 |
+
def __str__(self) -> str:
|
| 125 |
+
return "{}{}".format(*self._spec)
|
| 126 |
+
|
| 127 |
+
@property
|
| 128 |
+
def _canonical_spec(self) -> Tuple[str, str]:
|
| 129 |
+
return self._spec[0], canonicalize_version(self._spec[1])
|
| 130 |
+
|
| 131 |
+
def __hash__(self) -> int:
|
| 132 |
+
return hash(self._canonical_spec)
|
| 133 |
+
|
| 134 |
+
def __eq__(self, other: object) -> bool:
|
| 135 |
+
if isinstance(other, str):
|
| 136 |
+
try:
|
| 137 |
+
other = self.__class__(str(other))
|
| 138 |
+
except InvalidSpecifier:
|
| 139 |
+
return NotImplemented
|
| 140 |
+
elif not isinstance(other, self.__class__):
|
| 141 |
+
return NotImplemented
|
| 142 |
+
|
| 143 |
+
return self._canonical_spec == other._canonical_spec
|
| 144 |
+
|
| 145 |
+
def __ne__(self, other: object) -> bool:
|
| 146 |
+
if isinstance(other, str):
|
| 147 |
+
try:
|
| 148 |
+
other = self.__class__(str(other))
|
| 149 |
+
except InvalidSpecifier:
|
| 150 |
+
return NotImplemented
|
| 151 |
+
elif not isinstance(other, self.__class__):
|
| 152 |
+
return NotImplemented
|
| 153 |
+
|
| 154 |
+
return self._spec != other._spec
|
| 155 |
+
|
| 156 |
+
def _get_operator(self, op: str) -> CallableOperator:
|
| 157 |
+
operator_callable: CallableOperator = getattr(
|
| 158 |
+
self, f"_compare_{self._operators[op]}"
|
| 159 |
+
)
|
| 160 |
+
return operator_callable
|
| 161 |
+
|
| 162 |
+
def _coerce_version(self, version: UnparsedVersion) -> ParsedVersion:
|
| 163 |
+
if not isinstance(version, (LegacyVersion, Version)):
|
| 164 |
+
version = parse(version)
|
| 165 |
+
return version
|
| 166 |
+
|
| 167 |
+
@property
|
| 168 |
+
def operator(self) -> str:
|
| 169 |
+
return self._spec[0]
|
| 170 |
+
|
| 171 |
+
@property
|
| 172 |
+
def version(self) -> str:
|
| 173 |
+
return self._spec[1]
|
| 174 |
+
|
| 175 |
+
@property
|
| 176 |
+
def prereleases(self) -> Optional[bool]:
|
| 177 |
+
return self._prereleases
|
| 178 |
+
|
| 179 |
+
@prereleases.setter
|
| 180 |
+
def prereleases(self, value: bool) -> None:
|
| 181 |
+
self._prereleases = value
|
| 182 |
+
|
| 183 |
+
def __contains__(self, item: str) -> bool:
|
| 184 |
+
return self.contains(item)
|
| 185 |
+
|
| 186 |
+
def contains(
|
| 187 |
+
self, item: UnparsedVersion, prereleases: Optional[bool] = None
|
| 188 |
+
) -> bool:
|
| 189 |
+
|
| 190 |
+
# Determine if prereleases are to be allowed or not.
|
| 191 |
+
if prereleases is None:
|
| 192 |
+
prereleases = self.prereleases
|
| 193 |
+
|
| 194 |
+
# Normalize item to a Version or LegacyVersion, this allows us to have
|
| 195 |
+
# a shortcut for ``"2.0" in Specifier(">=2")
|
| 196 |
+
normalized_item = self._coerce_version(item)
|
| 197 |
+
|
| 198 |
+
# Determine if we should be supporting prereleases in this specifier
|
| 199 |
+
# or not, if we do not support prereleases than we can short circuit
|
| 200 |
+
# logic if this version is a prereleases.
|
| 201 |
+
if normalized_item.is_prerelease and not prereleases:
|
| 202 |
+
return False
|
| 203 |
+
|
| 204 |
+
# Actually do the comparison to determine if this item is contained
|
| 205 |
+
# within this Specifier or not.
|
| 206 |
+
operator_callable: CallableOperator = self._get_operator(self.operator)
|
| 207 |
+
return operator_callable(normalized_item, self.version)
|
| 208 |
+
|
| 209 |
+
def filter(
|
| 210 |
+
self, iterable: Iterable[VersionTypeVar], prereleases: Optional[bool] = None
|
| 211 |
+
) -> Iterable[VersionTypeVar]:
|
| 212 |
+
|
| 213 |
+
yielded = False
|
| 214 |
+
found_prereleases = []
|
| 215 |
+
|
| 216 |
+
kw = {"prereleases": prereleases if prereleases is not None else True}
|
| 217 |
+
|
| 218 |
+
# Attempt to iterate over all the values in the iterable and if any of
|
| 219 |
+
# them match, yield them.
|
| 220 |
+
for version in iterable:
|
| 221 |
+
parsed_version = self._coerce_version(version)
|
| 222 |
+
|
| 223 |
+
if self.contains(parsed_version, **kw):
|
| 224 |
+
# If our version is a prerelease, and we were not set to allow
|
| 225 |
+
# prereleases, then we'll store it for later in case nothing
|
| 226 |
+
# else matches this specifier.
|
| 227 |
+
if parsed_version.is_prerelease and not (
|
| 228 |
+
prereleases or self.prereleases
|
| 229 |
+
):
|
| 230 |
+
found_prereleases.append(version)
|
| 231 |
+
# Either this is not a prerelease, or we should have been
|
| 232 |
+
# accepting prereleases from the beginning.
|
| 233 |
+
else:
|
| 234 |
+
yielded = True
|
| 235 |
+
yield version
|
| 236 |
+
|
| 237 |
+
# Now that we've iterated over everything, determine if we've yielded
|
| 238 |
+
# any values, and if we have not and we have any prereleases stored up
|
| 239 |
+
# then we will go ahead and yield the prereleases.
|
| 240 |
+
if not yielded and found_prereleases:
|
| 241 |
+
for version in found_prereleases:
|
| 242 |
+
yield version
|
| 243 |
+
|
| 244 |
+
|
| 245 |
+
class LegacySpecifier(_IndividualSpecifier):
|
| 246 |
+
|
| 247 |
+
_regex_str = r"""
|
| 248 |
+
(?P<operator>(==|!=|<=|>=|<|>))
|
| 249 |
+
\s*
|
| 250 |
+
(?P<version>
|
| 251 |
+
[^,;\s)]* # Since this is a "legacy" specifier, and the version
|
| 252 |
+
# string can be just about anything, we match everything
|
| 253 |
+
# except for whitespace, a semi-colon for marker support,
|
| 254 |
+
# a closing paren since versions can be enclosed in
|
| 255 |
+
# them, and a comma since it's a version separator.
|
| 256 |
+
)
|
| 257 |
+
"""
|
| 258 |
+
|
| 259 |
+
_regex = re.compile(r"^\s*" + _regex_str + r"\s*$", re.VERBOSE | re.IGNORECASE)
|
| 260 |
+
|
| 261 |
+
_operators = {
|
| 262 |
+
"==": "equal",
|
| 263 |
+
"!=": "not_equal",
|
| 264 |
+
"<=": "less_than_equal",
|
| 265 |
+
">=": "greater_than_equal",
|
| 266 |
+
"<": "less_than",
|
| 267 |
+
">": "greater_than",
|
| 268 |
+
}
|
| 269 |
+
|
| 270 |
+
def __init__(self, spec: str = "", prereleases: Optional[bool] = None) -> None:
|
| 271 |
+
super().__init__(spec, prereleases)
|
| 272 |
+
|
| 273 |
+
warnings.warn(
|
| 274 |
+
"Creating a LegacyVersion has been deprecated and will be "
|
| 275 |
+
"removed in the next major release",
|
| 276 |
+
DeprecationWarning,
|
| 277 |
+
)
|
| 278 |
+
|
| 279 |
+
def _coerce_version(self, version: UnparsedVersion) -> LegacyVersion:
|
| 280 |
+
if not isinstance(version, LegacyVersion):
|
| 281 |
+
version = LegacyVersion(str(version))
|
| 282 |
+
return version
|
| 283 |
+
|
| 284 |
+
def _compare_equal(self, prospective: LegacyVersion, spec: str) -> bool:
|
| 285 |
+
return prospective == self._coerce_version(spec)
|
| 286 |
+
|
| 287 |
+
def _compare_not_equal(self, prospective: LegacyVersion, spec: str) -> bool:
|
| 288 |
+
return prospective != self._coerce_version(spec)
|
| 289 |
+
|
| 290 |
+
def _compare_less_than_equal(self, prospective: LegacyVersion, spec: str) -> bool:
|
| 291 |
+
return prospective <= self._coerce_version(spec)
|
| 292 |
+
|
| 293 |
+
def _compare_greater_than_equal(
|
| 294 |
+
self, prospective: LegacyVersion, spec: str
|
| 295 |
+
) -> bool:
|
| 296 |
+
return prospective >= self._coerce_version(spec)
|
| 297 |
+
|
| 298 |
+
def _compare_less_than(self, prospective: LegacyVersion, spec: str) -> bool:
|
| 299 |
+
return prospective < self._coerce_version(spec)
|
| 300 |
+
|
| 301 |
+
def _compare_greater_than(self, prospective: LegacyVersion, spec: str) -> bool:
|
| 302 |
+
return prospective > self._coerce_version(spec)
|
| 303 |
+
|
| 304 |
+
|
| 305 |
+
def _require_version_compare(
|
| 306 |
+
fn: Callable[["Specifier", ParsedVersion, str], bool]
|
| 307 |
+
) -> Callable[["Specifier", ParsedVersion, str], bool]:
|
| 308 |
+
@functools.wraps(fn)
|
| 309 |
+
def wrapped(self: "Specifier", prospective: ParsedVersion, spec: str) -> bool:
|
| 310 |
+
if not isinstance(prospective, Version):
|
| 311 |
+
return False
|
| 312 |
+
return fn(self, prospective, spec)
|
| 313 |
+
|
| 314 |
+
return wrapped
|
| 315 |
+
|
| 316 |
+
|
| 317 |
+
class Specifier(_IndividualSpecifier):
|
| 318 |
+
|
| 319 |
+
_regex_str = r"""
|
| 320 |
+
(?P<operator>(~=|==|!=|<=|>=|<|>|===))
|
| 321 |
+
(?P<version>
|
| 322 |
+
(?:
|
| 323 |
+
# The identity operators allow for an escape hatch that will
|
| 324 |
+
# do an exact string match of the version you wish to install.
|
| 325 |
+
# This will not be parsed by PEP 440 and we cannot determine
|
| 326 |
+
# any semantic meaning from it. This operator is discouraged
|
| 327 |
+
# but included entirely as an escape hatch.
|
| 328 |
+
(?<====) # Only match for the identity operator
|
| 329 |
+
\s*
|
| 330 |
+
[^\s]* # We just match everything, except for whitespace
|
| 331 |
+
# since we are only testing for strict identity.
|
| 332 |
+
)
|
| 333 |
+
|
|
| 334 |
+
(?:
|
| 335 |
+
# The (non)equality operators allow for wild card and local
|
| 336 |
+
# versions to be specified so we have to define these two
|
| 337 |
+
# operators separately to enable that.
|
| 338 |
+
(?<===|!=) # Only match for equals and not equals
|
| 339 |
+
|
| 340 |
+
\s*
|
| 341 |
+
v?
|
| 342 |
+
(?:[0-9]+!)? # epoch
|
| 343 |
+
[0-9]+(?:\.[0-9]+)* # release
|
| 344 |
+
(?: # pre release
|
| 345 |
+
[-_\.]?
|
| 346 |
+
(a|b|c|rc|alpha|beta|pre|preview)
|
| 347 |
+
[-_\.]?
|
| 348 |
+
[0-9]*
|
| 349 |
+
)?
|
| 350 |
+
(?: # post release
|
| 351 |
+
(?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
|
| 352 |
+
)?
|
| 353 |
+
|
| 354 |
+
# You cannot use a wild card and a dev or local version
|
| 355 |
+
# together so group them with a | and make them optional.
|
| 356 |
+
(?:
|
| 357 |
+
(?:[-_\.]?dev[-_\.]?[0-9]*)? # dev release
|
| 358 |
+
(?:\+[a-z0-9]+(?:[-_\.][a-z0-9]+)*)? # local
|
| 359 |
+
|
|
| 360 |
+
\.\* # Wild card syntax of .*
|
| 361 |
+
)?
|
| 362 |
+
)
|
| 363 |
+
|
|
| 364 |
+
(?:
|
| 365 |
+
# The compatible operator requires at least two digits in the
|
| 366 |
+
# release segment.
|
| 367 |
+
(?<=~=) # Only match for the compatible operator
|
| 368 |
+
|
| 369 |
+
\s*
|
| 370 |
+
v?
|
| 371 |
+
(?:[0-9]+!)? # epoch
|
| 372 |
+
[0-9]+(?:\.[0-9]+)+ # release (We have a + instead of a *)
|
| 373 |
+
(?: # pre release
|
| 374 |
+
[-_\.]?
|
| 375 |
+
(a|b|c|rc|alpha|beta|pre|preview)
|
| 376 |
+
[-_\.]?
|
| 377 |
+
[0-9]*
|
| 378 |
+
)?
|
| 379 |
+
(?: # post release
|
| 380 |
+
(?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
|
| 381 |
+
)?
|
| 382 |
+
(?:[-_\.]?dev[-_\.]?[0-9]*)? # dev release
|
| 383 |
+
)
|
| 384 |
+
|
|
| 385 |
+
(?:
|
| 386 |
+
# All other operators only allow a sub set of what the
|
| 387 |
+
# (non)equality operators do. Specifically they do not allow
|
| 388 |
+
# local versions to be specified nor do they allow the prefix
|
| 389 |
+
# matching wild cards.
|
| 390 |
+
(?<!==|!=|~=) # We have special cases for these
|
| 391 |
+
# operators so we want to make sure they
|
| 392 |
+
# don't match here.
|
| 393 |
+
|
| 394 |
+
\s*
|
| 395 |
+
v?
|
| 396 |
+
(?:[0-9]+!)? # epoch
|
| 397 |
+
[0-9]+(?:\.[0-9]+)* # release
|
| 398 |
+
(?: # pre release
|
| 399 |
+
[-_\.]?
|
| 400 |
+
(a|b|c|rc|alpha|beta|pre|preview)
|
| 401 |
+
[-_\.]?
|
| 402 |
+
[0-9]*
|
| 403 |
+
)?
|
| 404 |
+
(?: # post release
|
| 405 |
+
(?:-[0-9]+)|(?:[-_\.]?(post|rev|r)[-_\.]?[0-9]*)
|
| 406 |
+
)?
|
| 407 |
+
(?:[-_\.]?dev[-_\.]?[0-9]*)? # dev release
|
| 408 |
+
)
|
| 409 |
+
)
|
| 410 |
+
"""
|
| 411 |
+
|
| 412 |
+
_regex = re.compile(r"^\s*" + _regex_str + r"\s*$", re.VERBOSE | re.IGNORECASE)
|
| 413 |
+
|
| 414 |
+
_operators = {
|
| 415 |
+
"~=": "compatible",
|
| 416 |
+
"==": "equal",
|
| 417 |
+
"!=": "not_equal",
|
| 418 |
+
"<=": "less_than_equal",
|
| 419 |
+
">=": "greater_than_equal",
|
| 420 |
+
"<": "less_than",
|
| 421 |
+
">": "greater_than",
|
| 422 |
+
"===": "arbitrary",
|
| 423 |
+
}
|
| 424 |
+
|
| 425 |
+
@_require_version_compare
|
| 426 |
+
def _compare_compatible(self, prospective: ParsedVersion, spec: str) -> bool:
|
| 427 |
+
|
| 428 |
+
# Compatible releases have an equivalent combination of >= and ==. That
|
| 429 |
+
# is that ~=2.2 is equivalent to >=2.2,==2.*. This allows us to
|
| 430 |
+
# implement this in terms of the other specifiers instead of
|
| 431 |
+
# implementing it ourselves. The only thing we need to do is construct
|
| 432 |
+
# the other specifiers.
|
| 433 |
+
|
| 434 |
+
# We want everything but the last item in the version, but we want to
|
| 435 |
+
# ignore suffix segments.
|
| 436 |
+
prefix = ".".join(
|
| 437 |
+
list(itertools.takewhile(_is_not_suffix, _version_split(spec)))[:-1]
|
| 438 |
+
)
|
| 439 |
+
|
| 440 |
+
# Add the prefix notation to the end of our string
|
| 441 |
+
prefix += ".*"
|
| 442 |
+
|
| 443 |
+
return self._get_operator(">=")(prospective, spec) and self._get_operator("==")(
|
| 444 |
+
prospective, prefix
|
| 445 |
+
)
|
| 446 |
+
|
| 447 |
+
@_require_version_compare
|
| 448 |
+
def _compare_equal(self, prospective: ParsedVersion, spec: str) -> bool:
|
| 449 |
+
|
| 450 |
+
# We need special logic to handle prefix matching
|
| 451 |
+
if spec.endswith(".*"):
|
| 452 |
+
# In the case of prefix matching we want to ignore local segment.
|
| 453 |
+
prospective = Version(prospective.public)
|
| 454 |
+
# Split the spec out by dots, and pretend that there is an implicit
|
| 455 |
+
# dot in between a release segment and a pre-release segment.
|
| 456 |
+
split_spec = _version_split(spec[:-2]) # Remove the trailing .*
|
| 457 |
+
|
| 458 |
+
# Split the prospective version out by dots, and pretend that there
|
| 459 |
+
# is an implicit dot in between a release segment and a pre-release
|
| 460 |
+
# segment.
|
| 461 |
+
split_prospective = _version_split(str(prospective))
|
| 462 |
+
|
| 463 |
+
# Shorten the prospective version to be the same length as the spec
|
| 464 |
+
# so that we can determine if the specifier is a prefix of the
|
| 465 |
+
# prospective version or not.
|
| 466 |
+
shortened_prospective = split_prospective[: len(split_spec)]
|
| 467 |
+
|
| 468 |
+
# Pad out our two sides with zeros so that they both equal the same
|
| 469 |
+
# length.
|
| 470 |
+
padded_spec, padded_prospective = _pad_version(
|
| 471 |
+
split_spec, shortened_prospective
|
| 472 |
+
)
|
| 473 |
+
|
| 474 |
+
return padded_prospective == padded_spec
|
| 475 |
+
else:
|
| 476 |
+
# Convert our spec string into a Version
|
| 477 |
+
spec_version = Version(spec)
|
| 478 |
+
|
| 479 |
+
# If the specifier does not have a local segment, then we want to
|
| 480 |
+
# act as if the prospective version also does not have a local
|
| 481 |
+
# segment.
|
| 482 |
+
if not spec_version.local:
|
| 483 |
+
prospective = Version(prospective.public)
|
| 484 |
+
|
| 485 |
+
return prospective == spec_version
|
| 486 |
+
|
| 487 |
+
@_require_version_compare
|
| 488 |
+
def _compare_not_equal(self, prospective: ParsedVersion, spec: str) -> bool:
|
| 489 |
+
return not self._compare_equal(prospective, spec)
|
| 490 |
+
|
| 491 |
+
@_require_version_compare
|
| 492 |
+
def _compare_less_than_equal(self, prospective: ParsedVersion, spec: str) -> bool:
|
| 493 |
+
|
| 494 |
+
# NB: Local version identifiers are NOT permitted in the version
|
| 495 |
+
# specifier, so local version labels can be universally removed from
|
| 496 |
+
# the prospective version.
|
| 497 |
+
return Version(prospective.public) <= Version(spec)
|
| 498 |
+
|
| 499 |
+
@_require_version_compare
|
| 500 |
+
def _compare_greater_than_equal(
|
| 501 |
+
self, prospective: ParsedVersion, spec: str
|
| 502 |
+
) -> bool:
|
| 503 |
+
|
| 504 |
+
# NB: Local version identifiers are NOT permitted in the version
|
| 505 |
+
# specifier, so local version labels can be universally removed from
|
| 506 |
+
# the prospective version.
|
| 507 |
+
return Version(prospective.public) >= Version(spec)
|
| 508 |
+
|
| 509 |
+
@_require_version_compare
|
| 510 |
+
def _compare_less_than(self, prospective: ParsedVersion, spec_str: str) -> bool:
|
| 511 |
+
|
| 512 |
+
# Convert our spec to a Version instance, since we'll want to work with
|
| 513 |
+
# it as a version.
|
| 514 |
+
spec = Version(spec_str)
|
| 515 |
+
|
| 516 |
+
# Check to see if the prospective version is less than the spec
|
| 517 |
+
# version. If it's not we can short circuit and just return False now
|
| 518 |
+
# instead of doing extra unneeded work.
|
| 519 |
+
if not prospective < spec:
|
| 520 |
+
return False
|
| 521 |
+
|
| 522 |
+
# This special case is here so that, unless the specifier itself
|
| 523 |
+
# includes is a pre-release version, that we do not accept pre-release
|
| 524 |
+
# versions for the version mentioned in the specifier (e.g. <3.1 should
|
| 525 |
+
# not match 3.1.dev0, but should match 3.0.dev0).
|
| 526 |
+
if not spec.is_prerelease and prospective.is_prerelease:
|
| 527 |
+
if Version(prospective.base_version) == Version(spec.base_version):
|
| 528 |
+
return False
|
| 529 |
+
|
| 530 |
+
# If we've gotten to here, it means that prospective version is both
|
| 531 |
+
# less than the spec version *and* it's not a pre-release of the same
|
| 532 |
+
# version in the spec.
|
| 533 |
+
return True
|
| 534 |
+
|
| 535 |
+
@_require_version_compare
|
| 536 |
+
def _compare_greater_than(self, prospective: ParsedVersion, spec_str: str) -> bool:
|
| 537 |
+
|
| 538 |
+
# Convert our spec to a Version instance, since we'll want to work with
|
| 539 |
+
# it as a version.
|
| 540 |
+
spec = Version(spec_str)
|
| 541 |
+
|
| 542 |
+
# Check to see if the prospective version is greater than the spec
|
| 543 |
+
# version. If it's not we can short circuit and just return False now
|
| 544 |
+
# instead of doing extra unneeded work.
|
| 545 |
+
if not prospective > spec:
|
| 546 |
+
return False
|
| 547 |
+
|
| 548 |
+
# This special case is here so that, unless the specifier itself
|
| 549 |
+
# includes is a post-release version, that we do not accept
|
| 550 |
+
# post-release versions for the version mentioned in the specifier
|
| 551 |
+
# (e.g. >3.1 should not match 3.0.post0, but should match 3.2.post0).
|
| 552 |
+
if not spec.is_postrelease and prospective.is_postrelease:
|
| 553 |
+
if Version(prospective.base_version) == Version(spec.base_version):
|
| 554 |
+
return False
|
| 555 |
+
|
| 556 |
+
# Ensure that we do not allow a local version of the version mentioned
|
| 557 |
+
# in the specifier, which is technically greater than, to match.
|
| 558 |
+
if prospective.local is not None:
|
| 559 |
+
if Version(prospective.base_version) == Version(spec.base_version):
|
| 560 |
+
return False
|
| 561 |
+
|
| 562 |
+
# If we've gotten to here, it means that prospective version is both
|
| 563 |
+
# greater than the spec version *and* it's not a pre-release of the
|
| 564 |
+
# same version in the spec.
|
| 565 |
+
return True
|
| 566 |
+
|
| 567 |
+
def _compare_arbitrary(self, prospective: Version, spec: str) -> bool:
|
| 568 |
+
return str(prospective).lower() == str(spec).lower()
|
| 569 |
+
|
| 570 |
+
@property
|
| 571 |
+
def prereleases(self) -> bool:
|
| 572 |
+
|
| 573 |
+
# If there is an explicit prereleases set for this, then we'll just
|
| 574 |
+
# blindly use that.
|
| 575 |
+
if self._prereleases is not None:
|
| 576 |
+
return self._prereleases
|
| 577 |
+
|
| 578 |
+
# Look at all of our specifiers and determine if they are inclusive
|
| 579 |
+
# operators, and if they are if they are including an explicit
|
| 580 |
+
# prerelease.
|
| 581 |
+
operator, version = self._spec
|
| 582 |
+
if operator in ["==", ">=", "<=", "~=", "==="]:
|
| 583 |
+
# The == specifier can include a trailing .*, if it does we
|
| 584 |
+
# want to remove before parsing.
|
| 585 |
+
if operator == "==" and version.endswith(".*"):
|
| 586 |
+
version = version[:-2]
|
| 587 |
+
|
| 588 |
+
# Parse the version, and if it is a pre-release than this
|
| 589 |
+
# specifier allows pre-releases.
|
| 590 |
+
if parse(version).is_prerelease:
|
| 591 |
+
return True
|
| 592 |
+
|
| 593 |
+
return False
|
| 594 |
+
|
| 595 |
+
@prereleases.setter
|
| 596 |
+
def prereleases(self, value: bool) -> None:
|
| 597 |
+
self._prereleases = value
|
| 598 |
+
|
| 599 |
+
|
| 600 |
+
_prefix_regex = re.compile(r"^([0-9]+)((?:a|b|c|rc)[0-9]+)$")
|
| 601 |
+
|
| 602 |
+
|
| 603 |
+
def _version_split(version: str) -> List[str]:
|
| 604 |
+
result: List[str] = []
|
| 605 |
+
for item in version.split("."):
|
| 606 |
+
match = _prefix_regex.search(item)
|
| 607 |
+
if match:
|
| 608 |
+
result.extend(match.groups())
|
| 609 |
+
else:
|
| 610 |
+
result.append(item)
|
| 611 |
+
return result
|
| 612 |
+
|
| 613 |
+
|
| 614 |
+
def _is_not_suffix(segment: str) -> bool:
|
| 615 |
+
return not any(
|
| 616 |
+
segment.startswith(prefix) for prefix in ("dev", "a", "b", "rc", "post")
|
| 617 |
+
)
|
| 618 |
+
|
| 619 |
+
|
| 620 |
+
def _pad_version(left: List[str], right: List[str]) -> Tuple[List[str], List[str]]:
|
| 621 |
+
left_split, right_split = [], []
|
| 622 |
+
|
| 623 |
+
# Get the release segment of our versions
|
| 624 |
+
left_split.append(list(itertools.takewhile(lambda x: x.isdigit(), left)))
|
| 625 |
+
right_split.append(list(itertools.takewhile(lambda x: x.isdigit(), right)))
|
| 626 |
+
|
| 627 |
+
# Get the rest of our versions
|
| 628 |
+
left_split.append(left[len(left_split[0]) :])
|
| 629 |
+
right_split.append(right[len(right_split[0]) :])
|
| 630 |
+
|
| 631 |
+
# Insert our padding
|
| 632 |
+
left_split.insert(1, ["0"] * max(0, len(right_split[0]) - len(left_split[0])))
|
| 633 |
+
right_split.insert(1, ["0"] * max(0, len(left_split[0]) - len(right_split[0])))
|
| 634 |
+
|
| 635 |
+
return (list(itertools.chain(*left_split)), list(itertools.chain(*right_split)))
|
| 636 |
+
|
| 637 |
+
|
| 638 |
+
class SpecifierSet(BaseSpecifier):
|
| 639 |
+
def __init__(
|
| 640 |
+
self, specifiers: str = "", prereleases: Optional[bool] = None
|
| 641 |
+
) -> None:
|
| 642 |
+
|
| 643 |
+
# Split on , to break each individual specifier into it's own item, and
|
| 644 |
+
# strip each item to remove leading/trailing whitespace.
|
| 645 |
+
split_specifiers = [s.strip() for s in specifiers.split(",") if s.strip()]
|
| 646 |
+
|
| 647 |
+
# Parsed each individual specifier, attempting first to make it a
|
| 648 |
+
# Specifier and falling back to a LegacySpecifier.
|
| 649 |
+
parsed: Set[_IndividualSpecifier] = set()
|
| 650 |
+
for specifier in split_specifiers:
|
| 651 |
+
try:
|
| 652 |
+
parsed.add(Specifier(specifier))
|
| 653 |
+
except InvalidSpecifier:
|
| 654 |
+
parsed.add(LegacySpecifier(specifier))
|
| 655 |
+
|
| 656 |
+
# Turn our parsed specifiers into a frozen set and save them for later.
|
| 657 |
+
self._specs = frozenset(parsed)
|
| 658 |
+
|
| 659 |
+
# Store our prereleases value so we can use it later to determine if
|
| 660 |
+
# we accept prereleases or not.
|
| 661 |
+
self._prereleases = prereleases
|
| 662 |
+
|
| 663 |
+
def __repr__(self) -> str:
|
| 664 |
+
pre = (
|
| 665 |
+
f", prereleases={self.prereleases!r}"
|
| 666 |
+
if self._prereleases is not None
|
| 667 |
+
else ""
|
| 668 |
+
)
|
| 669 |
+
|
| 670 |
+
return "<SpecifierSet({!r}{})>".format(str(self), pre)
|
| 671 |
+
|
| 672 |
+
def __str__(self) -> str:
|
| 673 |
+
return ",".join(sorted(str(s) for s in self._specs))
|
| 674 |
+
|
| 675 |
+
def __hash__(self) -> int:
|
| 676 |
+
return hash(self._specs)
|
| 677 |
+
|
| 678 |
+
def __and__(self, other: Union["SpecifierSet", str]) -> "SpecifierSet":
|
| 679 |
+
if isinstance(other, str):
|
| 680 |
+
other = SpecifierSet(other)
|
| 681 |
+
elif not isinstance(other, SpecifierSet):
|
| 682 |
+
return NotImplemented
|
| 683 |
+
|
| 684 |
+
specifier = SpecifierSet()
|
| 685 |
+
specifier._specs = frozenset(self._specs | other._specs)
|
| 686 |
+
|
| 687 |
+
if self._prereleases is None and other._prereleases is not None:
|
| 688 |
+
specifier._prereleases = other._prereleases
|
| 689 |
+
elif self._prereleases is not None and other._prereleases is None:
|
| 690 |
+
specifier._prereleases = self._prereleases
|
| 691 |
+
elif self._prereleases == other._prereleases:
|
| 692 |
+
specifier._prereleases = self._prereleases
|
| 693 |
+
else:
|
| 694 |
+
raise ValueError(
|
| 695 |
+
"Cannot combine SpecifierSets with True and False prerelease "
|
| 696 |
+
"overrides."
|
| 697 |
+
)
|
| 698 |
+
|
| 699 |
+
return specifier
|
| 700 |
+
|
| 701 |
+
def __eq__(self, other: object) -> bool:
|
| 702 |
+
if isinstance(other, (str, _IndividualSpecifier)):
|
| 703 |
+
other = SpecifierSet(str(other))
|
| 704 |
+
elif not isinstance(other, SpecifierSet):
|
| 705 |
+
return NotImplemented
|
| 706 |
+
|
| 707 |
+
return self._specs == other._specs
|
| 708 |
+
|
| 709 |
+
def __ne__(self, other: object) -> bool:
|
| 710 |
+
if isinstance(other, (str, _IndividualSpecifier)):
|
| 711 |
+
other = SpecifierSet(str(other))
|
| 712 |
+
elif not isinstance(other, SpecifierSet):
|
| 713 |
+
return NotImplemented
|
| 714 |
+
|
| 715 |
+
return self._specs != other._specs
|
| 716 |
+
|
| 717 |
+
def __len__(self) -> int:
|
| 718 |
+
return len(self._specs)
|
| 719 |
+
|
| 720 |
+
def __iter__(self) -> Iterator[_IndividualSpecifier]:
|
| 721 |
+
return iter(self._specs)
|
| 722 |
+
|
| 723 |
+
@property
|
| 724 |
+
def prereleases(self) -> Optional[bool]:
|
| 725 |
+
|
| 726 |
+
# If we have been given an explicit prerelease modifier, then we'll
|
| 727 |
+
# pass that through here.
|
| 728 |
+
if self._prereleases is not None:
|
| 729 |
+
return self._prereleases
|
| 730 |
+
|
| 731 |
+
# If we don't have any specifiers, and we don't have a forced value,
|
| 732 |
+
# then we'll just return None since we don't know if this should have
|
| 733 |
+
# pre-releases or not.
|
| 734 |
+
if not self._specs:
|
| 735 |
+
return None
|
| 736 |
+
|
| 737 |
+
# Otherwise we'll see if any of the given specifiers accept
|
| 738 |
+
# prereleases, if any of them do we'll return True, otherwise False.
|
| 739 |
+
return any(s.prereleases for s in self._specs)
|
| 740 |
+
|
| 741 |
+
@prereleases.setter
|
| 742 |
+
def prereleases(self, value: bool) -> None:
|
| 743 |
+
self._prereleases = value
|
| 744 |
+
|
| 745 |
+
def __contains__(self, item: UnparsedVersion) -> bool:
|
| 746 |
+
return self.contains(item)
|
| 747 |
+
|
| 748 |
+
def contains(
|
| 749 |
+
self, item: UnparsedVersion, prereleases: Optional[bool] = None
|
| 750 |
+
) -> bool:
|
| 751 |
+
|
| 752 |
+
# Ensure that our item is a Version or LegacyVersion instance.
|
| 753 |
+
if not isinstance(item, (LegacyVersion, Version)):
|
| 754 |
+
item = parse(item)
|
| 755 |
+
|
| 756 |
+
# Determine if we're forcing a prerelease or not, if we're not forcing
|
| 757 |
+
# one for this particular filter call, then we'll use whatever the
|
| 758 |
+
# SpecifierSet thinks for whether or not we should support prereleases.
|
| 759 |
+
if prereleases is None:
|
| 760 |
+
prereleases = self.prereleases
|
| 761 |
+
|
| 762 |
+
# We can determine if we're going to allow pre-releases by looking to
|
| 763 |
+
# see if any of the underlying items supports them. If none of them do
|
| 764 |
+
# and this item is a pre-release then we do not allow it and we can
|
| 765 |
+
# short circuit that here.
|
| 766 |
+
# Note: This means that 1.0.dev1 would not be contained in something
|
| 767 |
+
# like >=1.0.devabc however it would be in >=1.0.debabc,>0.0.dev0
|
| 768 |
+
if not prereleases and item.is_prerelease:
|
| 769 |
+
return False
|
| 770 |
+
|
| 771 |
+
# We simply dispatch to the underlying specs here to make sure that the
|
| 772 |
+
# given version is contained within all of them.
|
| 773 |
+
# Note: This use of all() here means that an empty set of specifiers
|
| 774 |
+
# will always return True, this is an explicit design decision.
|
| 775 |
+
return all(s.contains(item, prereleases=prereleases) for s in self._specs)
|
| 776 |
+
|
| 777 |
+
def filter(
|
| 778 |
+
self, iterable: Iterable[VersionTypeVar], prereleases: Optional[bool] = None
|
| 779 |
+
) -> Iterable[VersionTypeVar]:
|
| 780 |
+
|
| 781 |
+
# Determine if we're forcing a prerelease or not, if we're not forcing
|
| 782 |
+
# one for this particular filter call, then we'll use whatever the
|
| 783 |
+
# SpecifierSet thinks for whether or not we should support prereleases.
|
| 784 |
+
if prereleases is None:
|
| 785 |
+
prereleases = self.prereleases
|
| 786 |
+
|
| 787 |
+
# If we have any specifiers, then we want to wrap our iterable in the
|
| 788 |
+
# filter method for each one, this will act as a logical AND amongst
|
| 789 |
+
# each specifier.
|
| 790 |
+
if self._specs:
|
| 791 |
+
for spec in self._specs:
|
| 792 |
+
iterable = spec.filter(iterable, prereleases=bool(prereleases))
|
| 793 |
+
return iterable
|
| 794 |
+
# If we do not have any specifiers, then we need to have a rough filter
|
| 795 |
+
# which will filter out any pre-releases, unless there are no final
|
| 796 |
+
# releases, and which will filter out LegacyVersion in general.
|
| 797 |
+
else:
|
| 798 |
+
filtered: List[VersionTypeVar] = []
|
| 799 |
+
found_prereleases: List[VersionTypeVar] = []
|
| 800 |
+
|
| 801 |
+
item: UnparsedVersion
|
| 802 |
+
parsed_version: Union[Version, LegacyVersion]
|
| 803 |
+
|
| 804 |
+
for item in iterable:
|
| 805 |
+
# Ensure that we some kind of Version class for this item.
|
| 806 |
+
if not isinstance(item, (LegacyVersion, Version)):
|
| 807 |
+
parsed_version = parse(item)
|
| 808 |
+
else:
|
| 809 |
+
parsed_version = item
|
| 810 |
+
|
| 811 |
+
# Filter out any item which is parsed as a LegacyVersion
|
| 812 |
+
if isinstance(parsed_version, LegacyVersion):
|
| 813 |
+
continue
|
| 814 |
+
|
| 815 |
+
# Store any item which is a pre-release for later unless we've
|
| 816 |
+
# already found a final version or we are accepting prereleases
|
| 817 |
+
if parsed_version.is_prerelease and not prereleases:
|
| 818 |
+
if not filtered:
|
| 819 |
+
found_prereleases.append(item)
|
| 820 |
+
else:
|
| 821 |
+
filtered.append(item)
|
| 822 |
+
|
| 823 |
+
# If we've found no items except for pre-releases, then we'll go
|
| 824 |
+
# ahead and use the pre-releases
|
| 825 |
+
if not filtered and found_prereleases and prereleases is None:
|
| 826 |
+
return found_prereleases
|
| 827 |
+
|
| 828 |
+
return filtered
|
env-llmeval/lib/python3.10/site-packages/sacrebleu/dataset/__main__.py
ADDED
|
@@ -0,0 +1,45 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
import sys
|
| 2 |
+
|
| 3 |
+
from . import DATASETS
|
| 4 |
+
|
| 5 |
+
try:
|
| 6 |
+
cmd = sys.argv[1]
|
| 7 |
+
except IndexError:
|
| 8 |
+
print(f"Usage: {sys.argv[0]} --check | --dump")
|
| 9 |
+
sys.exit(1)
|
| 10 |
+
|
| 11 |
+
if cmd == "--check":
|
| 12 |
+
import hashlib
|
| 13 |
+
import urllib.request
|
| 14 |
+
|
| 15 |
+
url_md5 = {}
|
| 16 |
+
|
| 17 |
+
for item in DATASETS.values():
|
| 18 |
+
if item.md5 is not None:
|
| 19 |
+
assert item.data
|
| 20 |
+
assert item.md5
|
| 21 |
+
assert len(item.data) == len(item.md5)
|
| 22 |
+
pairs = zip(item.data, item.md5)
|
| 23 |
+
for url, md5_hash in pairs:
|
| 24 |
+
url_md5[url] = md5_hash
|
| 25 |
+
|
| 26 |
+
for url, md5_hash in url_md5.items():
|
| 27 |
+
try:
|
| 28 |
+
print("Downloading ", url)
|
| 29 |
+
with urllib.request.urlopen(url) as f:
|
| 30 |
+
data = f.read()
|
| 31 |
+
except Exception as exc:
|
| 32 |
+
raise (exc)
|
| 33 |
+
|
| 34 |
+
if hashlib.md5(data).hexdigest() != md5_hash:
|
| 35 |
+
print("MD5 check failed for", url)
|
| 36 |
+
elif cmd == "--dump":
|
| 37 |
+
import re
|
| 38 |
+
|
| 39 |
+
# Dumps a table in markdown format
|
| 40 |
+
print(f'| {"Dataset":<30} | {"Description":<115} |')
|
| 41 |
+
header = "| " + "-" * 30 + " | " + "-" * 115 + " |"
|
| 42 |
+
print(header)
|
| 43 |
+
for name, item in DATASETS.items():
|
| 44 |
+
desc = re.sub(r"(http[s]?:\/\/\S+)", r"[URL](\1)", str(item.description))
|
| 45 |
+
print(f"| {name:<30} | {desc:<115} |")
|
env-llmeval/lib/python3.10/site-packages/sacrebleu/dataset/__pycache__/__main__.cpython-310.pyc
ADDED
|
Binary file (1.34 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/sacrebleu/dataset/__pycache__/base.cpython-310.pyc
ADDED
|
Binary file (7.35 kB). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/sacrebleu/dataset/__pycache__/iwslt_xml.cpython-310.pyc
ADDED
|
Binary file (472 Bytes). View file
|
|
|
env-llmeval/lib/python3.10/site-packages/sacrebleu/dataset/__pycache__/plain_text.cpython-310.pyc
ADDED
|
Binary file (1.57 kB). View file
|
|
|