peacock-data-public-datasets-idc-cronscript
/
venv
/lib
/python3.10
/site-packages
/tokenizers
/decoders
/__init__.pyi
# Generated content DO NOT EDIT | |
class Decoder: | |
""" | |
Base class for all decoders | |
This class is not supposed to be instantiated directly. Instead, any implementation of | |
a Decoder will return an instance of this class when instantiated. | |
""" | |
def decode(self, tokens): | |
""" | |
Decode the given list of tokens to a final string | |
Args: | |
tokens (:obj:`List[str]`): | |
The list of tokens to decode | |
Returns: | |
:obj:`str`: The decoded string | |
""" | |
pass | |
class BPEDecoder(Decoder): | |
""" | |
BPEDecoder Decoder | |
Args: | |
suffix (:obj:`str`, `optional`, defaults to :obj:`</w>`): | |
The suffix that was used to caracterize an end-of-word. This suffix will | |
be replaced by whitespaces during the decoding | |
""" | |
def __init__(self, suffix="</w>"): | |
pass | |
def decode(self, tokens): | |
""" | |
Decode the given list of tokens to a final string | |
Args: | |
tokens (:obj:`List[str]`): | |
The list of tokens to decode | |
Returns: | |
:obj:`str`: The decoded string | |
""" | |
pass | |
class ByteFallback(Decoder): | |
""" | |
ByteFallback Decoder | |
ByteFallback is a simple trick which converts tokens looking like `<0x61>` | |
to pure bytes, and attempts to make them into a string. If the tokens | |
cannot be decoded you will get � instead for each inconvertable byte token | |
""" | |
def __init__(self): | |
pass | |
def decode(self, tokens): | |
""" | |
Decode the given list of tokens to a final string | |
Args: | |
tokens (:obj:`List[str]`): | |
The list of tokens to decode | |
Returns: | |
:obj:`str`: The decoded string | |
""" | |
pass | |
class ByteLevel(Decoder): | |
""" | |
ByteLevel Decoder | |
This decoder is to be used in tandem with the :class:`~tokenizers.pre_tokenizers.ByteLevel` | |
:class:`~tokenizers.pre_tokenizers.PreTokenizer`. | |
""" | |
def __init__(self): | |
pass | |
def decode(self, tokens): | |
""" | |
Decode the given list of tokens to a final string | |
Args: | |
tokens (:obj:`List[str]`): | |
The list of tokens to decode | |
Returns: | |
:obj:`str`: The decoded string | |
""" | |
pass | |
class CTC(Decoder): | |
""" | |
CTC Decoder | |
Args: | |
pad_token (:obj:`str`, `optional`, defaults to :obj:`<pad>`): | |
The pad token used by CTC to delimit a new token. | |
word_delimiter_token (:obj:`str`, `optional`, defaults to :obj:`|`): | |
The word delimiter token. It will be replaced by a <space> | |
cleanup (:obj:`bool`, `optional`, defaults to :obj:`True`): | |
Whether to cleanup some tokenization artifacts. | |
Mainly spaces before punctuation, and some abbreviated english forms. | |
""" | |
def __init__(self, pad_token="<pad>", word_delimiter_token="|", cleanup=True): | |
pass | |
def decode(self, tokens): | |
""" | |
Decode the given list of tokens to a final string | |
Args: | |
tokens (:obj:`List[str]`): | |
The list of tokens to decode | |
Returns: | |
:obj:`str`: The decoded string | |
""" | |
pass | |
class Fuse(Decoder): | |
""" | |
Fuse Decoder | |
Fuse simply fuses every token into a single string. | |
This is the last step of decoding, this decoder exists only if | |
there is need to add other decoders *after* the fusion | |
""" | |
def __init__(self): | |
pass | |
def decode(self, tokens): | |
""" | |
Decode the given list of tokens to a final string | |
Args: | |
tokens (:obj:`List[str]`): | |
The list of tokens to decode | |
Returns: | |
:obj:`str`: The decoded string | |
""" | |
pass | |
class Metaspace(Decoder): | |
""" | |
Metaspace Decoder | |
Args: | |
replacement (:obj:`str`, `optional`, defaults to :obj:`▁`): | |
The replacement character. Must be exactly one character. By default we | |
use the `▁` (U+2581) meta symbol (Same as in SentencePiece). | |
prepend_scheme (:obj:`str`, `optional`, defaults to :obj:`"always"`): | |
Whether to add a space to the first word if there isn't already one. This | |
lets us treat `hello` exactly like `say hello`. | |
Choices: "always", "never", "first". First means the space is only added on the first | |
token (relevant when special tokens are used or other pre_tokenizer are used). | |
""" | |
def __init__(self, replacement="▁", prepend_scheme="always", split=True): | |
pass | |
def decode(self, tokens): | |
""" | |
Decode the given list of tokens to a final string | |
Args: | |
tokens (:obj:`List[str]`): | |
The list of tokens to decode | |
Returns: | |
:obj:`str`: The decoded string | |
""" | |
pass | |
class Replace(Decoder): | |
""" | |
Replace Decoder | |
This decoder is to be used in tandem with the :class:`~tokenizers.pre_tokenizers.Replace` | |
:class:`~tokenizers.pre_tokenizers.PreTokenizer`. | |
""" | |
def __init__(self, pattern, content): | |
pass | |
def decode(self, tokens): | |
""" | |
Decode the given list of tokens to a final string | |
Args: | |
tokens (:obj:`List[str]`): | |
The list of tokens to decode | |
Returns: | |
:obj:`str`: The decoded string | |
""" | |
pass | |
class Sequence(Decoder): | |
""" | |
Sequence Decoder | |
Args: | |
decoders (:obj:`List[Decoder]`) | |
The decoders that need to be chained | |
""" | |
def __init__(self, decoders): | |
pass | |
def decode(self, tokens): | |
""" | |
Decode the given list of tokens to a final string | |
Args: | |
tokens (:obj:`List[str]`): | |
The list of tokens to decode | |
Returns: | |
:obj:`str`: The decoded string | |
""" | |
pass | |
class Strip(Decoder): | |
""" | |
Strip normalizer | |
Strips n left characters of each token, or n right characters of each token | |
""" | |
def __init__(self, content, left=0, right=0): | |
pass | |
def decode(self, tokens): | |
""" | |
Decode the given list of tokens to a final string | |
Args: | |
tokens (:obj:`List[str]`): | |
The list of tokens to decode | |
Returns: | |
:obj:`str`: The decoded string | |
""" | |
pass | |
class WordPiece(Decoder): | |
""" | |
WordPiece Decoder | |
Args: | |
prefix (:obj:`str`, `optional`, defaults to :obj:`##`): | |
The prefix to use for subwords that are not a beginning-of-word | |
cleanup (:obj:`bool`, `optional`, defaults to :obj:`True`): | |
Whether to cleanup some tokenization artifacts. Mainly spaces before punctuation, | |
and some abbreviated english forms. | |
""" | |
def __init__(self, prefix="##", cleanup=True): | |
pass | |
def decode(self, tokens): | |
""" | |
Decode the given list of tokens to a final string | |
Args: | |
tokens (:obj:`List[str]`): | |
The list of tokens to decode | |
Returns: | |
:obj:`str`: The decoded string | |
""" | |
pass | |