peacock-data-public-datasets-idc-cronscript
/
venv
/lib
/python3.10
/site-packages
/tokenizers
/__init__.pyi
# Generated content DO NOT EDIT | |
class AddedToken: | |
""" | |
Represents a token that can be be added to a :class:`~tokenizers.Tokenizer`. | |
It can have special options that defines the way it should behave. | |
Args: | |
content (:obj:`str`): The content of the token | |
single_word (:obj:`bool`, defaults to :obj:`False`): | |
Defines whether this token should only match single words. If :obj:`True`, this | |
token will never match inside of a word. For example the token ``ing`` would match | |
on ``tokenizing`` if this option is :obj:`False`, but not if it is :obj:`True`. | |
The notion of "`inside of a word`" is defined by the word boundaries pattern in | |
regular expressions (ie. the token should start and end with word boundaries). | |
lstrip (:obj:`bool`, defaults to :obj:`False`): | |
Defines whether this token should strip all potential whitespaces on its left side. | |
If :obj:`True`, this token will greedily match any whitespace on its left. For | |
example if we try to match the token ``[MASK]`` with ``lstrip=True``, in the text | |
``"I saw a [MASK]"``, we would match on ``" [MASK]"``. (Note the space on the left). | |
rstrip (:obj:`bool`, defaults to :obj:`False`): | |
Defines whether this token should strip all potential whitespaces on its right | |
side. If :obj:`True`, this token will greedily match any whitespace on its right. | |
It works just like :obj:`lstrip` but on the right. | |
normalized (:obj:`bool`, defaults to :obj:`True` with :meth:`~tokenizers.Tokenizer.add_tokens` and :obj:`False` with :meth:`~tokenizers.Tokenizer.add_special_tokens`): | |
Defines whether this token should match against the normalized version of the input | |
text. For example, with the added token ``"yesterday"``, and a normalizer in charge of | |
lowercasing the text, the token could be extract from the input ``"I saw a lion | |
Yesterday"``. | |
special (:obj:`bool`, defaults to :obj:`False` with :meth:`~tokenizers.Tokenizer.add_tokens` and :obj:`False` with :meth:`~tokenizers.Tokenizer.add_special_tokens`): | |
Defines whether this token should be skipped when decoding. | |
""" | |
def __init__(self, content, single_word=False, lstrip=False, rstrip=False, normalized=True, special=False): | |
pass | |
@property | |
def content(self): | |
""" | |
Get the content of this :obj:`AddedToken` | |
""" | |
pass | |
@property | |
def lstrip(self): | |
""" | |
Get the value of the :obj:`lstrip` option | |
""" | |
pass | |
@property | |
def normalized(self): | |
""" | |
Get the value of the :obj:`normalized` option | |
""" | |
pass | |
@property | |
def rstrip(self): | |
""" | |
Get the value of the :obj:`rstrip` option | |
""" | |
pass | |
@property | |
def single_word(self): | |
""" | |
Get the value of the :obj:`single_word` option | |
""" | |
pass | |
@property | |
def special(self): | |
""" | |
Get the value of the :obj:`special` option | |
""" | |
pass | |
class Encoding: | |
""" | |
The :class:`~tokenizers.Encoding` represents the output of a :class:`~tokenizers.Tokenizer`. | |
""" | |
@property | |
def attention_mask(self): | |
""" | |
The attention mask | |
This indicates to the LM which tokens should be attended to, and which should not. | |
This is especially important when batching sequences, where we need to applying | |
padding. | |
Returns: | |
:obj:`List[int]`: The attention mask | |
""" | |
pass | |
def char_to_token(self, char_pos, sequence_index=0): | |
""" | |
Get the token that contains the char at the given position in the input sequence. | |
Args: | |
char_pos (:obj:`int`): | |
The position of a char in the input string | |
sequence_index (:obj:`int`, defaults to :obj:`0`): | |
The index of the sequence that contains the target char | |
Returns: | |
:obj:`int`: The index of the token that contains this char in the encoded sequence | |
""" | |
pass | |
def char_to_word(self, char_pos, sequence_index=0): | |
""" | |
Get the word that contains the char at the given position in the input sequence. | |
Args: | |
char_pos (:obj:`int`): | |
The position of a char in the input string | |
sequence_index (:obj:`int`, defaults to :obj:`0`): | |
The index of the sequence that contains the target char | |
Returns: | |
:obj:`int`: The index of the word that contains this char in the input sequence | |
""" | |
pass | |
@property | |
def ids(self): | |
""" | |
The generated IDs | |
The IDs are the main input to a Language Model. They are the token indices, | |
the numerical representations that a LM understands. | |
Returns: | |
:obj:`List[int]`: The list of IDs | |
""" | |
pass | |
@staticmethod | |
def merge(encodings, growing_offsets=True): | |
""" | |
Merge the list of encodings into one final :class:`~tokenizers.Encoding` | |
Args: | |
encodings (A :obj:`List` of :class:`~tokenizers.Encoding`): | |
The list of encodings that should be merged in one | |
growing_offsets (:obj:`bool`, defaults to :obj:`True`): | |
Whether the offsets should accumulate while merging | |
Returns: | |
:class:`~tokenizers.Encoding`: The resulting Encoding | |
""" | |
pass | |
@property | |
def n_sequences(self): | |
""" | |
The number of sequences represented | |
Returns: | |
:obj:`int`: The number of sequences in this :class:`~tokenizers.Encoding` | |
""" | |
pass | |
@property | |
def offsets(self): | |
""" | |
The offsets associated to each token | |
These offsets let's you slice the input string, and thus retrieve the original | |
part that led to producing the corresponding token. | |
Returns: | |
A :obj:`List` of :obj:`Tuple[int, int]`: The list of offsets | |
""" | |
pass | |
@property | |
def overflowing(self): | |
""" | |
A :obj:`List` of overflowing :class:`~tokenizers.Encoding` | |
When using truncation, the :class:`~tokenizers.Tokenizer` takes care of splitting | |
the output into as many pieces as required to match the specified maximum length. | |
This field lets you retrieve all the subsequent pieces. | |
When you use pairs of sequences, the overflowing pieces will contain enough | |
variations to cover all the possible combinations, while respecting the provided | |
maximum length. | |
""" | |
pass | |
def pad(self, length, direction="right", pad_id=0, pad_type_id=0, pad_token="[PAD]"): | |
""" | |
Pad the :class:`~tokenizers.Encoding` at the given length | |
Args: | |
length (:obj:`int`): | |
The desired length | |
direction: (:obj:`str`, defaults to :obj:`right`): | |
The expected padding direction. Can be either :obj:`right` or :obj:`left` | |
pad_id (:obj:`int`, defaults to :obj:`0`): | |
The ID corresponding to the padding token | |
pad_type_id (:obj:`int`, defaults to :obj:`0`): | |
The type ID corresponding to the padding token | |
pad_token (:obj:`str`, defaults to `[PAD]`): | |
The pad token to use | |
""" | |
pass | |
@property | |
def sequence_ids(self): | |
""" | |
The generated sequence indices. | |
They represent the index of the input sequence associated to each token. | |
The sequence id can be None if the token is not related to any input sequence, | |
like for example with special tokens. | |
Returns: | |
A :obj:`List` of :obj:`Optional[int]`: A list of optional sequence index. | |
""" | |
pass | |
def set_sequence_id(self, sequence_id): | |
""" | |
Set the given sequence index | |
Set the given sequence index for the whole range of tokens contained in this | |
:class:`~tokenizers.Encoding`. | |
""" | |
pass | |
@property | |
def special_tokens_mask(self): | |
""" | |
The special token mask | |
This indicates which tokens are special tokens, and which are not. | |
Returns: | |
:obj:`List[int]`: The special tokens mask | |
""" | |
pass | |
def token_to_chars(self, token_index): | |
""" | |
Get the offsets of the token at the given index. | |
The returned offsets are related to the input sequence that contains the | |
token. In order to determine in which input sequence it belongs, you | |
must call :meth:`~tokenizers.Encoding.token_to_sequence()`. | |
Args: | |
token_index (:obj:`int`): | |
The index of a token in the encoded sequence. | |
Returns: | |
:obj:`Tuple[int, int]`: The token offsets :obj:`(first, last + 1)` | |
""" | |
pass | |
def token_to_sequence(self, token_index): | |
""" | |
Get the index of the sequence represented by the given token. | |
In the general use case, this method returns :obj:`0` for a single sequence or | |
the first sequence of a pair, and :obj:`1` for the second sequence of a pair | |
Args: | |
token_index (:obj:`int`): | |
The index of a token in the encoded sequence. | |
Returns: | |
:obj:`int`: The sequence id of the given token | |
""" | |
pass | |
def token_to_word(self, token_index): | |
""" | |
Get the index of the word that contains the token in one of the input sequences. | |
The returned word index is related to the input sequence that contains | |
the token. In order to determine in which input sequence it belongs, you | |
must call :meth:`~tokenizers.Encoding.token_to_sequence()`. | |
Args: | |
token_index (:obj:`int`): | |
The index of a token in the encoded sequence. | |
Returns: | |
:obj:`int`: The index of the word in the relevant input sequence. | |
""" | |
pass | |
@property | |
def tokens(self): | |
""" | |
The generated tokens | |
They are the string representation of the IDs. | |
Returns: | |
:obj:`List[str]`: The list of tokens | |
""" | |
pass | |
def truncate(self, max_length, stride=0, direction="right"): | |
""" | |
Truncate the :class:`~tokenizers.Encoding` at the given length | |
If this :class:`~tokenizers.Encoding` represents multiple sequences, when truncating | |
this information is lost. It will be considered as representing a single sequence. | |
Args: | |
max_length (:obj:`int`): | |
The desired length | |
stride (:obj:`int`, defaults to :obj:`0`): | |
The length of previous content to be included in each overflowing piece | |
direction (:obj:`str`, defaults to :obj:`right`): | |
Truncate direction | |
""" | |
pass | |
@property | |
def type_ids(self): | |
""" | |
The generated type IDs | |
Generally used for tasks like sequence classification or question answering, | |
these tokens let the LM know which input sequence corresponds to each tokens. | |
Returns: | |
:obj:`List[int]`: The list of type ids | |
""" | |
pass | |
@property | |
def word_ids(self): | |
""" | |
The generated word indices. | |
They represent the index of the word associated to each token. | |
When the input is pre-tokenized, they correspond to the ID of the given input label, | |
otherwise they correspond to the words indices as defined by the | |
:class:`~tokenizers.pre_tokenizers.PreTokenizer` that was used. | |
For special tokens and such (any token that was generated from something that was | |
not part of the input), the output is :obj:`None` | |
Returns: | |
A :obj:`List` of :obj:`Optional[int]`: A list of optional word index. | |
""" | |
pass | |
def word_to_chars(self, word_index, sequence_index=0): | |
""" | |
Get the offsets of the word at the given index in one of the input sequences. | |
Args: | |
word_index (:obj:`int`): | |
The index of a word in one of the input sequences. | |
sequence_index (:obj:`int`, defaults to :obj:`0`): | |
The index of the sequence that contains the target word | |
Returns: | |
:obj:`Tuple[int, int]`: The range of characters (span) :obj:`(first, last + 1)` | |
""" | |
pass | |
def word_to_tokens(self, word_index, sequence_index=0): | |
""" | |
Get the encoded tokens corresponding to the word at the given index | |
in one of the input sequences. | |
Args: | |
word_index (:obj:`int`): | |
The index of a word in one of the input sequences. | |
sequence_index (:obj:`int`, defaults to :obj:`0`): | |
The index of the sequence that contains the target word | |
Returns: | |
:obj:`Tuple[int, int]`: The range of tokens: :obj:`(first, last + 1)` | |
""" | |
pass | |
@property | |
def words(self): | |
""" | |
The generated word indices. | |
.. warning:: | |
This is deprecated and will be removed in a future version. | |
Please use :obj:`~tokenizers.Encoding.word_ids` instead. | |
They represent the index of the word associated to each token. | |
When the input is pre-tokenized, they correspond to the ID of the given input label, | |
otherwise they correspond to the words indices as defined by the | |
:class:`~tokenizers.pre_tokenizers.PreTokenizer` that was used. | |
For special tokens and such (any token that was generated from something that was | |
not part of the input), the output is :obj:`None` | |
Returns: | |
A :obj:`List` of :obj:`Optional[int]`: A list of optional word index. | |
""" | |
pass | |
class NormalizedString: | |
""" | |
NormalizedString | |
A NormalizedString takes care of modifying an "original" string, to obtain a "normalized" one. | |
While making all the requested modifications, it keeps track of the alignment information | |
between the two versions of the string. | |
Args: | |
sequence: str: | |
The string sequence used to initialize this NormalizedString | |
""" | |
def append(self, s): | |
""" | |
Append the given sequence to the string | |
""" | |
pass | |
def clear(self): | |
""" | |
Clears the string | |
""" | |
pass | |
def filter(self, func): | |
""" | |
Filter each character of the string using the given func | |
""" | |
pass | |
def for_each(self, func): | |
""" | |
Calls the given function for each character of the string | |
""" | |
pass | |
def lowercase(self): | |
""" | |
Lowercase the string | |
""" | |
pass | |
def lstrip(self): | |
""" | |
Strip the left of the string | |
""" | |
pass | |
def map(self, func): | |
""" | |
Calls the given function for each character of the string | |
Replaces each character of the string using the returned value. Each | |
returned value **must** be a str of length 1 (ie a character). | |
""" | |
pass | |
def nfc(self): | |
""" | |
Runs the NFC normalization | |
""" | |
pass | |
def nfd(self): | |
""" | |
Runs the NFD normalization | |
""" | |
pass | |
def nfkc(self): | |
""" | |
Runs the NFKC normalization | |
""" | |
pass | |
def nfkd(self): | |
""" | |
Runs the NFKD normalization | |
""" | |
pass | |
@property | |
def normalized(self): | |
""" | |
The normalized part of the string | |
""" | |
pass | |
def prepend(self, s): | |
""" | |
Prepend the given sequence to the string | |
""" | |
pass | |
def replace(self, pattern, content): | |
""" | |
Replace the content of the given pattern with the provided content | |
Args: | |
pattern: Pattern: | |
A pattern used to match the string. Usually a string or a Regex | |
content: str: | |
The content to be used as replacement | |
""" | |
pass | |
def rstrip(self): | |
""" | |
Strip the right of the string | |
""" | |
pass | |
def slice(self, range): | |
""" | |
Slice the string using the given range | |
""" | |
pass | |
def split(self, pattern, behavior): | |
""" | |
Split the NormalizedString using the given pattern and the specified behavior | |
Args: | |
pattern: Pattern: | |
A pattern used to split the string. Usually a string or a regex built with `tokenizers.Regex` | |
behavior: SplitDelimiterBehavior: | |
The behavior to use when splitting. | |
Choices: "removed", "isolated", "merged_with_previous", "merged_with_next", | |
"contiguous" | |
Returns: | |
A list of NormalizedString, representing each split | |
""" | |
pass | |
def strip(self): | |
""" | |
Strip both ends of the string | |
""" | |
pass | |
def uppercase(self): | |
""" | |
Uppercase the string | |
""" | |
pass | |
class PreTokenizedString: | |
""" | |
PreTokenizedString | |
Wrapper over a string, that provides a way to normalize, pre-tokenize, tokenize the | |
underlying string, while keeping track of the alignment information (offsets). | |
The PreTokenizedString manages what we call `splits`. Each split represents a substring | |
which is a subpart of the original string, with the relevant offsets and tokens. | |
When calling one of the methods used to modify the PreTokenizedString (namely one of | |
`split`, `normalize` or `tokenize), only the `splits` that don't have any associated | |
tokens will get modified. | |
Args: | |
sequence: str: | |
The string sequence used to initialize this PreTokenizedString | |
""" | |
def __init__(self, sequence): | |
pass | |
def get_splits(self, offset_referential="original", offset_type="char"): | |
""" | |
Get the splits currently managed by the PreTokenizedString | |
Args: | |
offset_referential: :obj:`str` | |
Whether the returned splits should have offsets expressed relative | |
to the original string, or the normalized one. choices: "original", "normalized". | |
offset_type: :obj:`str` | |
Whether the returned splits should have offsets expressed in bytes or chars. | |
When slicing an str, we usually want to use chars, which is the default value. | |
Now in some cases it might be interesting to get these offsets expressed in bytes, | |
so it is possible to change this here. | |
choices: "char", "bytes" | |
Returns | |
A list of splits | |
""" | |
pass | |
def normalize(self, func): | |
""" | |
Normalize each split of the `PreTokenizedString` using the given `func` | |
Args: | |
func: Callable[[NormalizedString], None]: | |
The function used to normalize each underlying split. This function | |
does not need to return anything, just calling the methods on the provided | |
NormalizedString allow its modification. | |
""" | |
pass | |
def split(self, func): | |
""" | |
Split the PreTokenizedString using the given `func` | |
Args: | |
func: Callable[[index, NormalizedString], List[NormalizedString]]: | |
The function used to split each underlying split. | |
It is expected to return a list of `NormalizedString`, that represent the new | |
splits. If the given `NormalizedString` does not need any splitting, we can | |
just return it directly. | |
In order for the offsets to be tracked accurately, any returned `NormalizedString` | |
should come from calling either `.split` or `.slice` on the received one. | |
""" | |
pass | |
def to_encoding(self, type_id=0, word_idx=None): | |
""" | |
Return an Encoding generated from this PreTokenizedString | |
Args: | |
type_id: int = 0: | |
The type_id to be used on the generated Encoding. | |
word_idx: Optional[int] = None: | |
An optional word index to be used for each token of this Encoding. If provided, | |
all the word indices in the generated Encoding will use this value, instead | |
of the one automatically tracked during pre-tokenization. | |
Returns: | |
An Encoding | |
""" | |
pass | |
def tokenize(self, func): | |
""" | |
Tokenize each split of the `PreTokenizedString` using the given `func` | |
Args: | |
func: Callable[[str], List[Token]]: | |
The function used to tokenize each underlying split. This function must return | |
a list of Token generated from the input str. | |
""" | |
pass | |
class Regex: | |
""" | |
Instantiate a new Regex with the given pattern | |
""" | |
def __init__(self, pattern): | |
pass | |
class Token: | |
pass | |
class Tokenizer: | |
""" | |
A :obj:`Tokenizer` works as a pipeline. It processes some raw text as input | |
and outputs an :class:`~tokenizers.Encoding`. | |
Args: | |
model (:class:`~tokenizers.models.Model`): | |
The core algorithm that this :obj:`Tokenizer` should be using. | |
""" | |
def __init__(self, model): | |
pass | |
def add_special_tokens(self, tokens): | |
""" | |
Add the given special tokens to the Tokenizer. | |
If these tokens are already part of the vocabulary, it just let the Tokenizer know about | |
them. If they don't exist, the Tokenizer creates them, giving them a new id. | |
These special tokens will never be processed by the model (ie won't be split into | |
multiple tokens), and they can be removed from the output when decoding. | |
Args: | |
tokens (A :obj:`List` of :class:`~tokenizers.AddedToken` or :obj:`str`): | |
The list of special tokens we want to add to the vocabulary. Each token can either | |
be a string or an instance of :class:`~tokenizers.AddedToken` for more | |
customization. | |
Returns: | |
:obj:`int`: The number of tokens that were created in the vocabulary | |
""" | |
pass | |
def add_tokens(self, tokens): | |
""" | |
Add the given tokens to the vocabulary | |
The given tokens are added only if they don't already exist in the vocabulary. | |
Each token then gets a new attributed id. | |
Args: | |
tokens (A :obj:`List` of :class:`~tokenizers.AddedToken` or :obj:`str`): | |
The list of tokens we want to add to the vocabulary. Each token can be either a | |
string or an instance of :class:`~tokenizers.AddedToken` for more customization. | |
Returns: | |
:obj:`int`: The number of tokens that were created in the vocabulary | |
""" | |
pass | |
def decode(self, ids, skip_special_tokens=True): | |
""" | |
Decode the given list of ids back to a string | |
This is used to decode anything coming back from a Language Model | |
Args: | |
ids (A :obj:`List/Tuple` of :obj:`int`): | |
The list of ids that we want to decode | |
skip_special_tokens (:obj:`bool`, defaults to :obj:`True`): | |
Whether the special tokens should be removed from the decoded string | |
Returns: | |
:obj:`str`: The decoded string | |
""" | |
pass | |
def decode_batch(self, sequences, skip_special_tokens=True): | |
""" | |
Decode a batch of ids back to their corresponding string | |
Args: | |
sequences (:obj:`List` of :obj:`List[int]`): | |
The batch of sequences we want to decode | |
skip_special_tokens (:obj:`bool`, defaults to :obj:`True`): | |
Whether the special tokens should be removed from the decoded strings | |
Returns: | |
:obj:`List[str]`: A list of decoded strings | |
""" | |
pass | |
@property | |
def decoder(self): | |
""" | |
The `optional` :class:`~tokenizers.decoders.Decoder` in use by the Tokenizer | |
""" | |
pass | |
def enable_padding( | |
self, direction="right", pad_id=0, pad_type_id=0, pad_token="[PAD]", length=None, pad_to_multiple_of=None | |
): | |
""" | |
Enable the padding | |
Args: | |
direction (:obj:`str`, `optional`, defaults to :obj:`right`): | |
The direction in which to pad. Can be either ``right`` or ``left`` | |
pad_to_multiple_of (:obj:`int`, `optional`): | |
If specified, the padding length should always snap to the next multiple of the | |
given value. For example if we were going to pad witha length of 250 but | |
``pad_to_multiple_of=8`` then we will pad to 256. | |
pad_id (:obj:`int`, defaults to 0): | |
The id to be used when padding | |
pad_type_id (:obj:`int`, defaults to 0): | |
The type id to be used when padding | |
pad_token (:obj:`str`, defaults to :obj:`[PAD]`): | |
The pad token to be used when padding | |
length (:obj:`int`, `optional`): | |
If specified, the length at which to pad. If not specified we pad using the size of | |
the longest sequence in a batch. | |
""" | |
pass | |
def enable_truncation(self, max_length, stride=0, strategy="longest_first", direction="right"): | |
""" | |
Enable truncation | |
Args: | |
max_length (:obj:`int`): | |
The max length at which to truncate | |
stride (:obj:`int`, `optional`): | |
The length of the previous first sequence to be included in the overflowing | |
sequence | |
strategy (:obj:`str`, `optional`, defaults to :obj:`longest_first`): | |
The strategy used to truncation. Can be one of ``longest_first``, ``only_first`` or | |
``only_second``. | |
direction (:obj:`str`, defaults to :obj:`right`): | |
Truncate direction | |
""" | |
pass | |
def encode(self, sequence, pair=None, is_pretokenized=False, add_special_tokens=True): | |
""" | |
Encode the given sequence and pair. This method can process raw text sequences | |
as well as already pre-tokenized sequences. | |
Example: | |
Here are some examples of the inputs that are accepted:: | |
encode("A single sequence")` | |
encode("A sequence", "And its pair")` | |
encode([ "A", "pre", "tokenized", "sequence" ], is_pretokenized=True)` | |
encode( | |
[ "A", "pre", "tokenized", "sequence" ], [ "And", "its", "pair" ], | |
is_pretokenized=True | |
) | |
Args: | |
sequence (:obj:`~tokenizers.InputSequence`): | |
The main input sequence we want to encode. This sequence can be either raw | |
text or pre-tokenized, according to the ``is_pretokenized`` argument: | |
- If ``is_pretokenized=False``: :class:`~tokenizers.TextInputSequence` | |
- If ``is_pretokenized=True``: :class:`~tokenizers.PreTokenizedInputSequence` | |
pair (:obj:`~tokenizers.InputSequence`, `optional`): | |
An optional input sequence. The expected format is the same that for ``sequence``. | |
is_pretokenized (:obj:`bool`, defaults to :obj:`False`): | |
Whether the input is already pre-tokenized | |
add_special_tokens (:obj:`bool`, defaults to :obj:`True`): | |
Whether to add the special tokens | |
Returns: | |
:class:`~tokenizers.Encoding`: The encoded result | |
""" | |
pass | |
def encode_batch(self, input, is_pretokenized=False, add_special_tokens=True): | |
""" | |
Encode the given batch of inputs. This method accept both raw text sequences | |
as well as already pre-tokenized sequences. | |
Example: | |
Here are some examples of the inputs that are accepted:: | |
encode_batch([ | |
"A single sequence", | |
("A tuple with a sequence", "And its pair"), | |
[ "A", "pre", "tokenized", "sequence" ], | |
([ "A", "pre", "tokenized", "sequence" ], "And its pair") | |
]) | |
Args: | |
input (A :obj:`List`/:obj:`Tuple` of :obj:`~tokenizers.EncodeInput`): | |
A list of single sequences or pair sequences to encode. Each sequence | |
can be either raw text or pre-tokenized, according to the ``is_pretokenized`` | |
argument: | |
- If ``is_pretokenized=False``: :class:`~tokenizers.TextEncodeInput` | |
- If ``is_pretokenized=True``: :class:`~tokenizers.PreTokenizedEncodeInput` | |
is_pretokenized (:obj:`bool`, defaults to :obj:`False`): | |
Whether the input is already pre-tokenized | |
add_special_tokens (:obj:`bool`, defaults to :obj:`True`): | |
Whether to add the special tokens | |
Returns: | |
A :obj:`List` of :class:`~tokenizers.Encoding`: The encoded batch | |
""" | |
pass | |
@property | |
def encode_special_tokens(self): | |
""" | |
Modifies the tokenizer in order to use or not the special tokens | |
during encoding. | |
Args: | |
value (:obj:`bool`): | |
Whether to use the special tokens or not | |
""" | |
pass | |
@staticmethod | |
def from_buffer(buffer): | |
""" | |
Instantiate a new :class:`~tokenizers.Tokenizer` from the given buffer. | |
Args: | |
buffer (:obj:`bytes`): | |
A buffer containing a previously serialized :class:`~tokenizers.Tokenizer` | |
Returns: | |
:class:`~tokenizers.Tokenizer`: The new tokenizer | |
""" | |
pass | |
@staticmethod | |
def from_file(path): | |
""" | |
Instantiate a new :class:`~tokenizers.Tokenizer` from the file at the given path. | |
Args: | |
path (:obj:`str`): | |
A path to a local JSON file representing a previously serialized | |
:class:`~tokenizers.Tokenizer` | |
Returns: | |
:class:`~tokenizers.Tokenizer`: The new tokenizer | |
""" | |
pass | |
@staticmethod | |
def from_pretrained(identifier, revision="main", auth_token=None): | |
""" | |
Instantiate a new :class:`~tokenizers.Tokenizer` from an existing file on the | |
Hugging Face Hub. | |
Args: | |
identifier (:obj:`str`): | |
The identifier of a Model on the Hugging Face Hub, that contains | |
a tokenizer.json file | |
revision (:obj:`str`, defaults to `main`): | |
A branch or commit id | |
auth_token (:obj:`str`, `optional`, defaults to `None`): | |
An optional auth token used to access private repositories on the | |
Hugging Face Hub | |
Returns: | |
:class:`~tokenizers.Tokenizer`: The new tokenizer | |
""" | |
pass | |
@staticmethod | |
def from_str(json): | |
""" | |
Instantiate a new :class:`~tokenizers.Tokenizer` from the given JSON string. | |
Args: | |
json (:obj:`str`): | |
A valid JSON string representing a previously serialized | |
:class:`~tokenizers.Tokenizer` | |
Returns: | |
:class:`~tokenizers.Tokenizer`: The new tokenizer | |
""" | |
pass | |
def get_added_tokens_decoder(self): | |
""" | |
Get the underlying vocabulary | |
Returns: | |
:obj:`Dict[int, AddedToken]`: The vocabulary | |
""" | |
pass | |
def get_vocab(self, with_added_tokens=True): | |
""" | |
Get the underlying vocabulary | |
Args: | |
with_added_tokens (:obj:`bool`, defaults to :obj:`True`): | |
Whether to include the added tokens | |
Returns: | |
:obj:`Dict[str, int]`: The vocabulary | |
""" | |
pass | |
def get_vocab_size(self, with_added_tokens=True): | |
""" | |
Get the size of the underlying vocabulary | |
Args: | |
with_added_tokens (:obj:`bool`, defaults to :obj:`True`): | |
Whether to include the added tokens | |
Returns: | |
:obj:`int`: The size of the vocabulary | |
""" | |
pass | |
def id_to_token(self, id): | |
""" | |
Convert the given id to its corresponding token if it exists | |
Args: | |
id (:obj:`int`): | |
The id to convert | |
Returns: | |
:obj:`Optional[str]`: An optional token, :obj:`None` if out of vocabulary | |
""" | |
pass | |
@property | |
def model(self): | |
""" | |
The :class:`~tokenizers.models.Model` in use by the Tokenizer | |
""" | |
pass | |
def no_padding(self): | |
""" | |
Disable padding | |
""" | |
pass | |
def no_truncation(self): | |
""" | |
Disable truncation | |
""" | |
pass | |
@property | |
def normalizer(self): | |
""" | |
The `optional` :class:`~tokenizers.normalizers.Normalizer` in use by the Tokenizer | |
""" | |
pass | |
def num_special_tokens_to_add(self, is_pair): | |
""" | |
Return the number of special tokens that would be added for single/pair sentences. | |
:param is_pair: Boolean indicating if the input would be a single sentence or a pair | |
:return: | |
""" | |
pass | |
@property | |
def padding(self): | |
""" | |
Get the current padding parameters | |
`Cannot be set, use` :meth:`~tokenizers.Tokenizer.enable_padding` `instead` | |
Returns: | |
(:obj:`dict`, `optional`): | |
A dict with the current padding parameters if padding is enabled | |
""" | |
pass | |
def post_process(self, encoding, pair=None, add_special_tokens=True): | |
""" | |
Apply all the post-processing steps to the given encodings. | |
The various steps are: | |
1. Truncate according to the set truncation params (provided with | |
:meth:`~tokenizers.Tokenizer.enable_truncation`) | |
2. Apply the :class:`~tokenizers.processors.PostProcessor` | |
3. Pad according to the set padding params (provided with | |
:meth:`~tokenizers.Tokenizer.enable_padding`) | |
Args: | |
encoding (:class:`~tokenizers.Encoding`): | |
The :class:`~tokenizers.Encoding` corresponding to the main sequence. | |
pair (:class:`~tokenizers.Encoding`, `optional`): | |
An optional :class:`~tokenizers.Encoding` corresponding to the pair sequence. | |
add_special_tokens (:obj:`bool`): | |
Whether to add the special tokens | |
Returns: | |
:class:`~tokenizers.Encoding`: The final post-processed encoding | |
""" | |
pass | |
@property | |
def post_processor(self): | |
""" | |
The `optional` :class:`~tokenizers.processors.PostProcessor` in use by the Tokenizer | |
""" | |
pass | |
@property | |
def pre_tokenizer(self): | |
""" | |
The `optional` :class:`~tokenizers.pre_tokenizers.PreTokenizer` in use by the Tokenizer | |
""" | |
pass | |
def save(self, path, pretty=True): | |
""" | |
Save the :class:`~tokenizers.Tokenizer` to the file at the given path. | |
Args: | |
path (:obj:`str`): | |
A path to a file in which to save the serialized tokenizer. | |
pretty (:obj:`bool`, defaults to :obj:`True`): | |
Whether the JSON file should be pretty formatted. | |
""" | |
pass | |
def to_str(self, pretty=False): | |
""" | |
Gets a serialized string representing this :class:`~tokenizers.Tokenizer`. | |
Args: | |
pretty (:obj:`bool`, defaults to :obj:`False`): | |
Whether the JSON string should be pretty formatted. | |
Returns: | |
:obj:`str`: A string representing the serialized Tokenizer | |
""" | |
pass | |
def token_to_id(self, token): | |
""" | |
Convert the given token to its corresponding id if it exists | |
Args: | |
token (:obj:`str`): | |
The token to convert | |
Returns: | |
:obj:`Optional[int]`: An optional id, :obj:`None` if out of vocabulary | |
""" | |
pass | |
def train(self, files, trainer=None): | |
""" | |
Train the Tokenizer using the given files. | |
Reads the files line by line, while keeping all the whitespace, even new lines. | |
If you want to train from data store in-memory, you can check | |
:meth:`~tokenizers.Tokenizer.train_from_iterator` | |
Args: | |
files (:obj:`List[str]`): | |
A list of path to the files that we should use for training | |
trainer (:obj:`~tokenizers.trainers.Trainer`, `optional`): | |
An optional trainer that should be used to train our Model | |
""" | |
pass | |
def train_from_iterator(self, iterator, trainer=None, length=None): | |
""" | |
Train the Tokenizer using the provided iterator. | |
You can provide anything that is a Python Iterator | |
* A list of sequences :obj:`List[str]` | |
* A generator that yields :obj:`str` or :obj:`List[str]` | |
* A Numpy array of strings | |
* ... | |
Args: | |
iterator (:obj:`Iterator`): | |
Any iterator over strings or list of strings | |
trainer (:obj:`~tokenizers.trainers.Trainer`, `optional`): | |
An optional trainer that should be used to train our Model | |
length (:obj:`int`, `optional`): | |
The total number of sequences in the iterator. This is used to | |
provide meaningful progress tracking | |
""" | |
pass | |
@property | |
def truncation(self): | |
""" | |
Get the currently set truncation parameters | |
`Cannot set, use` :meth:`~tokenizers.Tokenizer.enable_truncation` `instead` | |
Returns: | |
(:obj:`dict`, `optional`): | |
A dict with the current truncation parameters if truncation is enabled | |
""" | |
pass | |