code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
'''simple docstring''' from __future__ import annotations def UpperCamelCase_ ( _UpperCAmelCase : list[int | str] ) -> None: """simple docstring""" create_state_space_tree(_UpperCAmelCase , [] , 0 , [0 for i in range(len(_UpperCAmelCase ) )] ) def UpperCamelCase_ ( _UpperCAmelCase : list[int | str] , _UpperCAmelCase : list[int | str] , _UpperCAmelCase : int , _UpperCAmelCase : list[int] , ) -> None: """simple docstring""" if index == len(_UpperCAmelCase ): print(_UpperCAmelCase ) return for i in range(len(_UpperCAmelCase ) ): if not index_used[i]: current_sequence.append(sequence[i] ) _UpperCAmelCase : List[str] = True create_state_space_tree(_UpperCAmelCase , _UpperCAmelCase , index + 1 , _UpperCAmelCase ) current_sequence.pop() _UpperCAmelCase : int = False __SCREAMING_SNAKE_CASE : list[int | str] = [3, 1, 2, 4] generate_all_permutations(sequence) __SCREAMING_SNAKE_CASE : list[int | str] = ["A", "B", "C"] generate_all_permutations(sequence_a)
31
from __future__ import annotations def __A ( __lowerCamelCase , __lowerCamelCase ) -> float: a = sorted(numsa + numsa ) a , a = divmod(len(__lowerCamelCase ) , 2 ) if mod == 1: return all_numbers[div] else: return (all_numbers[div] + all_numbers[div - 1]) / 2 if __name__ == "__main__": import doctest doctest.testmod() __UpperCamelCase : Tuple = [float(x) for x in input("Enter the elements of first array: ").split()] __UpperCamelCase : List[Any] = [float(x) for x in input("Enter the elements of second array: ").split()] print(F'The median of two arrays is: {median_of_two_arrays(array_a, array_a)}')
228
0
"""simple docstring""" import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class _UpperCAmelCase ( enum.Enum ): a__ : str = 0 a__ : List[Any] = 1 a__ : str = 2 @add_end_docstrings(_lowerCAmelCase ) class _UpperCAmelCase ( _lowerCAmelCase ): a__ : Dict = "\n In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The\n voice of Nicholas's young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western\n Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision\n and denounces one of the men as a horse thief. Although his father initially slaps him for making such an\n accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of\n the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop,\n begging for his blessing. <eod> </s> <eos>\n " def __init__( self : Optional[Any] , *_lowercase : Any , **_lowercase : Optional[int] ): super().__init__(*_lowercase , **_lowercase ) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == '''tf''' else MODEL_FOR_CAUSAL_LM_MAPPING ) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. __UpperCAmelCase = None if self.model.config.prefix is not None: __UpperCAmelCase = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. __UpperCAmelCase = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = self._sanitize_parameters(prefix=_lowercase , **self._forward_params ) __UpperCAmelCase = {**self._preprocess_params, **preprocess_params} __UpperCAmelCase = {**self._forward_params, **forward_params} def a ( self : Any , _lowercase : Optional[Any]=None , _lowercase : List[str]=None , _lowercase : int=None , _lowercase : Union[str, Any]=None , _lowercase : Union[str, Any]=None , _lowercase : Union[str, Any]=None , _lowercase : Union[str, Any]=None , _lowercase : List[Any]=None , **_lowercase : str , ): __UpperCAmelCase = {} if prefix is not None: __UpperCAmelCase = prefix if prefix: __UpperCAmelCase = self.tokenizer( _lowercase , padding=_lowercase , add_special_tokens=_lowercase , return_tensors=self.framework ) __UpperCAmelCase = prefix_inputs['''input_ids'''].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( F'''{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected''' ''' [None, \'hole\']''' ) __UpperCAmelCase = handle_long_generation preprocess_params.update(_lowercase ) __UpperCAmelCase = generate_kwargs __UpperCAmelCase = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError('''`return_text` is mutually exclusive with `return_full_text`''' ) if return_tensors is not None: raise ValueError('''`return_full_text` is mutually exclusive with `return_tensors`''' ) __UpperCAmelCase = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError('''`return_text` is mutually exclusive with `return_tensors`''' ) __UpperCAmelCase = ReturnType.TENSORS if return_type is not None: __UpperCAmelCase = return_type if clean_up_tokenization_spaces is not None: __UpperCAmelCase = clean_up_tokenization_spaces if stop_sequence is not None: __UpperCAmelCase = self.tokenizer.encode(_lowercase , add_special_tokens=_lowercase ) if len(_lowercase ) > 1: warnings.warn( '''Stopping on a multiple token sequence is not yet supported on transformers. The first token of''' ''' the stop sequence will be used as the stop sequence string in the interim.''' ) __UpperCAmelCase = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def a ( self : Optional[int] , *_lowercase : Optional[int] , **_lowercase : Any ): # Parse arguments if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({'''add_space_before_punct_symbol''': True} ) return super()._parse_and_tokenize(*_lowercase , **_lowercase ) def __call__( self : List[str] , _lowercase : str , **_lowercase : Optional[Any] ): return super().__call__(_lowercase , **_lowercase ) def a ( self : Union[str, Any] , _lowercase : Any , _lowercase : Dict="" , _lowercase : Union[str, Any]=None , **_lowercase : Tuple ): __UpperCAmelCase = self.tokenizer( prefix + prompt_text , padding=_lowercase , add_special_tokens=_lowercase , return_tensors=self.framework ) __UpperCAmelCase = prompt_text if handle_long_generation == "hole": __UpperCAmelCase = inputs['''input_ids'''].shape[-1] if "max_new_tokens" in generate_kwargs: __UpperCAmelCase = generate_kwargs['''max_new_tokens'''] else: __UpperCAmelCase = generate_kwargs.get('''max_length''' , self.model.config.max_length ) - cur_len if new_tokens < 0: raise ValueError('''We cannot infer how many new tokens are expected''' ) if cur_len + new_tokens > self.tokenizer.model_max_length: __UpperCAmelCase = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( '''We cannot use `hole` to handle this generation the number of desired tokens exceeds the''' ''' models max length''' ) __UpperCAmelCase = inputs['''input_ids'''][:, -keep_length:] if "attention_mask" in inputs: __UpperCAmelCase = inputs['''attention_mask'''][:, -keep_length:] return inputs def a ( self : Union[str, Any] , _lowercase : List[str] , **_lowercase : Optional[int] ): __UpperCAmelCase = model_inputs['''input_ids'''] __UpperCAmelCase = model_inputs.get('''attention_mask''' , _lowercase ) # Allow empty prompts if input_ids.shape[1] == 0: __UpperCAmelCase = None __UpperCAmelCase = None __UpperCAmelCase = 1 else: __UpperCAmelCase = input_ids.shape[0] __UpperCAmelCase = model_inputs.pop('''prompt_text''' ) # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. __UpperCAmelCase = generate_kwargs.pop('''prefix_length''' , 0 ) if prefix_length > 0: __UpperCAmelCase = '''max_new_tokens''' in generate_kwargs or ( '''generation_config''' in generate_kwargs and generate_kwargs['''generation_config'''].max_new_tokens is not None ) if not has_max_new_tokens: __UpperCAmelCase = generate_kwargs.get('''max_length''' ) or self.model.config.max_length generate_kwargs["max_length"] += prefix_length __UpperCAmelCase = '''min_new_tokens''' in generate_kwargs or ( '''generation_config''' in generate_kwargs and generate_kwargs['''generation_config'''].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL __UpperCAmelCase = self.model.generate(input_ids=_lowercase , attention_mask=_lowercase , **_lowercase ) __UpperCAmelCase = generated_sequence.shape[0] if self.framework == "pt": __UpperCAmelCase = generated_sequence.reshape(_lowercase , out_b // in_b , *generated_sequence.shape[1:] ) elif self.framework == "tf": __UpperCAmelCase = tf.reshape(_lowercase , (in_b, out_b // in_b, *generated_sequence.shape[1:]) ) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def a ( self : Optional[int] , _lowercase : Union[str, Any] , _lowercase : Optional[int]=ReturnType.FULL_TEXT , _lowercase : List[str]=True ): __UpperCAmelCase = model_outputs['''generated_sequence'''][0] __UpperCAmelCase = model_outputs['''input_ids'''] __UpperCAmelCase = model_outputs['''prompt_text'''] __UpperCAmelCase = generated_sequence.numpy().tolist() __UpperCAmelCase = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: __UpperCAmelCase = {'''generated_token_ids''': sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text __UpperCAmelCase = self.tokenizer.decode( _lowercase , skip_special_tokens=_lowercase , clean_up_tokenization_spaces=_lowercase , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: __UpperCAmelCase = 0 else: __UpperCAmelCase = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=_lowercase , clean_up_tokenization_spaces=_lowercase , ) ) if return_type == ReturnType.FULL_TEXT: __UpperCAmelCase = prompt_text + text[prompt_length:] else: __UpperCAmelCase = text[prompt_length:] __UpperCAmelCase = {'''generated_text''': all_text} records.append(_lowercase ) return records
86
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional from packaging import version if TYPE_CHECKING: from ... import PreTrainedTokenizer, TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import is_torch_available, logging _lowercase : List[Any] = logging.get_logger(__name__) _lowercase : int = { 'bigscience/bloom': 'https://huggingface.co/bigscience/bloom/resolve/main/config.json', 'bigscience/bloom-560m': 'https://huggingface.co/bigscience/bloom-560m/blob/main/config.json', 'bigscience/bloom-1b1': 'https://huggingface.co/bigscience/bloom-1b1/blob/main/config.json', 'bigscience/bloom-1b7': 'https://huggingface.co/bigscience/bloom-1b7/blob/main/config.json', 'bigscience/bloom-3b': 'https://huggingface.co/bigscience/bloom-3b/blob/main/config.json', 'bigscience/bloom-7b1': 'https://huggingface.co/bigscience/bloom-7b1/blob/main/config.json', } class _UpperCAmelCase ( _lowerCAmelCase ): a__ : Optional[Any] = "bloom" a__ : List[Any] = ["past_key_values"] a__ : Optional[Any] = { "num_hidden_layers": "n_layer", "num_attention_heads": "n_head", } def __init__( self : Union[str, Any] , _lowercase : Dict=25_08_80 , _lowercase : str=64 , _lowercase : int=2 , _lowercase : Union[str, Any]=8 , _lowercase : Optional[Any]=1E-5 , _lowercase : Dict=0.02 , _lowercase : Optional[int]=True , _lowercase : Any=1 , _lowercase : Dict=2 , _lowercase : Optional[Any]=False , _lowercase : Union[str, Any]=0.0 , _lowercase : str=0.0 , _lowercase : str=1 , _lowercase : int=False , **_lowercase : List[str] , ): __UpperCAmelCase = vocab_size # Backward compatibility with n_embed kwarg __UpperCAmelCase = kwargs.pop('''n_embed''' , _lowercase ) __UpperCAmelCase = hidden_size if n_embed is None else n_embed __UpperCAmelCase = n_layer __UpperCAmelCase = n_head __UpperCAmelCase = layer_norm_epsilon __UpperCAmelCase = initializer_range __UpperCAmelCase = use_cache __UpperCAmelCase = pretraining_tp __UpperCAmelCase = apply_residual_connection_post_layernorm __UpperCAmelCase = hidden_dropout __UpperCAmelCase = attention_dropout __UpperCAmelCase = bos_token_id __UpperCAmelCase = eos_token_id __UpperCAmelCase = slow_but_exact super().__init__(bos_token_id=_lowercase , eos_token_id=_lowercase , **_lowercase ) class _UpperCAmelCase ( _lowerCAmelCase ): a__ : List[str] = version.parse("1.12" ) def __init__( self : Optional[int] , _lowercase : PretrainedConfig , _lowercase : str = "default" , _lowercase : List[PatchingSpec] = None , _lowercase : bool = False , ): super().__init__(_lowercase , task=_lowercase , patching_specs=_lowercase , use_past=_lowercase ) if not getattr(self._config , '''pad_token_id''' , _lowercase ): # TODO: how to do that better? __UpperCAmelCase = 0 @property def a ( self : Optional[int] ): __UpperCAmelCase = OrderedDict({'''input_ids''': {0: '''batch''', 1: '''sequence'''}} ) if self.use_past: # BLOOM stores values on dynamic axis 2. For more details see: https://github.com/huggingface/transformers/pull/18344 self.fill_with_past_key_values_(_lowercase , direction='''inputs''' , inverted_values_shape=_lowercase ) __UpperCAmelCase = {0: '''batch''', 1: '''past_sequence + sequence'''} else: __UpperCAmelCase = {0: '''batch''', 1: '''sequence'''} return common_inputs @property def a ( self : Any ): return self._config.n_layer @property def a ( self : Tuple ): return self._config.n_head @property def a ( self : Dict ): return 1E-3 def a ( self : List[str] , _lowercase : "PreTrainedTokenizer" , _lowercase : int = -1 , _lowercase : int = -1 , _lowercase : bool = False , _lowercase : Optional["TensorType"] = None , ): __UpperCAmelCase = super(_lowercase , self ).generate_dummy_inputs( _lowercase , batch_size=_lowercase , seq_length=_lowercase , is_pair=_lowercase , framework=_lowercase ) # We need to order the input in the way they appears in the forward() __UpperCAmelCase = OrderedDict({'''input_ids''': common_inputs['''input_ids''']} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError('''Cannot generate dummy past_keys inputs without PyTorch installed.''' ) else: import torch __UpperCAmelCase , __UpperCAmelCase = common_inputs['''input_ids'''].shape # Not using the same length for past_key_values __UpperCAmelCase = seqlen + 2 __UpperCAmelCase = self._config.hidden_size // self.num_attention_heads __UpperCAmelCase = ( batch * self.num_attention_heads, head_dim, past_key_values_length, ) __UpperCAmelCase = ( batch * self.num_attention_heads, past_key_values_length, head_dim, ) __UpperCAmelCase = [ (torch.zeros(_lowercase ), torch.zeros(_lowercase )) for _ in range(self.num_layers ) ] __UpperCAmelCase = common_inputs['''attention_mask'''] if self.use_past: __UpperCAmelCase = ordered_inputs['''attention_mask'''].dtype __UpperCAmelCase = torch.cat( [ordered_inputs['''attention_mask'''], torch.ones(_lowercase , _lowercase , dtype=_lowercase )] , dim=1 ) return ordered_inputs @property def a ( self : Any ): return 13
86
1
import argparse import glob import logging import os import sys import time from collections import defaultdict from pathlib import Path from typing import Dict, List, Tuple import numpy as np import pytorch_lightning as pl import torch from callbacks import SeqaSeqLoggingCallback, get_checkpoint_callback, get_early_stopping_callback from torch import nn from torch.utils.data import DataLoader from transformers import MBartTokenizer, TaForConditionalGeneration from transformers.models.bart.modeling_bart import shift_tokens_right from utils import ( ROUGE_KEYS, LegacySeqaSeqDataset, SeqaSeqDataset, assert_all_frozen, calculate_bleu, calculate_rouge, check_output_dir, flatten_list, freeze_embeds, freeze_params, get_git_info, label_smoothed_nll_loss, lmap, pickle_save, save_git_info, save_json, use_task_specific_params, ) # need the parent dir module sys.path.insert(2, str(Path(__file__).resolve().parents[1])) from lightning_base import BaseTransformer, add_generic_args, generic_train # noqa UpperCAmelCase__ : Any = logging.getLogger(__name__) class UpperCAmelCase ( A__ ): '''simple docstring''' __UpperCamelCase : Union[str, Any] = '''summarization''' __UpperCamelCase : int = ['''loss'''] __UpperCamelCase : Dict = ROUGE_KEYS __UpperCamelCase : Any = '''rouge2''' def __init__( self : Any , lowerCAmelCase_ : Union[str, Any] , **lowerCAmelCase_ : Dict ): """simple docstring""" if hparams.sortish_sampler and hparams.gpus > 1: _A: Any = False elif hparams.max_tokens_per_batch is not None: if hparams.gpus > 1: raise NotImplementedError('''Dynamic Batch size does not work for multi-gpu training''' ) if hparams.sortish_sampler: raise ValueError('''--sortish_sampler and --max_tokens_per_batch may not be used simultaneously''' ) super().__init__(_lowerCAmelCase , num_labels=_lowerCAmelCase , mode=self.mode , **_lowerCAmelCase ) use_task_specific_params(self.model , '''summarization''' ) save_git_info(self.hparams.output_dir ) _A: Tuple = Path(self.output_dir ) / '''metrics.json''' _A: Optional[Any] = Path(self.output_dir ) / '''hparams.pkl''' pickle_save(self.hparams , self.hparams_save_path ) _A: Dict = 0 _A: Dict = defaultdict(_lowerCAmelCase ) _A: Dict = self.config.model_type _A: Any = self.config.tgt_vocab_size if self.model_type == '''fsmt''' else self.config.vocab_size _A: Tuple = { '''data_dir''': self.hparams.data_dir, '''max_source_length''': self.hparams.max_source_length, '''prefix''': self.model.config.prefix or '''''', } _A: Tuple = { '''train''': self.hparams.n_train, '''val''': self.hparams.n_val, '''test''': self.hparams.n_test, } _A: Optional[int] = {k: v if v >= 0 else None for k, v in n_observations_per_split.items()} _A: List[Any] = { '''train''': self.hparams.max_target_length, '''val''': self.hparams.val_max_target_length, '''test''': self.hparams.test_max_target_length, } assert self.target_lens["train"] <= self.target_lens["val"], F"""target_lens: {self.target_lens}""" assert self.target_lens["train"] <= self.target_lens["test"], F"""target_lens: {self.target_lens}""" if self.hparams.freeze_embeds: freeze_embeds(self.model ) if self.hparams.freeze_encoder: freeze_params(self.model.get_encoder() ) assert_all_frozen(self.model.get_encoder() ) _A: Dict = get_git_info()['''repo_sha'''] _A: Dict = hparams.num_workers _A: str = None # default to config if self.model.config.decoder_start_token_id is None and isinstance(self.tokenizer , _lowerCAmelCase ): _A: Optional[Any] = self.tokenizer.lang_code_to_id[hparams.tgt_lang] _A: Optional[int] = self.decoder_start_token_id _A: Dict = ( SeqaSeqDataset if hasattr(self.tokenizer , '''prepare_seq2seq_batch''' ) else LegacySeqaSeqDataset ) _A: List[str] = False _A: Optional[Any] = self.model.config.num_beams if self.hparams.eval_beams is None else self.hparams.eval_beams if self.hparams.eval_max_gen_length is not None: _A: Optional[int] = self.hparams.eval_max_gen_length else: _A: Any = self.model.config.max_length _A: Union[str, Any] = self.default_val_metric if self.hparams.val_metric is None else self.hparams.val_metric def __magic_name__ ( self : str , lowerCAmelCase_ : Dict[str, torch.Tensor] ): """simple docstring""" _A: Optional[Any] = { k: self.tokenizer.batch_decode(v.tolist() ) if '''mask''' not in k else v.shape for k, v in batch.items() } save_json(_lowerCAmelCase , Path(self.output_dir ) / '''text_batch.json''' ) save_json({k: v.tolist() for k, v in batch.items()} , Path(self.output_dir ) / '''tok_batch.json''' ) _A: List[Any] = True return readable_batch def __magic_name__ ( self : Tuple , lowerCAmelCase_ : List[Any] , **lowerCAmelCase_ : Dict ): """simple docstring""" return self.model(_lowerCAmelCase , **_lowerCAmelCase ) def __magic_name__ ( self : Tuple , lowerCAmelCase_ : List[int] ): """simple docstring""" _A: int = self.tokenizer.batch_decode( _lowerCAmelCase , skip_special_tokens=_lowerCAmelCase , clean_up_tokenization_spaces=_lowerCAmelCase ) return lmap(str.strip , _lowerCAmelCase ) def __magic_name__ ( self : Any , lowerCAmelCase_ : dict ): """simple docstring""" _A: Tuple = self.tokenizer.pad_token_id _A , _A: Any = batch['''input_ids'''], batch['''attention_mask'''] _A: Dict = batch['''labels'''] if isinstance(self.model , _lowerCAmelCase ): _A: Optional[Any] = self.model._shift_right(_lowerCAmelCase ) else: _A: List[str] = shift_tokens_right(_lowerCAmelCase , _lowerCAmelCase ) if not self.already_saved_batch: # This would be slightly better if it only happened on rank zero _A: List[str] = decoder_input_ids self.save_readable_batch(_lowerCAmelCase ) _A: Optional[Any] = self(_lowerCAmelCase , attention_mask=_lowerCAmelCase , decoder_input_ids=_lowerCAmelCase , use_cache=_lowerCAmelCase ) _A: Optional[Any] = outputs['''logits'''] if self.hparams.label_smoothing == 0: # Same behavior as modeling_bart.py, besides ignoring pad_token_id _A: Dict = nn.CrossEntropyLoss(ignore_index=_lowerCAmelCase ) assert lm_logits.shape[-1] == self.vocab_size _A: Dict = ce_loss_fct(lm_logits.view(-1 , lm_logits.shape[-1] ) , tgt_ids.view(-1 ) ) else: _A: List[str] = nn.functional.log_softmax(_lowerCAmelCase , dim=-1 ) _A , _A: Dict = label_smoothed_nll_loss( _lowerCAmelCase , _lowerCAmelCase , self.hparams.label_smoothing , ignore_index=_lowerCAmelCase ) return (loss,) @property def __magic_name__ ( self : List[Any] ): """simple docstring""" return self.tokenizer.pad_token_id def __magic_name__ ( self : int , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Dict ): """simple docstring""" _A: List[str] = self._step(_lowerCAmelCase ) _A: Tuple = dict(zip(self.loss_names , _lowerCAmelCase ) ) # tokens per batch _A: Tuple = batch['''input_ids'''].ne(self.pad ).sum() + batch['''labels'''].ne(self.pad ).sum() _A: Dict = batch['''input_ids'''].shape[0] _A: Tuple = batch['''input_ids'''].eq(self.pad ).sum() _A: Tuple = batch['''input_ids'''].eq(self.pad ).float().mean() # TODO(SS): make a wandb summary metric for this return {"loss": loss_tensors[0], "log": logs} def __magic_name__ ( self : int , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : List[Any] ): """simple docstring""" return self._generative_step(_lowerCAmelCase ) def __magic_name__ ( self : int , lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Optional[Any]="val" ): """simple docstring""" self.step_count += 1 _A: Union[str, Any] = {k: torch.stack([x[k] for x in outputs] ).mean() for k in self.loss_names} _A: str = losses['''loss'''] _A: List[str] = { k: np.array([x[k] for x in outputs] ).mean() for k in self.metric_names + ['''gen_time''', '''gen_len'''] } _A: int = ( generative_metrics[self.val_metric] if self.val_metric in generative_metrics else losses[self.val_metric] ) _A: Dict = torch.tensor(_lowerCAmelCase ).type_as(_lowerCAmelCase ) generative_metrics.update({k: v.item() for k, v in losses.items()} ) losses.update(_lowerCAmelCase ) _A: Dict = {F"""{prefix}_avg_{k}""": x for k, x in losses.items()} _A: List[Any] = self.step_count self.metrics[prefix].append(_lowerCAmelCase ) # callback writes this to self.metrics_save_path _A: Optional[int] = flatten_list([x['''preds'''] for x in outputs] ) return { "log": all_metrics, "preds": preds, F"""{prefix}_loss""": loss, F"""{prefix}_{self.val_metric}""": metric_tensor, } def __magic_name__ ( self : Union[str, Any] , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : Dict ): """simple docstring""" return calculate_rouge(_lowerCAmelCase , _lowerCAmelCase ) def __magic_name__ ( self : int , lowerCAmelCase_ : dict ): """simple docstring""" _A: Union[str, Any] = time.time() # parser.add_argument('--eval_max_gen_length', type=int, default=None, help='never generate more than n tokens') _A: Optional[Any] = self.model.generate( batch['''input_ids'''] , attention_mask=batch['''attention_mask'''] , use_cache=_lowerCAmelCase , decoder_start_token_id=self.decoder_start_token_id , num_beams=self.eval_beams , max_length=self.eval_max_length , ) _A: Tuple = (time.time() - ta) / batch['''input_ids'''].shape[0] _A: Any = self.ids_to_clean_text(_lowerCAmelCase ) _A: int = self.ids_to_clean_text(batch['''labels'''] ) _A: Dict = self._step(_lowerCAmelCase ) _A: Any = dict(zip(self.loss_names , _lowerCAmelCase ) ) _A: Optional[int] = self.calc_generative_metrics(_lowerCAmelCase , _lowerCAmelCase ) _A: Any = np.mean(lmap(_lowerCAmelCase , _lowerCAmelCase ) ) base_metrics.update(gen_time=_lowerCAmelCase , gen_len=_lowerCAmelCase , preds=_lowerCAmelCase , target=_lowerCAmelCase , **_lowerCAmelCase ) return base_metrics def __magic_name__ ( self : str , lowerCAmelCase_ : Optional[Any] , lowerCAmelCase_ : List[str] ): """simple docstring""" return self._generative_step(_lowerCAmelCase ) def __magic_name__ ( self : List[Any] , lowerCAmelCase_ : Dict ): """simple docstring""" return self.validation_epoch_end(_lowerCAmelCase , prefix='''test''' ) def __magic_name__ ( self : Optional[Any] , lowerCAmelCase_ : Tuple ): """simple docstring""" _A: Optional[Any] = self.n_obs[type_path] _A: Union[str, Any] = self.target_lens[type_path] _A: List[str] = self.dataset_class( self.tokenizer , type_path=_lowerCAmelCase , n_obs=_lowerCAmelCase , max_target_length=_lowerCAmelCase , **self.dataset_kwargs , ) return dataset def __magic_name__ ( self : Any , lowerCAmelCase_ : str , lowerCAmelCase_ : int , lowerCAmelCase_ : bool = False ): """simple docstring""" _A: Any = self.get_dataset(_lowerCAmelCase ) if self.hparams.sortish_sampler and type_path != "test" and type_path != "val": _A: Dict = dataset.make_sortish_sampler(_lowerCAmelCase , distributed=self.hparams.gpus > 1 ) return DataLoader( _lowerCAmelCase , batch_size=_lowerCAmelCase , collate_fn=dataset.collate_fn , shuffle=_lowerCAmelCase , num_workers=self.num_workers , sampler=_lowerCAmelCase , ) elif self.hparams.max_tokens_per_batch is not None and type_path != "test" and type_path != "val": _A: Any = dataset.make_dynamic_sampler( self.hparams.max_tokens_per_batch , distributed=self.hparams.gpus > 1 ) return DataLoader( _lowerCAmelCase , batch_sampler=_lowerCAmelCase , collate_fn=dataset.collate_fn , num_workers=self.num_workers , ) else: return DataLoader( _lowerCAmelCase , batch_size=_lowerCAmelCase , collate_fn=dataset.collate_fn , shuffle=_lowerCAmelCase , num_workers=self.num_workers , sampler=_lowerCAmelCase , ) def __magic_name__ ( self : Union[str, Any] ): """simple docstring""" _A: int = self.get_dataloader('''train''' , batch_size=self.hparams.train_batch_size , shuffle=_lowerCAmelCase ) return dataloader def __magic_name__ ( self : Optional[int] ): """simple docstring""" return self.get_dataloader('''val''' , batch_size=self.hparams.eval_batch_size ) def __magic_name__ ( self : Any ): """simple docstring""" return self.get_dataloader('''test''' , batch_size=self.hparams.eval_batch_size ) @staticmethod def __magic_name__ ( lowerCAmelCase_ : List[Any] , lowerCAmelCase_ : Dict ): """simple docstring""" BaseTransformer.add_model_specific_args(_lowerCAmelCase , _lowerCAmelCase ) add_generic_args(_lowerCAmelCase , _lowerCAmelCase ) parser.add_argument( '''--max_source_length''' , default=1_0_2_4 , type=_lowerCAmelCase , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--max_target_length''' , default=5_6 , type=_lowerCAmelCase , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--val_max_target_length''' , default=1_4_2 , type=_lowerCAmelCase , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument( '''--test_max_target_length''' , default=1_4_2 , type=_lowerCAmelCase , help=( '''The maximum total input sequence length after tokenization. Sequences longer ''' '''than this will be truncated, sequences shorter will be padded.''' ) , ) parser.add_argument('''--freeze_encoder''' , action='''store_true''' ) parser.add_argument('''--freeze_embeds''' , action='''store_true''' ) parser.add_argument('''--sortish_sampler''' , action='''store_true''' , default=_lowerCAmelCase ) parser.add_argument('''--overwrite_output_dir''' , action='''store_true''' , default=_lowerCAmelCase ) parser.add_argument('''--max_tokens_per_batch''' , type=_lowerCAmelCase , default=_lowerCAmelCase ) parser.add_argument('''--logger_name''' , type=_lowerCAmelCase , choices=['''default''', '''wandb''', '''wandb_shared'''] , default='''default''' ) parser.add_argument('''--n_train''' , type=_lowerCAmelCase , default=-1 , required=_lowerCAmelCase , help='''# examples. -1 means use all.''' ) parser.add_argument('''--n_val''' , type=_lowerCAmelCase , default=5_0_0 , required=_lowerCAmelCase , help='''# examples. -1 means use all.''' ) parser.add_argument('''--n_test''' , type=_lowerCAmelCase , default=-1 , required=_lowerCAmelCase , help='''# examples. -1 means use all.''' ) parser.add_argument( '''--task''' , type=_lowerCAmelCase , default='''summarization''' , required=_lowerCAmelCase , help='''# examples. -1 means use all.''' ) parser.add_argument('''--label_smoothing''' , type=_lowerCAmelCase , default=0.0 , required=_lowerCAmelCase ) parser.add_argument('''--src_lang''' , type=_lowerCAmelCase , default='''''' , required=_lowerCAmelCase ) parser.add_argument('''--tgt_lang''' , type=_lowerCAmelCase , default='''''' , required=_lowerCAmelCase ) parser.add_argument('''--eval_beams''' , type=_lowerCAmelCase , default=_lowerCAmelCase , required=_lowerCAmelCase ) parser.add_argument( '''--val_metric''' , type=_lowerCAmelCase , default=_lowerCAmelCase , required=_lowerCAmelCase , choices=['''bleu''', '''rouge2''', '''loss''', None] ) parser.add_argument('''--eval_max_gen_length''' , type=_lowerCAmelCase , default=_lowerCAmelCase , help='''never generate more than n tokens''' ) parser.add_argument('''--save_top_k''' , type=_lowerCAmelCase , default=1 , required=_lowerCAmelCase , help='''How many checkpoints to save''' ) parser.add_argument( '''--early_stopping_patience''' , type=_lowerCAmelCase , default=-1 , required=_lowerCAmelCase , help=( '''-1 means never early stop. early_stopping_patience is measured in validation checks, not epochs. So''' ''' val_check_interval will effect it.''' ) , ) return parser class UpperCAmelCase ( A__ ): '''simple docstring''' __UpperCamelCase : List[str] = '''translation''' __UpperCamelCase : int = ['''loss'''] __UpperCamelCase : List[str] = ['''bleu'''] __UpperCamelCase : Optional[int] = '''bleu''' def __init__( self : Optional[int] , lowerCAmelCase_ : List[str] , **lowerCAmelCase_ : Union[str, Any] ): """simple docstring""" super().__init__(_lowerCAmelCase , **_lowerCAmelCase ) _A: Tuple = hparams.src_lang _A: List[str] = hparams.tgt_lang def __magic_name__ ( self : Tuple , lowerCAmelCase_ : Optional[int] , lowerCAmelCase_ : Union[str, Any] ): """simple docstring""" return calculate_bleu(_lowerCAmelCase , _lowerCAmelCase ) def lowerCamelCase__ ( a , a=None ) -> SummarizationModule: Path(args.output_dir ).mkdir(exist_ok=a ) check_output_dir(a , expected_items=3 ) if model is None: if "summarization" in args.task: _A: int = SummarizationModule(a ) else: _A: Any = TranslationModule(a ) _A: Tuple = Path(args.data_dir ).name if ( args.logger_name == "default" or args.fast_dev_run or str(args.output_dir ).startswith('''/tmp''' ) or str(args.output_dir ).startswith('''/var''' ) ): _A: List[str] = True # don't pollute wandb logs unnecessarily elif args.logger_name == "wandb": from pytorch_lightning.loggers import WandbLogger _A: int = os.environ.get('''WANDB_PROJECT''' , a ) _A: Optional[Any] = WandbLogger(name=model.output_dir.name , project=a ) elif args.logger_name == "wandb_shared": from pytorch_lightning.loggers import WandbLogger _A: str = WandbLogger(name=model.output_dir.name , project=f"""hf_{dataset}""" ) if args.early_stopping_patience >= 0: _A: Dict = get_early_stopping_callback(model.val_metric , args.early_stopping_patience ) else: _A: str = False _A: int = args.val_metric == '''loss''' _A: Tuple = generic_train( a , a , logging_callback=SeqaSeqLoggingCallback() , checkpoint_callback=get_checkpoint_callback( args.output_dir , model.val_metric , args.save_top_k , a ) , early_stopping_callback=a , logger=a , ) pickle_save(model.hparams , model.output_dir / '''hparams.pkl''' ) if not args.do_predict: return model _A: str = '''''' _A: Union[str, Any] = sorted(glob.glob(os.path.join(args.output_dir , '''*.ckpt''' ) , recursive=a ) ) if checkpoints: _A: List[str] = checkpoints[-1] _A: str = checkpoints[-1] trainer.logger.log_hyperparams(model.hparams ) # test() without a model tests using the best checkpoint automatically trainer.test() return model if __name__ == "__main__": UpperCAmelCase__ : Tuple = argparse.ArgumentParser() UpperCAmelCase__ : List[Any] = pl.Trainer.add_argparse_args(parser) UpperCAmelCase__ : Dict = SummarizationModule.add_model_specific_args(parser, os.getcwd()) UpperCAmelCase__ : Union[str, Any] = parser.parse_args() main(args)
121
'''simple docstring''' from math import asin, atan, cos, radians, sin, sqrt, tan _lowerCamelCase : List[Any] = 6_3_7_8_1_3_7.0 _lowerCamelCase : List[Any] = 6_3_5_6_7_5_2.3_1_4_2_4_5 _lowerCamelCase : Optional[int] = 637_8137 def __a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) ->float: """simple docstring""" A = (AXIS_A - AXIS_B) / AXIS_A A = atan((1 - flattening) * tan(radians(UpperCAmelCase ) ) ) A = atan((1 - flattening) * tan(radians(UpperCAmelCase ) ) ) A = radians(UpperCAmelCase ) A = radians(UpperCAmelCase ) # Equation A = sin((phi_a - phi_a) / 2 ) A = sin((lambda_a - lambda_a) / 2 ) # Square both values sin_sq_phi *= sin_sq_phi sin_sq_lambda *= sin_sq_lambda A = sqrt(sin_sq_phi + (cos(UpperCAmelCase ) * cos(UpperCAmelCase ) * sin_sq_lambda) ) return 2 * RADIUS * asin(UpperCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
258
0
import logging import os from .state import PartialState class lowerCAmelCase__( logging.LoggerAdapter ): '''simple docstring''' @staticmethod def UpperCamelCase_ ( __lowerCamelCase ) -> Union[str, Any]: _SCREAMING_SNAKE_CASE : Optional[Any] = PartialState() return not main_process_only or (main_process_only and state.is_main_process) def UpperCamelCase_ ( self , __lowerCamelCase , __lowerCamelCase , *__lowerCamelCase , **__lowerCamelCase ) -> Tuple: if PartialState._shared_state == {}: raise RuntimeError( "You must initialize the accelerate state by calling either `PartialState()` or `Accelerator()` before using the logging utility." ) _SCREAMING_SNAKE_CASE : int = kwargs.pop("main_process_only" , __lowerCamelCase ) _SCREAMING_SNAKE_CASE : Optional[Any] = kwargs.pop("in_order" , __lowerCamelCase ) if self.isEnabledFor(__lowerCamelCase ): if self._should_log(__lowerCamelCase ): _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : str = self.process(__lowerCamelCase , __lowerCamelCase ) self.logger.log(__lowerCamelCase , __lowerCamelCase , *__lowerCamelCase , **__lowerCamelCase ) elif in_order: _SCREAMING_SNAKE_CASE : Tuple = PartialState() for i in range(state.num_processes ): if i == state.process_index: _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE : Optional[int] = self.process(__lowerCamelCase , __lowerCamelCase ) self.logger.log(__lowerCamelCase , __lowerCamelCase , *__lowerCamelCase , **__lowerCamelCase ) state.wait_for_everyone() def lowerCamelCase__ (__lowerCamelCase, __lowerCamelCase = None ): if log_level is None: _SCREAMING_SNAKE_CASE : Tuple = os.environ.get("ACCELERATE_LOG_LEVEL", __lowerCamelCase ) _SCREAMING_SNAKE_CASE : List[str] = logging.getLogger(__lowerCamelCase ) if log_level is not None: logger.setLevel(log_level.upper() ) logger.root.setLevel(log_level.upper() ) return MultiProcessAdapter(__lowerCamelCase, {} )
325
from math import factorial def lowerCamelCase__ (__lowerCamelCase, __lowerCamelCase ): # If either of the conditions are true, the function is being asked # to calculate a factorial of a negative number, which is not possible if n < k or k < 0: raise ValueError("Please enter positive integers for n and k where n >= k" ) return factorial(__lowerCamelCase ) // (factorial(__lowerCamelCase ) * factorial(n - k )) if __name__ == "__main__": print( 'The number of five-card hands possible from a standard', f"fifty-two card deck is: {combinations(52, 5)}\n", ) print( 'If a class of 40 students must be arranged into groups of', f"4 for group projects, there are {combinations(40, 4)} ways", 'to arrange them.\n', ) print( 'If 10 teams are competing in a Formula One race, there', f"are {combinations(10, 3)} ways that first, second and", 'third place can be awarded.', )
325
1
import logging from transformers import PretrainedConfig __A : Union[str, Any] = logging.getLogger(__name__) __A : List[str] = { "bertabs-finetuned-cnndm": "https://huggingface.co/remi/bertabs-finetuned-cnndm-extractive-abstractive-summarization/resolve/main/config.json", } class __A ( UpperCamelCase_ ): lowerCAmelCase_ : str = """bertabs""" def __init__( self : Any , UpperCAmelCase_ : Union[str, Any]=30522 , UpperCAmelCase_ : Optional[int]=512 , UpperCAmelCase_ : List[Any]=6 , UpperCAmelCase_ : List[Any]=512 , UpperCAmelCase_ : Optional[int]=8 , UpperCAmelCase_ : str=512 , UpperCAmelCase_ : int=0.2 , UpperCAmelCase_ : int=6 , UpperCAmelCase_ : Dict=768 , UpperCAmelCase_ : Any=8 , UpperCAmelCase_ : List[Any]=2048 , UpperCAmelCase_ : Union[str, Any]=0.2 , **UpperCAmelCase_ : int , ): super().__init__(**_a ) lowerCAmelCase : str = vocab_size lowerCAmelCase : Optional[Any] = max_pos lowerCAmelCase : Tuple = enc_layers lowerCAmelCase : Optional[Any] = enc_hidden_size lowerCAmelCase : Optional[Any] = enc_heads lowerCAmelCase : Union[str, Any] = enc_ff_size lowerCAmelCase : Optional[Any] = enc_dropout lowerCAmelCase : str = dec_layers lowerCAmelCase : Optional[Any] = dec_hidden_size lowerCAmelCase : Optional[int] = dec_heads lowerCAmelCase : Tuple = dec_ff_size lowerCAmelCase : List[Any] = dec_dropout
138
"""simple docstring""" import inspect from typing import Optional, Union import numpy as np import PIL import torch from torch.nn import functional as F from torchvision import transforms from transformers import CLIPFeatureExtractor, CLIPModel, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, DPMSolverMultistepScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.utils import ( PIL_INTERPOLATION, randn_tensor, ) def _lowercase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Union[str, Any]: if isinstance(__lowerCAmelCase , torch.Tensor ): return image elif isinstance(__lowerCAmelCase , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ : Any = [image] if isinstance(image[0] , PIL.Image.Image ): SCREAMING_SNAKE_CASE__ : List[str] = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION["""lanczos"""] ) )[None, :] for i in image] SCREAMING_SNAKE_CASE__ : List[Any] = np.concatenate(__lowerCAmelCase , axis=0 ) SCREAMING_SNAKE_CASE__ : List[str] = np.array(__lowerCAmelCase ).astype(np.floataa ) / 255.0 SCREAMING_SNAKE_CASE__ : Optional[Any] = image.transpose(0 , 3 , 1 , 2 ) SCREAMING_SNAKE_CASE__ : Tuple = 2.0 * image - 1.0 SCREAMING_SNAKE_CASE__ : Tuple = torch.from_numpy(__lowerCAmelCase ) elif isinstance(image[0] , torch.Tensor ): SCREAMING_SNAKE_CASE__ : Tuple = torch.cat(__lowerCAmelCase , dim=0 ) return image def _lowercase ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=0.9_995 ) -> Union[str, Any]: if not isinstance(__lowerCAmelCase , np.ndarray ): SCREAMING_SNAKE_CASE__ : Dict = True SCREAMING_SNAKE_CASE__ : int = va.device SCREAMING_SNAKE_CASE__ : str = va.cpu().numpy() SCREAMING_SNAKE_CASE__ : str = va.cpu().numpy() SCREAMING_SNAKE_CASE__ : Any = np.sum(va * va / (np.linalg.norm(__lowerCAmelCase ) * np.linalg.norm(__lowerCAmelCase )) ) if np.abs(__lowerCAmelCase ) > DOT_THRESHOLD: SCREAMING_SNAKE_CASE__ : Tuple = (1 - t) * va + t * va else: SCREAMING_SNAKE_CASE__ : Optional[int] = np.arccos(__lowerCAmelCase ) SCREAMING_SNAKE_CASE__ : Optional[Any] = np.sin(__lowerCAmelCase ) SCREAMING_SNAKE_CASE__ : str = theta_a * t SCREAMING_SNAKE_CASE__ : Tuple = np.sin(__lowerCAmelCase ) SCREAMING_SNAKE_CASE__ : List[str] = np.sin(theta_a - theta_t ) / sin_theta_a SCREAMING_SNAKE_CASE__ : Optional[int] = sin_theta_t / sin_theta_a SCREAMING_SNAKE_CASE__ : List[Any] = sa * va + sa * va if inputs_are_torch: SCREAMING_SNAKE_CASE__ : str = torch.from_numpy(__lowerCAmelCase ).to(__lowerCAmelCase ) return va def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> Any: SCREAMING_SNAKE_CASE__ : Tuple = F.normalize(__lowerCAmelCase , dim=-1 ) SCREAMING_SNAKE_CASE__ : Optional[int] = F.normalize(__lowerCAmelCase , dim=-1 ) return (x - y).norm(dim=-1 ).div(2 ).arcsin().pow(2 ).mul(2 ) def _lowercase ( __lowerCAmelCase , __lowerCAmelCase ) -> Union[str, Any]: for param in model.parameters(): SCREAMING_SNAKE_CASE__ : int = value class __a (UpperCamelCase_): '''simple docstring''' def __init__( self , _a , _a , _a , _a , _a , _a , _a , _a=None , _a=None , _a=None , ) -> Union[str, Any]: """simple docstring""" super().__init__() self.register_modules( vae=_a , text_encoder=_a , clip_model=_a , tokenizer=_a , unet=_a , scheduler=_a , feature_extractor=_a , coca_model=_a , coca_tokenizer=_a , coca_transform=_a , ) SCREAMING_SNAKE_CASE__ : Optional[Any] = ( feature_extractor.size if isinstance(feature_extractor.size , _a ) else feature_extractor.size["""shortest_edge"""] ) SCREAMING_SNAKE_CASE__ : List[Any] = transforms.Normalize(mean=feature_extractor.image_mean , std=feature_extractor.image_std ) set_requires_grad(self.text_encoder , _a ) set_requires_grad(self.clip_model , _a ) def _a ( self , _a = "auto" ) -> Dict: """simple docstring""" if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(_a ) def _a ( self ) -> List[str]: """simple docstring""" self.enable_attention_slicing(_a ) def _a ( self ) -> List[Any]: """simple docstring""" set_requires_grad(self.vae , _a ) def _a ( self ) -> Dict: """simple docstring""" set_requires_grad(self.vae , _a ) def _a ( self ) -> Optional[Any]: """simple docstring""" set_requires_grad(self.unet , _a ) def _a ( self ) -> int: """simple docstring""" set_requires_grad(self.unet , _a ) def _a ( self , _a , _a , _a ) -> Tuple: """simple docstring""" SCREAMING_SNAKE_CASE__ : Dict = min(int(num_inference_steps * strength ) , _a ) SCREAMING_SNAKE_CASE__ : Optional[int] = max(num_inference_steps - init_timestep , 0 ) SCREAMING_SNAKE_CASE__ : Dict = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def _a ( self , _a , _a , _a , _a , _a , _a=None ) -> Optional[Any]: """simple docstring""" if not isinstance(_a , torch.Tensor ): raise ValueError(f'''`image` has to be of type `torch.Tensor` but is {type(_a )}''' ) SCREAMING_SNAKE_CASE__ : Optional[Any] = image.to(device=_a , dtype=_a ) if isinstance(_a , _a ): SCREAMING_SNAKE_CASE__ : int = [ self.vae.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(_a ) ] SCREAMING_SNAKE_CASE__ : Tuple = torch.cat(_a , dim=0 ) else: SCREAMING_SNAKE_CASE__ : int = self.vae.encode(_a ).latent_dist.sample(_a ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor SCREAMING_SNAKE_CASE__ : Optional[Any] = 0.18_215 * init_latents SCREAMING_SNAKE_CASE__ : List[str] = init_latents.repeat_interleave(_a , dim=0 ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = randn_tensor(init_latents.shape , generator=_a , device=_a , dtype=_a ) # get latents SCREAMING_SNAKE_CASE__ : Any = self.scheduler.add_noise(_a , _a , _a ) SCREAMING_SNAKE_CASE__ : Optional[Any] = init_latents return latents def _a ( self , _a ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.coca_transform(_a ).unsqueeze(0 ) with torch.no_grad(), torch.cuda.amp.autocast(): SCREAMING_SNAKE_CASE__ : List[Any] = self.coca_model.generate(transformed_image.to(device=self.device , dtype=self.coca_model.dtype ) ) SCREAMING_SNAKE_CASE__ : str = self.coca_tokenizer.decode(generated[0].cpu().numpy() ) return generated.split("""<end_of_text>""" )[0].replace("""<start_of_text>""" , """""" ).rstrip(""" .,""" ) def _a ( self , _a , _a ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : str = self.feature_extractor.preprocess(_a ) SCREAMING_SNAKE_CASE__ : str = torch.from_numpy(clip_image_input["""pixel_values"""][0] ).unsqueeze(0 ).to(self.device ).half() SCREAMING_SNAKE_CASE__ : Any = self.clip_model.get_image_features(_a ) SCREAMING_SNAKE_CASE__ : int = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=_a ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = image_embeddings_clip.repeat_interleave(_a , dim=0 ) return image_embeddings_clip @torch.enable_grad() def _a ( self , _a , _a , _a , _a , _a , _a , _a , ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE__ : int = latents.detach().requires_grad_() SCREAMING_SNAKE_CASE__ : str = self.scheduler.scale_model_input(_a , _a ) # predict the noise residual SCREAMING_SNAKE_CASE__ : Any = self.unet(_a , _a , encoder_hidden_states=_a ).sample if isinstance(self.scheduler , (PNDMScheduler, DDIMScheduler, DPMSolverMultistepScheduler) ): SCREAMING_SNAKE_CASE__ : Optional[Any] = self.scheduler.alphas_cumprod[timestep] SCREAMING_SNAKE_CASE__ : List[Any] = 1 - alpha_prod_t # compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf SCREAMING_SNAKE_CASE__ : Optional[Any] = (latents - beta_prod_t ** 0.5 * noise_pred) / alpha_prod_t ** 0.5 SCREAMING_SNAKE_CASE__ : List[str] = torch.sqrt(_a ) SCREAMING_SNAKE_CASE__ : Dict = pred_original_sample * (fac) + latents * (1 - fac) elif isinstance(self.scheduler , _a ): SCREAMING_SNAKE_CASE__ : Optional[int] = self.scheduler.sigmas[index] SCREAMING_SNAKE_CASE__ : Dict = latents - sigma * noise_pred else: raise ValueError(f'''scheduler type {type(self.scheduler )} not supported''' ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor SCREAMING_SNAKE_CASE__ : Optional[Any] = 1 / 0.18_215 * sample SCREAMING_SNAKE_CASE__ : Optional[Any] = self.vae.decode(_a ).sample SCREAMING_SNAKE_CASE__ : Any = (image / 2 + 0.5).clamp(0 , 1 ) SCREAMING_SNAKE_CASE__ : Any = transforms.Resize(self.feature_extractor_size )(_a ) SCREAMING_SNAKE_CASE__ : Dict = self.normalize(_a ).to(latents.dtype ) SCREAMING_SNAKE_CASE__ : Tuple = self.clip_model.get_image_features(_a ) SCREAMING_SNAKE_CASE__ : int = image_embeddings_clip / image_embeddings_clip.norm(p=2 , dim=-1 , keepdim=_a ) SCREAMING_SNAKE_CASE__ : Union[str, Any] = spherical_dist_loss(_a , _a ).mean() * clip_guidance_scale SCREAMING_SNAKE_CASE__ : Optional[Any] = -torch.autograd.grad(_a , _a )[0] if isinstance(self.scheduler , _a ): SCREAMING_SNAKE_CASE__ : Any = latents.detach() + grads * (sigma**2) SCREAMING_SNAKE_CASE__ : Optional[int] = noise_pred_original else: SCREAMING_SNAKE_CASE__ : Union[str, Any] = noise_pred_original - torch.sqrt(_a ) * grads return noise_pred, latents @torch.no_grad() def __call__( self , _a , _a , _a = None , _a = None , _a = 512 , _a = 512 , _a = 0.6 , _a = 50 , _a = 7.5 , _a = 1 , _a = 0.0 , _a = 100 , _a = None , _a = "pil" , _a = True , _a = 0.8 , _a = 0.1 , _a = 0.1 , ) -> int: """simple docstring""" if isinstance(_a , _a ) and len(_a ) != batch_size: raise ValueError(f'''You have passed {batch_size} batch_size, but only {len(_a )} generators.''' ) if height % 8 != 0 or width % 8 != 0: raise ValueError(f'''`height` and `width` have to be divisible by 8 but are {height} and {width}.''' ) if isinstance(_a , torch.Generator ) and batch_size > 1: SCREAMING_SNAKE_CASE__ : Optional[Any] = [generator] + [None] * (batch_size - 1) SCREAMING_SNAKE_CASE__ : List[Any] = [ ("""model""", self.coca_model is None), ("""tokenizer""", self.coca_tokenizer is None), ("""transform""", self.coca_transform is None), ] SCREAMING_SNAKE_CASE__ : Optional[int] = [x[0] for x in coca_is_none if x[1]] SCREAMING_SNAKE_CASE__ : Union[str, Any] = """, """.join(_a ) # generate prompts with coca model if prompt is None if content_prompt is None: if len(_a ): raise ValueError( f'''Content prompt is None and CoCa [{coca_is_none_str}] is None.''' f'''Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.''' ) SCREAMING_SNAKE_CASE__ : Any = self.get_image_description(_a ) if style_prompt is None: if len(_a ): raise ValueError( f'''Style prompt is None and CoCa [{coca_is_none_str}] is None.''' f''' Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.''' ) SCREAMING_SNAKE_CASE__ : Tuple = self.get_image_description(_a ) # get prompt text embeddings for content and style SCREAMING_SNAKE_CASE__ : Any = self.tokenizer( _a , padding="""max_length""" , max_length=self.tokenizer.model_max_length , truncation=_a , return_tensors="""pt""" , ) SCREAMING_SNAKE_CASE__ : Any = self.text_encoder(content_text_input.input_ids.to(self.device ) )[0] SCREAMING_SNAKE_CASE__ : Dict = self.tokenizer( _a , padding="""max_length""" , max_length=self.tokenizer.model_max_length , truncation=_a , return_tensors="""pt""" , ) SCREAMING_SNAKE_CASE__ : List[str] = self.text_encoder(style_text_input.input_ids.to(self.device ) )[0] SCREAMING_SNAKE_CASE__ : Union[str, Any] = slerp(_a , _a , _a ) # duplicate text embeddings for each generation per prompt SCREAMING_SNAKE_CASE__ : int = text_embeddings.repeat_interleave(_a , dim=0 ) # set timesteps SCREAMING_SNAKE_CASE__ : Union[str, Any] = """offset""" in set(inspect.signature(self.scheduler.set_timesteps ).parameters.keys() ) SCREAMING_SNAKE_CASE__ : Tuple = {} if accepts_offset: SCREAMING_SNAKE_CASE__ : List[str] = 1 self.scheduler.set_timesteps(_a , **_a ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand self.scheduler.timesteps.to(self.device ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Union[str, Any] = self.get_timesteps(_a , _a , self.device ) SCREAMING_SNAKE_CASE__ : List[str] = timesteps[:1].repeat(_a ) # Preprocess image SCREAMING_SNAKE_CASE__ : str = preprocess(_a , _a , _a ) SCREAMING_SNAKE_CASE__ : Dict = self.prepare_latents( _a , _a , _a , text_embeddings.dtype , self.device , _a ) SCREAMING_SNAKE_CASE__ : List[Any] = preprocess(_a , _a , _a ) SCREAMING_SNAKE_CASE__ : Any = self.prepare_latents( _a , _a , _a , text_embeddings.dtype , self.device , _a ) SCREAMING_SNAKE_CASE__ : List[Any] = slerp(_a , _a , _a ) if clip_guidance_scale > 0: SCREAMING_SNAKE_CASE__ : List[str] = self.get_clip_image_embeddings(_a , _a ) SCREAMING_SNAKE_CASE__ : List[Any] = self.get_clip_image_embeddings(_a , _a ) SCREAMING_SNAKE_CASE__ : Dict = slerp( _a , _a , _a ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. SCREAMING_SNAKE_CASE__ : str = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: SCREAMING_SNAKE_CASE__ : Union[str, Any] = content_text_input.input_ids.shape[-1] SCREAMING_SNAKE_CASE__ : str = self.tokenizer([""""""] , padding="""max_length""" , max_length=_a , return_tensors="""pt""" ) SCREAMING_SNAKE_CASE__ : Tuple = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt SCREAMING_SNAKE_CASE__ : Tuple = uncond_embeddings.repeat_interleave(_a , dim=0 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes SCREAMING_SNAKE_CASE__ : Union[str, Any] = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. SCREAMING_SNAKE_CASE__ : Dict = (batch_size, self.unet.config.in_channels, height // 8, width // 8) SCREAMING_SNAKE_CASE__ : Any = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not work reproducibly on mps SCREAMING_SNAKE_CASE__ : List[str] = torch.randn(_a , generator=_a , device="""cpu""" , dtype=_a ).to( self.device ) else: SCREAMING_SNAKE_CASE__ : Any = torch.randn(_a , generator=_a , device=self.device , dtype=_a ) else: if latents.shape != latents_shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {latents_shape}''' ) SCREAMING_SNAKE_CASE__ : Optional[Any] = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler SCREAMING_SNAKE_CASE__ : List[str] = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] SCREAMING_SNAKE_CASE__ : Union[str, Any] = """eta""" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) SCREAMING_SNAKE_CASE__ : str = {} if accepts_eta: SCREAMING_SNAKE_CASE__ : Optional[Any] = eta # check if the scheduler accepts generator SCREAMING_SNAKE_CASE__ : int = """generator""" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) if accepts_generator: SCREAMING_SNAKE_CASE__ : Optional[Any] = generator with self.progress_bar(total=_a ): for i, t in enumerate(_a ): # expand the latents if we are doing classifier free guidance SCREAMING_SNAKE_CASE__ : List[Any] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents SCREAMING_SNAKE_CASE__ : List[str] = self.scheduler.scale_model_input(_a , _a ) # predict the noise residual SCREAMING_SNAKE_CASE__ : List[Any] = self.unet(_a , _a , encoder_hidden_states=_a ).sample # perform classifier free guidance if do_classifier_free_guidance: SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Optional[Any] = noise_pred.chunk(2 ) SCREAMING_SNAKE_CASE__ : Optional[Any] = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # perform clip guidance if clip_guidance_scale > 0: SCREAMING_SNAKE_CASE__ : List[Any] = ( text_embeddings.chunk(2 )[1] if do_classifier_free_guidance else text_embeddings ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[str] = self.cond_fn( _a , _a , _a , _a , _a , _a , _a , ) # compute the previous noisy sample x_t -> x_t-1 SCREAMING_SNAKE_CASE__ : Any = self.scheduler.step(_a , _a , _a , **_a ).prev_sample # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor SCREAMING_SNAKE_CASE__ : List[Any] = 1 / 0.18_215 * latents SCREAMING_SNAKE_CASE__ : int = self.vae.decode(_a ).sample SCREAMING_SNAKE_CASE__ : str = (image / 2 + 0.5).clamp(0 , 1 ) SCREAMING_SNAKE_CASE__ : Optional[Any] = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": SCREAMING_SNAKE_CASE__ : int = self.numpy_to_pil(_a ) if not return_dict: return (image, None) return StableDiffusionPipelineOutput(images=_a , nsfw_content_detected=_a )
132
0
def UpperCAmelCase__ ( UpperCAmelCase__, UpperCAmelCase__ ) -> Optional[Any]: A_ = len(UpperCAmelCase__ ) A_ = len(UpperCAmelCase__ ) A_ = [[False for _ in range(m + 1 )] for _ in range(n + 1 )] A_ = True for i in range(UpperCAmelCase__ ): for j in range(m + 1 ): if dp[i][j]: if j < m and a[i].upper() == b[j]: A_ = True if a[i].islower(): A_ = True return dp[n][m] if __name__ == "__main__": import doctest doctest.testmod()
361
'''simple docstring''' import copy import unittest from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_MULTIPLE_CHOICE_MAPPING, MODEL_FOR_QUESTION_ANSWERING_MAPPING, MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, LayoutLMvaForQuestionAnswering, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaModel, ) from transformers.models.layoutlmva.modeling_layoutlmva import LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class A__ : def __init__( self , UpperCamelCase__ , UpperCamelCase__=2 , UpperCamelCase__=3 , UpperCamelCase__=4 , UpperCamelCase__=2 , UpperCamelCase__=7 , UpperCamelCase__=True , UpperCamelCase__=True , UpperCamelCase__=True , UpperCamelCase__=True , UpperCamelCase__=99 , UpperCamelCase__=36 , UpperCamelCase__=3 , UpperCamelCase__=4 , UpperCamelCase__=37 , UpperCamelCase__="gelu" , UpperCamelCase__=0.1 , UpperCamelCase__=0.1 , UpperCamelCase__=512 , UpperCamelCase__=16 , UpperCamelCase__=2 , UpperCamelCase__=0.02 , UpperCamelCase__=6 , UpperCamelCase__=6 , UpperCamelCase__=3 , UpperCamelCase__=4 , UpperCamelCase__=None , UpperCamelCase__=1000 , ) -> Optional[int]: '''simple docstring''' A_ = parent A_ = batch_size A_ = num_channels A_ = image_size A_ = patch_size A_ = text_seq_length A_ = is_training A_ = use_input_mask A_ = use_token_type_ids A_ = use_labels A_ = vocab_size A_ = hidden_size A_ = num_hidden_layers A_ = num_attention_heads A_ = intermediate_size A_ = hidden_act A_ = hidden_dropout_prob A_ = attention_probs_dropout_prob A_ = max_position_embeddings A_ = type_vocab_size A_ = type_sequence_label_size A_ = initializer_range A_ = coordinate_size A_ = shape_size A_ = num_labels A_ = num_choices A_ = scope A_ = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) A_ = text_seq_length A_ = (image_size // patch_size) ** 2 + 1 A_ = self.text_seq_length + self.image_seq_length def snake_case_ ( self ) -> Any: '''simple docstring''' A_ = ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) A_ = ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: A_ = bbox[i, j, 3] A_ = bbox[i, j, 1] A_ = t if bbox[i, j, 2] < bbox[i, j, 0]: A_ = bbox[i, j, 2] A_ = bbox[i, j, 0] A_ = t A_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ = None if self.use_input_mask: A_ = random_attention_mask([self.batch_size, self.text_seq_length] ) A_ = None if self.use_token_type_ids: A_ = ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) A_ = None A_ = None if self.use_labels: A_ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A_ = ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) A_ = LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> str: '''simple docstring''' A_ = LayoutLMvaModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() # text + image A_ = model(UpperCamelCase__ , pixel_values=UpperCamelCase__ ) A_ = model( UpperCamelCase__ , bbox=UpperCamelCase__ , pixel_values=UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ ) A_ = model(UpperCamelCase__ , bbox=UpperCamelCase__ , pixel_values=UpperCamelCase__ , token_type_ids=UpperCamelCase__ ) A_ = model(UpperCamelCase__ , bbox=UpperCamelCase__ , pixel_values=UpperCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only A_ = model(UpperCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only A_ = model(pixel_values=UpperCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Optional[Any]: '''simple docstring''' A_ = self.num_labels A_ = LayoutLMvaForSequenceClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A_ = model( UpperCamelCase__ , bbox=UpperCamelCase__ , pixel_values=UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , labels=UpperCamelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> str: '''simple docstring''' A_ = self.num_labels A_ = LayoutLMvaForTokenClassification(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A_ = model( UpperCamelCase__ , bbox=UpperCamelCase__ , pixel_values=UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , labels=UpperCamelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> int: '''simple docstring''' A_ = LayoutLMvaForQuestionAnswering(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() A_ = model( UpperCamelCase__ , bbox=UpperCamelCase__ , pixel_values=UpperCamelCase__ , attention_mask=UpperCamelCase__ , token_type_ids=UpperCamelCase__ , start_positions=UpperCamelCase__ , end_positions=UpperCamelCase__ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def snake_case_ ( self ) -> int: '''simple docstring''' A_ = self.prepare_config_and_inputs() ( ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ( A_ ) , ) = config_and_inputs A_ = { """input_ids""": input_ids, """bbox""": bbox, """pixel_values""": pixel_values, """token_type_ids""": token_type_ids, """attention_mask""": input_mask, } return config, inputs_dict @require_torch class A__ ( _snake_case , _snake_case , unittest.TestCase ): lowercase = False lowercase = False lowercase = False lowercase = ( ( LayoutLMvaModel, LayoutLMvaForSequenceClassification, LayoutLMvaForTokenClassification, LayoutLMvaForQuestionAnswering, ) if is_torch_available() else () ) lowercase = ( {"document-question-answering": LayoutLMvaForQuestionAnswering, "feature-extraction": LayoutLMvaModel} if is_torch_available() else {} ) def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Dict: '''simple docstring''' # `DocumentQuestionAnsweringPipeline` is expected to work with this model, but it combines the text and visual # embedding along the sequence dimension (dim 1), which causes an error during post-processing as `p_mask` has # the sequence dimension of the text embedding only. # (see the line `embedding_output = torch.cat([embedding_output, visual_embeddings], dim=1)`) return True def snake_case_ ( self ) -> str: '''simple docstring''' A_ = LayoutLMvaModelTester(self ) A_ = ConfigTester(self , config_class=UpperCamelCase__ , hidden_size=37 ) def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__=False ) -> Union[str, Any]: '''simple docstring''' A_ = copy.deepcopy(UpperCamelCase__ ) if model_class in get_values(UpperCamelCase__ ): A_ = { k: v.unsqueeze(1 ).expand(-1 , self.model_tester.num_choices , -1 ).contiguous() if isinstance(UpperCamelCase__ , torch.Tensor ) and v.ndim > 1 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(UpperCamelCase__ ): A_ = torch.ones(self.model_tester.batch_size , dtype=torch.long , device=UpperCamelCase__ ) elif model_class in get_values(UpperCamelCase__ ): A_ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=UpperCamelCase__ ) A_ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=UpperCamelCase__ ) elif model_class in [ *get_values(UpperCamelCase__ ), ]: A_ = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=UpperCamelCase__ ) elif model_class in [ *get_values(UpperCamelCase__ ), ]: A_ = torch.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=torch.long , device=UpperCamelCase__ , ) return inputs_dict def snake_case_ ( self ) -> Optional[Any]: '''simple docstring''' self.config_tester.run_common_tests() def snake_case_ ( self ) -> Dict: '''simple docstring''' A_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def snake_case_ ( self ) -> Dict: '''simple docstring''' A_ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A_ = type self.model_tester.create_and_check_model(*UpperCamelCase__ ) def snake_case_ ( self ) -> Dict: '''simple docstring''' A_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*UpperCamelCase__ ) def snake_case_ ( self ) -> List[str]: '''simple docstring''' A_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCamelCase__ ) def snake_case_ ( self ) -> Tuple: '''simple docstring''' A_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCamelCase__ ) @slow def snake_case_ ( self ) -> Union[str, Any]: '''simple docstring''' for model_name in LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ = LayoutLMvaModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def UpperCAmelCase__ ( ) -> Dict: A_ = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch class A__ ( unittest.TestCase ): @cached_property def snake_case_ ( self ) -> List[Any]: '''simple docstring''' return LayoutLMvaImageProcessor(apply_ocr=UpperCamelCase__ ) if is_vision_available() else None @slow def snake_case_ ( self ) -> Optional[Any]: '''simple docstring''' A_ = LayoutLMvaModel.from_pretrained("""microsoft/layoutlmv3-base""" ).to(UpperCamelCase__ ) A_ = self.default_image_processor A_ = prepare_img() A_ = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ).pixel_values.to(UpperCamelCase__ ) A_ = torch.tensor([[1, 2]] ) A_ = torch.tensor([[1, 2, 3, 4], [5, 6, 7, 8]] ).unsqueeze(0 ) # forward pass A_ = model( input_ids=input_ids.to(UpperCamelCase__ ) , bbox=bbox.to(UpperCamelCase__ ) , pixel_values=pixel_values.to(UpperCamelCase__ ) , ) # verify the logits A_ = torch.Size((1, 199, 768) ) self.assertEqual(outputs.last_hidden_state.shape , UpperCamelCase__ ) A_ = torch.tensor( [[-0.0529, 0.3618, 0.1632], [-0.1587, -0.1667, -0.0400], [-0.1557, -0.1671, -0.0505]] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , UpperCamelCase__ , atol=1e-4 ) )
101
0
"""simple docstring""" import multiprocessing from typing import TYPE_CHECKING, Optional, Union from .. import Dataset, Features, config from ..formatting import query_table from ..packaged_modules.sql.sql import Sql from ..utils import logging from .abc import AbstractDatasetInputStream if TYPE_CHECKING: import sqlitea import sqlalchemy class UpperCamelCase ( lowerCAmelCase__ ): def __init__( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__ = None, lowerCAmelCase__ = None, lowerCAmelCase__ = False, **lowerCAmelCase__, ) -> Dict: super().__init__(features=lowerCAmelCase__, cache_dir=lowerCAmelCase__, keep_in_memory=lowerCAmelCase__, **lowerCAmelCase__) snake_case_ = Sql( cache_dir=lowerCAmelCase__, features=lowerCAmelCase__, sql=lowerCAmelCase__, con=lowerCAmelCase__, **lowerCAmelCase__, ) def a_ ( self) -> int: snake_case_ = None snake_case_ = None snake_case_ = None snake_case_ = None self.builder.download_and_prepare( download_config=lowerCAmelCase__, download_mode=lowerCAmelCase__, verification_mode=lowerCAmelCase__, base_path=lowerCAmelCase__, ) # Build dataset for splits snake_case_ = self.builder.as_dataset( split='train', verification_mode=lowerCAmelCase__, in_memory=self.keep_in_memory) return dataset class UpperCamelCase : def __init__( self, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__ = None, lowerCAmelCase__ = None, **lowerCAmelCase__, ) -> Optional[Any]: if num_proc is not None and num_proc <= 0: raise ValueError(f'num_proc {num_proc} must be an integer > 0.') snake_case_ = dataset snake_case_ = name snake_case_ = con snake_case_ = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE snake_case_ = num_proc snake_case_ = to_sql_kwargs def a_ ( self) -> int: snake_case_ = self.to_sql_kwargs.pop('sql', lowerCAmelCase__) snake_case_ = self.to_sql_kwargs.pop('con', lowerCAmelCase__) snake_case_ = self.to_sql_kwargs.pop('index', lowerCAmelCase__) snake_case_ = self._write(index=lowerCAmelCase__, **self.to_sql_kwargs) return written def a_ ( self, lowerCAmelCase__) -> Any: snake_case_ , snake_case_ , snake_case_ = args snake_case_ = {**to_sql_kwargs, 'if_exists': 'append'} if offset > 0 else to_sql_kwargs snake_case_ = query_table( table=self.dataset.data, key=slice(lowerCAmelCase__, offset + self.batch_size), indices=self.dataset._indices, ) snake_case_ = batch.to_pandas() snake_case_ = df.to_sql(self.name, self.con, index=lowerCAmelCase__, **lowerCAmelCase__) return num_rows or len(lowerCAmelCase__) def a_ ( self, lowerCAmelCase__, **lowerCAmelCase__) -> int: snake_case_ = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0, len(self.dataset), self.batch_size), unit='ba', disable=not logging.is_progress_bar_enabled(), desc='Creating SQL from Arrow format', ): written += self._batch_sql((offset, index, to_sql_kwargs)) else: snake_case_ , snake_case_ = len(self.dataset), self.batch_size with multiprocessing.Pool(self.num_proc) as pool: for num_rows in logging.tqdm( pool.imap( self._batch_sql, [(offset, index, to_sql_kwargs) for offset in range(0, lowerCAmelCase__, lowerCAmelCase__)], ), total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size, unit='ba', disable=not logging.is_progress_bar_enabled(), desc='Creating SQL from Arrow format', ): written += num_rows return written
69
'''simple docstring''' import math import time from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class UpperCAmelCase ( UpperCamelCase__ ): def __init__( self :Optional[Any] , *lowercase_ :int , lowercase_ :Any=None , lowercase_ :List[str]=None , **lowercase_ :Any )-> Any: super().__init__(*lowercase_ , **lowercase_ ) A__ = eval_examples A__ = post_process_function def UpperCAmelCase_ ( self :str , lowercase_ :str=None , lowercase_ :Optional[int]=None , lowercase_ :Optional[int]=None , lowercase_ :str = "eval" )-> Union[str, Any]: A__ = self.eval_dataset if eval_dataset is None else eval_dataset A__ = self.get_eval_dataloader(lowercase_ ) A__ = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. A__ = self.compute_metrics A__ = None A__ = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop A__ = time.time() try: A__ = eval_loop( lowercase_ , description="Evaluation" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=lowercase_ , metric_key_prefix=lowercase_ , ) finally: A__ = compute_metrics A__ = self.args.eval_batch_size * self.args.world_size if F"{metric_key_prefix}_jit_compilation_time" in output.metrics: start_time += output.metrics[F"{metric_key_prefix}_jit_compilation_time"] output.metrics.update( speed_metrics( lowercase_ , lowercase_ , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default A__ = self.post_process_function(lowercase_ , lowercase_ , output.predictions ) A__ = self.compute_metrics(lowercase_ ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"{metric_key_prefix}_" ): A__ = metrics.pop(lowercase_ ) metrics.update(output.metrics ) else: A__ = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(lowercase_ ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) A__ = self.callback_handler.on_evaluate(self.args , self.state , self.control , lowercase_ ) return metrics def UpperCAmelCase_ ( self :List[str] , lowercase_ :List[Any] , lowercase_ :str , lowercase_ :Any=None , lowercase_ :str = "test" )-> List[Any]: A__ = self.get_test_dataloader(lowercase_ ) # Temporarily disable metric computation, we will do it in the loop here. A__ = self.compute_metrics A__ = None A__ = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop A__ = time.time() try: A__ = eval_loop( lowercase_ , description="Prediction" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=lowercase_ , metric_key_prefix=lowercase_ , ) finally: A__ = compute_metrics A__ = self.args.eval_batch_size * self.args.world_size if F"{metric_key_prefix}_jit_compilation_time" in output.metrics: start_time += output.metrics[F"{metric_key_prefix}_jit_compilation_time"] output.metrics.update( speed_metrics( lowercase_ , lowercase_ , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output A__ = self.post_process_function(lowercase_ , lowercase_ , output.predictions , "predict" ) A__ = self.compute_metrics(lowercase_ ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"{metric_key_prefix}_" ): A__ = metrics.pop(lowercase_ ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=lowercase_ )
237
0
import os import torch from ..logging import get_logger from .constants import FSDP_PYTORCH_VERSION, MODEL_NAME, OPTIMIZER_NAME from .versions import is_torch_version if is_torch_version(">=", FSDP_PYTORCH_VERSION): import torch.distributed.checkpoint as dist_cp from torch.distributed.checkpoint.default_planner import DefaultLoadPlanner, DefaultSavePlanner from torch.distributed.checkpoint.optimizer import load_sharded_optimizer_state_dict from torch.distributed.fsdp.fully_sharded_data_parallel import FullyShardedDataParallel as FSDP from torch.distributed.fsdp.fully_sharded_data_parallel import StateDictType _a : Any= get_logger(__name__) def __UpperCAmelCase ( UpperCAmelCase_ : int , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : str=0 ) -> int: '''simple docstring''' os.makedirs(a__ , exist_ok=a__ ) with FSDP.state_dict_type( a__ , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): __snake_case : List[Any] = model.state_dict() if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: __snake_case : Optional[int] = F"{MODEL_NAME}.bin" if model_index == 0 else F"{MODEL_NAME}_{model_index}.bin" __snake_case : Any = os.path.join(a__ , a__ ) if accelerator.process_index == 0: logger.info(F"Saving model to {output_model_file}" ) torch.save(a__ , a__ ) logger.info(F"Model saved to {output_model_file}" ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: __snake_case : Dict = ( F"{MODEL_NAME}_rank{accelerator.process_index}.bin" if model_index == 0 else F"{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin" ) __snake_case : Optional[Any] = os.path.join(a__ , a__ ) logger.info(F"Saving model to {output_model_file}" ) torch.save(a__ , a__ ) logger.info(F"Model saved to {output_model_file}" ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: __snake_case : Union[str, Any] = os.path.join(a__ , F"{MODEL_NAME}_{model_index}" ) os.makedirs(a__ , exist_ok=a__ ) logger.info(F"Saving model to {ckpt_dir}" ) __snake_case : List[Any] = {'model': state_dict} dist_cp.save_state_dict( state_dict=a__ , storage_writer=dist_cp.FileSystemWriter(a__ ) , planner=DefaultSavePlanner() , ) logger.info(F"Model saved to {ckpt_dir}" ) def __UpperCAmelCase ( UpperCAmelCase_ : Dict , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : int=0 ) -> Union[str, Any]: '''simple docstring''' accelerator.wait_for_everyone() with FSDP.state_dict_type( a__ , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if type(a__ ) != FSDP and accelerator.process_index != 0: if not fsdp_plugin.sync_module_states: raise ValueError( 'Set the `sync_module_states` flag to `True` so that model states are synced across processes when ' 'initializing FSDP object' ) return __snake_case : Optional[int] = F"{MODEL_NAME}.bin" if model_index == 0 else F"{MODEL_NAME}_{model_index}.bin" __snake_case : List[Any] = os.path.join(a__ , a__ ) logger.info(F"Loading model from {input_model_file}" ) __snake_case : Any = torch.load(a__ ) logger.info(F"Model loaded from {input_model_file}" ) elif fsdp_plugin.state_dict_type == StateDictType.LOCAL_STATE_DICT: __snake_case : Optional[Any] = ( F"{MODEL_NAME}_rank{accelerator.process_index}.bin" if model_index == 0 else F"{MODEL_NAME}_{model_index}_rank{accelerator.process_index}.bin" ) __snake_case : Any = os.path.join(a__ , a__ ) logger.info(F"Loading model from {input_model_file}" ) __snake_case : Optional[int] = torch.load(a__ ) logger.info(F"Model loaded from {input_model_file}" ) elif fsdp_plugin.state_dict_type == StateDictType.SHARDED_STATE_DICT: __snake_case : Dict = ( os.path.join(a__ , F"{MODEL_NAME}_{model_index}" ) if F"{MODEL_NAME}" not in input_dir else input_dir ) logger.info(F"Loading model from {ckpt_dir}" ) __snake_case : Any = {'model': model.state_dict()} dist_cp.load_state_dict( state_dict=a__ , storage_reader=dist_cp.FileSystemReader(a__ ) , planner=DefaultLoadPlanner() , ) __snake_case : Any = state_dict['model'] logger.info(F"Model loaded from {ckpt_dir}" ) model.load_state_dict(a__ ) def __UpperCAmelCase ( UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Dict , UpperCAmelCase_ : Optional[Any]=0 ) -> List[str]: '''simple docstring''' os.makedirs(a__ , exist_ok=a__ ) with FSDP.state_dict_type( a__ , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): __snake_case : Union[str, Any] = FSDP.optim_state_dict(a__ , a__ ) if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: if accelerator.process_index == 0: __snake_case : Optional[int] = ( F"{OPTIMIZER_NAME}.bin" if optimizer_index == 0 else F"{OPTIMIZER_NAME}_{optimizer_index}.bin" ) __snake_case : Tuple = os.path.join(a__ , a__ ) logger.info(F"Saving Optimizer state to {output_optimizer_file}" ) torch.save(a__ , a__ ) logger.info(F"Optimizer state saved in {output_optimizer_file}" ) else: __snake_case : Tuple = os.path.join(a__ , F"{OPTIMIZER_NAME}_{optimizer_index}" ) os.makedirs(a__ , exist_ok=a__ ) logger.info(F"Saving Optimizer state to {ckpt_dir}" ) dist_cp.save_state_dict( state_dict={'optimizer': optim_state} , storage_writer=dist_cp.FileSystemWriter(a__ ) , planner=DefaultSavePlanner() , ) logger.info(F"Optimizer state saved in {ckpt_dir}" ) def __UpperCAmelCase ( UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Any , UpperCAmelCase_ : str , UpperCAmelCase_ : Any , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : str=0 ) -> str: '''simple docstring''' accelerator.wait_for_everyone() with FSDP.state_dict_type( a__ , fsdp_plugin.state_dict_type , fsdp_plugin.state_dict_config , fsdp_plugin.optim_state_dict_config ): if fsdp_plugin.state_dict_type == StateDictType.FULL_STATE_DICT: __snake_case : List[str] = None # below check should work but currently it isn't working (mostly opytorch issue), # in the meantime disabling it at the cost of excess memory usage # if accelerator.process_index == 0 or not fsdp_plugin.optim_state_dict_config.rank0_only: __snake_case : Optional[Any] = ( F"{OPTIMIZER_NAME}.bin" if optimizer_index == 0 else F"{OPTIMIZER_NAME}_{optimizer_index}.bin" ) __snake_case : Any = os.path.join(a__ , a__ ) logger.info(F"Loading Optimizer state from {input_optimizer_file}" ) __snake_case : Optional[Any] = torch.load(a__ ) logger.info(F"Optimizer state loaded from {input_optimizer_file}" ) else: __snake_case : Tuple = ( os.path.join(a__ , F"{OPTIMIZER_NAME}_{optimizer_index}" ) if F"{OPTIMIZER_NAME}" not in input_dir else input_dir ) logger.info(F"Loading Optimizer from {ckpt_dir}" ) __snake_case : str = load_sharded_optimizer_state_dict( model_state_dict=model.state_dict() , optimizer_key='optimizer' , storage_reader=dist_cp.FileSystemReader(a__ ) , ) __snake_case : str = optim_state['optimizer'] logger.info(F"Optimizer loaded from {ckpt_dir}" ) __snake_case : Dict = FSDP.optim_state_dict_to_load(a__ , a__ , a__ ) optimizer.load_state_dict(a__ )
361
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import TransformeraDModel, VQDiffusionPipeline, VQDiffusionScheduler, VQModel from diffusers.pipelines.vq_diffusion.pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings from diffusers.utils import load_numpy, slow, torch_device from diffusers.utils.testing_utils import require_torch_gpu _a : List[str]= False class UpperCamelCase ( unittest.TestCase ): def _lowercase (self : List[str]) -> List[Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _lowercase (self : Optional[Any]) -> Union[str, Any]: return 12 @property def _lowercase (self : Dict) -> Union[str, Any]: return 12 @property def _lowercase (self : int) -> Tuple: return 32 @property def _lowercase (self : Optional[int]) -> Dict: torch.manual_seed(0) __snake_case : Any = VQModel( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=3 , num_vq_embeddings=self.num_embed , vq_embed_dim=3 , ) return model @property def _lowercase (self : List[Any]) -> Optional[int]: __snake_case : Dict = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip') return tokenizer @property def _lowercase (self : Union[str, Any]) -> Optional[int]: torch.manual_seed(0) __snake_case : Any = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModel(_A) @property def _lowercase (self : Union[str, Any]) -> Dict: torch.manual_seed(0) __snake_case : Any = 12 __snake_case : int = 12 __snake_case : List[Any] = { 'attention_bias': True, 'cross_attention_dim': 32, 'attention_head_dim': height * width, 'num_attention_heads': 1, 'num_vector_embeds': self.num_embed, 'num_embeds_ada_norm': self.num_embeds_ada_norm, 'norm_num_groups': 32, 'sample_size': width, 'activation_fn': 'geglu-approximate', } __snake_case : Union[str, Any] = TransformeraDModel(**_A) return model def _lowercase (self : Union[str, Any]) -> Dict: __snake_case : Tuple = 'cpu' __snake_case : List[str] = self.dummy_vqvae __snake_case : str = self.dummy_text_encoder __snake_case : Optional[Any] = self.dummy_tokenizer __snake_case : Dict = self.dummy_transformer __snake_case : Optional[int] = VQDiffusionScheduler(self.num_embed) __snake_case : List[Any] = LearnedClassifierFreeSamplingEmbeddings(learnable=_A) __snake_case : List[Any] = VQDiffusionPipeline( vqvae=_A , text_encoder=_A , tokenizer=_A , transformer=_A , scheduler=_A , learned_classifier_free_sampling_embeddings=_A , ) __snake_case : List[Any] = pipe.to(_A) pipe.set_progress_bar_config(disable=_A) __snake_case : Optional[Any] = 'teddy bear playing in the pool' __snake_case : str = torch.Generator(device=_A).manual_seed(0) __snake_case : Union[str, Any] = pipe([prompt] , generator=_A , num_inference_steps=2 , output_type='np') __snake_case : Optional[int] = output.images __snake_case : int = torch.Generator(device=_A).manual_seed(0) __snake_case : Tuple = pipe( [prompt] , generator=_A , output_type='np' , return_dict=_A , num_inference_steps=2)[0] __snake_case : str = image[0, -3:, -3:, -1] __snake_case : Union[str, Any] = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 24, 24, 3) __snake_case : str = np.array([0.6_551, 0.6_168, 0.5_008, 0.5_676, 0.5_659, 0.4_295, 0.6_073, 0.5_599, 0.4_992]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1E-2 def _lowercase (self : Tuple) -> Optional[int]: __snake_case : Optional[Any] = 'cpu' __snake_case : Optional[int] = self.dummy_vqvae __snake_case : List[str] = self.dummy_text_encoder __snake_case : Optional[int] = self.dummy_tokenizer __snake_case : Optional[Any] = self.dummy_transformer __snake_case : Union[str, Any] = VQDiffusionScheduler(self.num_embed) __snake_case : Optional[int] = LearnedClassifierFreeSamplingEmbeddings( learnable=_A , hidden_size=self.text_embedder_hidden_size , length=tokenizer.model_max_length) __snake_case : Union[str, Any] = VQDiffusionPipeline( vqvae=_A , text_encoder=_A , tokenizer=_A , transformer=_A , scheduler=_A , learned_classifier_free_sampling_embeddings=_A , ) __snake_case : Union[str, Any] = pipe.to(_A) pipe.set_progress_bar_config(disable=_A) __snake_case : Union[str, Any] = 'teddy bear playing in the pool' __snake_case : Optional[int] = torch.Generator(device=_A).manual_seed(0) __snake_case : Tuple = pipe([prompt] , generator=_A , num_inference_steps=2 , output_type='np') __snake_case : Optional[Any] = output.images __snake_case : str = torch.Generator(device=_A).manual_seed(0) __snake_case : Dict = pipe( [prompt] , generator=_A , output_type='np' , return_dict=_A , num_inference_steps=2)[0] __snake_case : int = image[0, -3:, -3:, -1] __snake_case : Tuple = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 24, 24, 3) __snake_case : Optional[Any] = np.array([0.6_693, 0.6_075, 0.4_959, 0.5_701, 0.5_583, 0.4_333, 0.6_171, 0.5_684, 0.4_988]) assert np.abs(image_slice.flatten() - expected_slice).max() < 2.0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice).max() < 1E-2 @slow @require_torch_gpu class UpperCamelCase ( unittest.TestCase ): def _lowercase (self : Any) -> Union[str, Any]: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def _lowercase (self : Tuple) -> Optional[int]: __snake_case : List[Any] = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/vq_diffusion/teddy_bear_pool_classifier_free_sampling.npy') __snake_case : Union[str, Any] = VQDiffusionPipeline.from_pretrained('microsoft/vq-diffusion-ithq') __snake_case : Tuple = pipeline.to(_A) pipeline.set_progress_bar_config(disable=_A) # requires GPU generator for gumbel softmax # don't use GPU generator in tests though __snake_case : Optional[int] = torch.Generator(device=_A).manual_seed(0) __snake_case : int = pipeline( 'teddy bear playing in the pool' , num_images_per_prompt=1 , generator=_A , output_type='np' , ) __snake_case : int = output.images[0] assert image.shape == (2_56, 2_56, 3) assert np.abs(expected_image - image).max() < 2.0
95
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) a_ : Tuple = { 'configuration_gpt_bigcode': ['GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GPTBigCodeConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : Union[str, Any] = [ 'GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST', 'GPTBigCodeForSequenceClassification', 'GPTBigCodeForTokenClassification', 'GPTBigCodeForCausalLM', 'GPTBigCodeModel', 'GPTBigCodePreTrainedModel', ] if TYPE_CHECKING: from .configuration_gpt_bigcode import GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTBigCodeConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_bigcode import ( GPT_BIGCODE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTBigCodeForCausalLM, GPTBigCodeForSequenceClassification, GPTBigCodeForTokenClassification, GPTBigCodeModel, GPTBigCodePreTrainedModel, ) else: import sys a_ : int = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
137
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger a_ : List[Any] = get_logger(__name__) class _snake_case ( enum.Enum ): _lowercase : Any = '''all_checks''' _lowercase : str = '''basic_checks''' _lowercase : str = '''no_checks''' class _snake_case ( A__ ): pass class _snake_case ( A__ ): pass class _snake_case ( A__ ): pass class _snake_case ( A__ ): pass def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None): if expected_checksums is None: logger.info('Unable to verify checksums.') return if len(set(_UpperCAmelCase) - set(_UpperCAmelCase)) > 0: raise ExpectedMoreDownloadedFiles(str(set(_UpperCAmelCase) - set(_UpperCAmelCase))) if len(set(_UpperCAmelCase) - set(_UpperCAmelCase)) > 0: raise UnexpectedDownloadedFile(str(set(_UpperCAmelCase) - set(_UpperCAmelCase))) SCREAMING_SNAKE_CASE = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] SCREAMING_SNAKE_CASE = ' for ' + verification_name if verification_name is not None else '' if len(_UpperCAmelCase) > 0: raise NonMatchingChecksumError( F'''Checksums didn\'t match{for_verification_name}:\n''' F'''{bad_urls}\n''' 'Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error') logger.info('All the checksums matched successfully' + for_verification_name) class _snake_case ( A__ ): pass class _snake_case ( A__ ): pass class _snake_case ( A__ ): pass class _snake_case ( A__ ): pass def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase): if expected_splits is None: logger.info('Unable to verify splits sizes.') return if len(set(_UpperCAmelCase) - set(_UpperCAmelCase)) > 0: raise ExpectedMoreSplits(str(set(_UpperCAmelCase) - set(_UpperCAmelCase))) if len(set(_UpperCAmelCase) - set(_UpperCAmelCase)) > 0: raise UnexpectedSplits(str(set(_UpperCAmelCase) - set(_UpperCAmelCase))) SCREAMING_SNAKE_CASE = [ {'expected': expected_splits[name], 'recorded': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(_UpperCAmelCase) > 0: raise NonMatchingSplitsSizesError(str(_UpperCAmelCase)) logger.info('All the splits matched successfully.') def lowerCamelCase__ (_UpperCAmelCase , _UpperCAmelCase = True): if record_checksum: SCREAMING_SNAKE_CASE = shaaaa() with open(_UpperCAmelCase , 'rb') as f: for chunk in iter(lambda: f.read(1 << 20) , B''): m.update(_UpperCAmelCase) SCREAMING_SNAKE_CASE = m.hexdigest() else: SCREAMING_SNAKE_CASE = None return {"num_bytes": os.path.getsize(_UpperCAmelCase), "checksum": checksum} def lowerCamelCase__ (_UpperCAmelCase): if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
137
1
import inspect import unittest from transformers import RegNetConfig, is_flax_available from transformers.testing_utils import require_flax, slow from transformers.utils import cached_property, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.models.regnet.modeling_flax_regnet import FlaxRegNetForImageClassification, FlaxRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def __init__( self , _A , _A=3 , _A=32 , _A=3 , _A=10 , _A=[10, 20, 30, 40] , _A=[1, 1, 2, 1] , _A=True , _A=True , _A="relu" , _A=3 , _A=None , ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ = parent SCREAMING_SNAKE_CASE_ = batch_size SCREAMING_SNAKE_CASE_ = image_size SCREAMING_SNAKE_CASE_ = num_channels SCREAMING_SNAKE_CASE_ = embeddings_size SCREAMING_SNAKE_CASE_ = hidden_sizes SCREAMING_SNAKE_CASE_ = depths SCREAMING_SNAKE_CASE_ = is_training SCREAMING_SNAKE_CASE_ = use_labels SCREAMING_SNAKE_CASE_ = hidden_act SCREAMING_SNAKE_CASE_ = num_labels SCREAMING_SNAKE_CASE_ = scope SCREAMING_SNAKE_CASE_ = len(_A ) def _UpperCamelCase ( self ) -> Any: SCREAMING_SNAKE_CASE_ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) SCREAMING_SNAKE_CASE_ = self.get_config() return config, pixel_values def _UpperCamelCase ( self ) -> Tuple: return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def _UpperCamelCase ( self , _A , _A ) -> str: SCREAMING_SNAKE_CASE_ = FlaxRegNetModel(config=_A ) SCREAMING_SNAKE_CASE_ = model(_A ) # Output shape (b, c, h, w) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _UpperCamelCase ( self , _A , _A ) -> Dict: SCREAMING_SNAKE_CASE_ = self.num_labels SCREAMING_SNAKE_CASE_ = FlaxRegNetForImageClassification(config=_A ) SCREAMING_SNAKE_CASE_ = model(_A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCamelCase ( self ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = config_and_inputs SCREAMING_SNAKE_CASE_ = {'''pixel_values''': pixel_values} return config, inputs_dict @require_flax class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" UpperCAmelCase_ =(FlaxRegNetModel, FlaxRegNetForImageClassification) if is_flax_available() else () UpperCAmelCase_ =False UpperCAmelCase_ =False UpperCAmelCase_ =False def _UpperCamelCase ( self ) -> None: SCREAMING_SNAKE_CASE_ = FlaxRegNetModelTester(self ) SCREAMING_SNAKE_CASE_ = ConfigTester(self , config_class=_A , has_text_modality=_A ) def _UpperCamelCase ( self ) -> List[Any]: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _UpperCamelCase ( self ) -> List[Any]: return def _UpperCamelCase ( self ) -> List[str]: SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_A ) def _UpperCamelCase ( self ) -> Any: SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_A ) @unittest.skip(reason='''RegNet does not use inputs_embeds''' ) def _UpperCamelCase ( self ) -> Optional[Any]: pass @unittest.skip(reason='''RegNet does not support input and output embeddings''' ) def _UpperCamelCase ( self ) -> Optional[int]: pass def _UpperCamelCase ( self ) -> List[Any]: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = model_class(_A ) SCREAMING_SNAKE_CASE_ = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE_ = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE_ = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _A ) def _UpperCamelCase ( self ) -> Any: def check_hidden_states_output(_A , _A , _A ): SCREAMING_SNAKE_CASE_ = model_class(_A ) SCREAMING_SNAKE_CASE_ = model(**self._prepare_for_class(_A , _A ) ) SCREAMING_SNAKE_CASE_ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states SCREAMING_SNAKE_CASE_ = self.model_tester.num_stages self.assertEqual(len(_A ) , expected_num_stages + 1 ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_ = True check_hidden_states_output(_A , _A , _A ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] SCREAMING_SNAKE_CASE_ = True check_hidden_states_output(_A , _A , _A ) def _UpperCamelCase ( self ) -> Tuple: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): SCREAMING_SNAKE_CASE_ = self._prepare_for_class(_A , _A ) SCREAMING_SNAKE_CASE_ = model_class(_A ) @jax.jit def model_jitted(_A , **_A ): return model(pixel_values=_A , **_A ) with self.subTest('''JIT Enabled''' ): SCREAMING_SNAKE_CASE_ = model_jitted(**_A ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): SCREAMING_SNAKE_CASE_ = model_jitted(**_A ).to_tuple() self.assertEqual(len(_A ) , len(_A ) ) for jitted_output, output in zip(_A , _A ): self.assertEqual(jitted_output.shape , output.shape ) def A__ ( ): SCREAMING_SNAKE_CASE_ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_flax class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def _UpperCamelCase ( self ) -> Tuple: return AutoImageProcessor.from_pretrained('''facebook/regnet-y-040''' ) if is_vision_available() else None @slow def _UpperCamelCase ( self ) -> List[str]: SCREAMING_SNAKE_CASE_ = FlaxRegNetForImageClassification.from_pretrained('''facebook/regnet-y-040''' ) SCREAMING_SNAKE_CASE_ = self.default_image_processor SCREAMING_SNAKE_CASE_ = prepare_img() SCREAMING_SNAKE_CASE_ = image_processor(images=_A , return_tensors='''np''' ) SCREAMING_SNAKE_CASE_ = model(**_A ) # verify the logits SCREAMING_SNAKE_CASE_ = (1, 1000) self.assertEqual(outputs.logits.shape , _A ) SCREAMING_SNAKE_CASE_ = jnp.array([-0.4180, -1.5051, -3.4836] ) self.assertTrue(jnp.allclose(outputs.logits[0, :3] , _A , atol=1E-4 ) )
257
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ConditionalDetrImageProcessor class UpperCamelCase__ ( unittest.TestCase ): """simple docstring""" def __init__( self , _A , _A=7 , _A=3 , _A=30 , _A=400 , _A=True , _A=None , _A=True , _A=[0.5, 0.5, 0.5] , _A=[0.5, 0.5, 0.5] , _A=True , _A=1 / 255 , _A=True , ) -> Optional[Any]: # by setting size["longest_edge"] > max_resolution we're effectively not testing this :p SCREAMING_SNAKE_CASE_ = size if size is not None else {'''shortest_edge''': 18, '''longest_edge''': 1333} SCREAMING_SNAKE_CASE_ = parent SCREAMING_SNAKE_CASE_ = batch_size SCREAMING_SNAKE_CASE_ = num_channels SCREAMING_SNAKE_CASE_ = min_resolution SCREAMING_SNAKE_CASE_ = max_resolution SCREAMING_SNAKE_CASE_ = do_resize SCREAMING_SNAKE_CASE_ = size SCREAMING_SNAKE_CASE_ = do_normalize SCREAMING_SNAKE_CASE_ = image_mean SCREAMING_SNAKE_CASE_ = image_std SCREAMING_SNAKE_CASE_ = do_rescale SCREAMING_SNAKE_CASE_ = rescale_factor SCREAMING_SNAKE_CASE_ = do_pad def _UpperCamelCase ( self ) -> Any: return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def _UpperCamelCase ( self , _A , _A=False ) -> str: if not batched: SCREAMING_SNAKE_CASE_ = image_inputs[0] if isinstance(_A , Image.Image ): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = image.size else: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = image.shape[1], image.shape[2] if w < h: SCREAMING_SNAKE_CASE_ = int(self.size['''shortest_edge'''] * h / w ) SCREAMING_SNAKE_CASE_ = self.size['''shortest_edge'''] elif w > h: SCREAMING_SNAKE_CASE_ = self.size['''shortest_edge'''] SCREAMING_SNAKE_CASE_ = int(self.size['''shortest_edge'''] * w / h ) else: SCREAMING_SNAKE_CASE_ = self.size['''shortest_edge'''] SCREAMING_SNAKE_CASE_ = self.size['''shortest_edge'''] else: SCREAMING_SNAKE_CASE_ = [] for image in image_inputs: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) SCREAMING_SNAKE_CASE_ = max(_A , key=lambda _A : item[0] )[0] SCREAMING_SNAKE_CASE_ = max(_A , key=lambda _A : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" UpperCAmelCase_ =ConditionalDetrImageProcessor if is_vision_available() else None def _UpperCamelCase ( self ) -> Any: SCREAMING_SNAKE_CASE_ = ConditionalDetrImageProcessingTester(self ) @property def _UpperCamelCase ( self ) -> str: return self.image_processor_tester.prepare_image_processor_dict() def _UpperCamelCase ( self ) -> Optional[Any]: SCREAMING_SNAKE_CASE_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_A , '''image_mean''' ) ) self.assertTrue(hasattr(_A , '''image_std''' ) ) self.assertTrue(hasattr(_A , '''do_normalize''' ) ) self.assertTrue(hasattr(_A , '''do_resize''' ) ) self.assertTrue(hasattr(_A , '''size''' ) ) def _UpperCamelCase ( self ) -> str: SCREAMING_SNAKE_CASE_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''shortest_edge''': 18, '''longest_edge''': 1333} ) self.assertEqual(image_processor.do_pad , _A ) SCREAMING_SNAKE_CASE_ = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=_A ) self.assertEqual(image_processor.size , {'''shortest_edge''': 42, '''longest_edge''': 84} ) self.assertEqual(image_processor.do_pad , _A ) def _UpperCamelCase ( self ) -> Any: pass def _UpperCamelCase ( self ) -> int: # Initialize image_processing SCREAMING_SNAKE_CASE_ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A ) for image in image_inputs: self.assertIsInstance(_A , Image.Image ) # Test not batched input SCREAMING_SNAKE_CASE_ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.image_processor_tester.get_expected_values(_A , batched=_A ) SCREAMING_SNAKE_CASE_ = image_processing(_A , return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _UpperCamelCase ( self ) -> List[Any]: # Initialize image_processing SCREAMING_SNAKE_CASE_ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors SCREAMING_SNAKE_CASE_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , numpify=_A ) for image in image_inputs: self.assertIsInstance(_A , np.ndarray ) # Test not batched input SCREAMING_SNAKE_CASE_ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched SCREAMING_SNAKE_CASE_ = image_processing(_A , return_tensors='''pt''' ).pixel_values SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def _UpperCamelCase ( self ) -> Union[str, Any]: # Initialize image_processing SCREAMING_SNAKE_CASE_ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE_ = prepare_image_inputs(self.image_processor_tester , equal_resolution=_A , torchify=_A ) for image in image_inputs: self.assertIsInstance(_A , torch.Tensor ) # Test not batched input SCREAMING_SNAKE_CASE_ = image_processing(image_inputs[0] , return_tensors='''pt''' ).pixel_values SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.image_processor_tester.get_expected_values(_A ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched SCREAMING_SNAKE_CASE_ = image_processing(_A , return_tensors='''pt''' ).pixel_values SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = self.image_processor_tester.get_expected_values(_A , batched=_A ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def _UpperCamelCase ( self ) -> str: # prepare image and target SCREAMING_SNAKE_CASE_ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' , '''r''' ) as f: SCREAMING_SNAKE_CASE_ = json.loads(f.read() ) SCREAMING_SNAKE_CASE_ = {'''image_id''': 39769, '''annotations''': target} # encode them SCREAMING_SNAKE_CASE_ = ConditionalDetrImageProcessor.from_pretrained('''microsoft/conditional-detr-resnet-50''' ) SCREAMING_SNAKE_CASE_ = image_processing(images=_A , annotations=_A , return_tensors='''pt''' ) # verify pixel values SCREAMING_SNAKE_CASE_ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['''pixel_values'''].shape , _A ) SCREAMING_SNAKE_CASE_ = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , _A , atol=1E-4 ) ) # verify area SCREAMING_SNAKE_CASE_ = torch.tensor([5887.9600, 1_1250.2061, 48_9353.8438, 83_7122.7500, 14_7967.5156, 16_5732.3438] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , _A ) ) # verify boxes SCREAMING_SNAKE_CASE_ = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , _A ) SCREAMING_SNAKE_CASE_ = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , _A , atol=1E-3 ) ) # verify image_id SCREAMING_SNAKE_CASE_ = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , _A ) ) # verify is_crowd SCREAMING_SNAKE_CASE_ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , _A ) ) # verify class_labels SCREAMING_SNAKE_CASE_ = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , _A ) ) # verify orig_size SCREAMING_SNAKE_CASE_ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , _A ) ) # verify size SCREAMING_SNAKE_CASE_ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , _A ) ) @slow def _UpperCamelCase ( self ) -> Tuple: # prepare image, target and masks_path SCREAMING_SNAKE_CASE_ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' , '''r''' ) as f: SCREAMING_SNAKE_CASE_ = json.loads(f.read() ) SCREAMING_SNAKE_CASE_ = {'''file_name''': '''000000039769.png''', '''image_id''': 39769, '''segments_info''': target} SCREAMING_SNAKE_CASE_ = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them SCREAMING_SNAKE_CASE_ = ConditionalDetrImageProcessor(format='''coco_panoptic''' ) SCREAMING_SNAKE_CASE_ = image_processing(images=_A , annotations=_A , masks_path=_A , return_tensors='''pt''' ) # verify pixel values SCREAMING_SNAKE_CASE_ = torch.Size([1, 3, 800, 1066] ) self.assertEqual(encoding['''pixel_values'''].shape , _A ) SCREAMING_SNAKE_CASE_ = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] , _A , atol=1E-4 ) ) # verify area SCREAMING_SNAKE_CASE_ = torch.tensor([14_7979.6875, 16_5527.0469, 48_4638.5938, 1_1292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] , _A ) ) # verify boxes SCREAMING_SNAKE_CASE_ = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape , _A ) SCREAMING_SNAKE_CASE_ = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] , _A , atol=1E-3 ) ) # verify image_id SCREAMING_SNAKE_CASE_ = torch.tensor([39769] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] , _A ) ) # verify is_crowd SCREAMING_SNAKE_CASE_ = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] , _A ) ) # verify class_labels SCREAMING_SNAKE_CASE_ = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] , _A ) ) # verify masks SCREAMING_SNAKE_CASE_ = 822873 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() , _A ) # verify orig_size SCREAMING_SNAKE_CASE_ = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] , _A ) ) # verify size SCREAMING_SNAKE_CASE_ = torch.tensor([800, 1066] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] , _A ) )
257
1
"""simple docstring""" # Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING, Dict, Optional import numpy as np import pyarrow as pa from .. import config from ..utils.logging import get_logger from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import jax import jaxlib _lowercase = get_logger() _lowercase = None class lowerCAmelCase_ ( TensorFormatter[Mapping, '''jax.Array''', Mapping] ): '''simple docstring''' def __init__( self : Optional[int] ,A_ : Any=None ,A_ : int=None ,**A_ : str ) -> Union[str, Any]: super().__init__(features=A_ ) import jax from jaxlib.xla_client import Device if isinstance(A_ ,A_ ): raise ValueError( F'Expected {device} to be a `str` not {type(A_ )}, as `jaxlib.xla_extension.Device` ' 'is not serializable neither with `pickle` nor with `dill`. Instead you can surround ' 'the device with `str()` to get its string identifier that will be internally mapped ' 'to the actual `jaxlib.xla_extension.Device`.' ) A = device if isinstance(A_ ,A_ ) else str(jax.devices()[0] ) # using global variable since `jaxlib.xla_extension.Device` is not serializable neither # with `pickle` nor with `dill`, so we need to use a global variable instead global DEVICE_MAPPING if DEVICE_MAPPING is None: A = self._map_devices_to_str() if self.device not in list(DEVICE_MAPPING.keys() ): logger.warning( F'Device with string identifier {self.device} not listed among the available ' F'devices: {list(DEVICE_MAPPING.keys() )}, so falling back to the default ' F'device: {str(jax.devices()[0] )}.' ) A = str(jax.devices()[0] ) A = jnp_array_kwargs @staticmethod def _SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: import jax return {str(A_ ): device for device in jax.devices()} def _SCREAMING_SNAKE_CASE ( self : Tuple ,A_ : Dict ) -> Union[str, Any]: import jax import jax.numpy as jnp if isinstance(A_ ,A_ ) and column: if all( isinstance(A_ ,jax.Array ) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column ): return jnp.stack(A_ ,axis=0 ) return column def _SCREAMING_SNAKE_CASE ( self : Tuple ,A_ : List[str] ) -> Union[str, Any]: import jax import jax.numpy as jnp if isinstance(A_ ,(str, bytes, type(A_ )) ): return value elif isinstance(A_ ,(np.character, np.ndarray) ) and np.issubdtype(value.dtype ,np.character ): return value.tolist() A = {} if isinstance(A_ ,(np.number, np.ndarray) ) and np.issubdtype(value.dtype ,np.integer ): # the default int precision depends on the jax config # see https://jax.readthedocs.io/en/latest/notebooks/Common_Gotchas_in_JAX.html#double-64bit-precision if jax.config.jax_enable_xaa: A = {'dtype': jnp.intaa} else: A = {'dtype': jnp.intaa} elif isinstance(A_ ,(np.number, np.ndarray) ) and np.issubdtype(value.dtype ,np.floating ): A = {'dtype': jnp.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(A_ ,PIL.Image.Image ): A = np.asarray(A_ ) # using global variable since `jaxlib.xla_extension.Device` is not serializable neither # with `pickle` nor with `dill`, so we need to use a global variable instead global DEVICE_MAPPING if DEVICE_MAPPING is None: A = self._map_devices_to_str() with jax.default_device(DEVICE_MAPPING[self.device] ): # calling jnp.array on a np.ndarray does copy the data # see https://github.com/google/jax/issues/4486 return jnp.array(A_ ,**{**default_dtype, **self.jnp_array_kwargs} ) def _SCREAMING_SNAKE_CASE ( self : int ,A_ : Optional[int] ) -> Union[str, Any]: import jax # support for torch, tf, jax etc. if config.TORCH_AVAILABLE and "torch" in sys.modules: import torch if isinstance(A_ ,torch.Tensor ): return self._tensorize(data_struct.detach().cpu().numpy()[()] ) if hasattr(A_ ,'__array__' ) and not isinstance(A_ ,jax.Array ): A = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(A_ ,np.ndarray ): if data_struct.dtype == object: # jax arrays cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(A_ ) for substruct in data_struct] ) elif isinstance(A_ ,(list, tuple) ): return self._consolidate([self.recursive_tensorize(A_ ) for substruct in data_struct] ) return self._tensorize(A_ ) def _SCREAMING_SNAKE_CASE ( self : str ,A_ : str ) -> int: return map_nested(self._recursive_tensorize ,A_ ,map_list=A_ ) def _SCREAMING_SNAKE_CASE ( self : Optional[int] ,A_ : Tuple ) -> List[str]: A = self.numpy_arrow_extractor().extract_row(A_ ) A = self.python_features_decoder.decode_row(A_ ) return self.recursive_tensorize(A_ ) def _SCREAMING_SNAKE_CASE ( self : int ,A_ : Optional[int] ) -> List[str]: A = self.numpy_arrow_extractor().extract_column(A_ ) A = self.python_features_decoder.decode_column(A_ ,pa_table.column_names[0] ) A = self.recursive_tensorize(A_ ) A = self._consolidate(A_ ) return column def _SCREAMING_SNAKE_CASE ( self : str ,A_ : Dict ) -> List[Any]: A = self.numpy_arrow_extractor().extract_batch(A_ ) A = self.python_features_decoder.decode_batch(A_ ) A = self.recursive_tensorize(A_ ) for column_name in batch: A = self._consolidate(batch[column_name] ) return batch
74
from __future__ import annotations from collections.abc import Callable from typing import Generic, TypeVar snake_case_ = TypeVar('''T''') snake_case_ = TypeVar('''U''') class SCREAMING_SNAKE_CASE__ (Generic[T, U] ): def __init__( self , a , a): lowercase__ : List[Any] = key lowercase__ : List[Any] = val lowercase__ : DoubleLinkedListNode[T, U] | None = None lowercase__ : DoubleLinkedListNode[T, U] | None = None def __repr__( self): return ( f"""Node: key: {self.key}, val: {self.val}, """ f"""has next: {bool(self.next)}, has prev: {bool(self.prev)}""" ) class SCREAMING_SNAKE_CASE__ (Generic[T, U] ): def __init__( self): lowercase__ : DoubleLinkedListNode[T, U] = DoubleLinkedListNode(a , a) lowercase__ : DoubleLinkedListNode[T, U] = DoubleLinkedListNode(a , a) lowercase__ , lowercase__ : Union[str, Any] = self.rear, self.head def __repr__( self): lowercase__ : Any = ['DoubleLinkedList'] lowercase__ : List[str] = self.head while node.next is not None: rep.append(str(a)) lowercase__ : Tuple = node.next rep.append(str(self.rear)) return ",\n ".join(a) def snake_case_ ( self , a): lowercase__ : Optional[Any] = self.rear.prev # All nodes other than self.head are guaranteed to have non-None previous assert previous is not None lowercase__ : Dict = node lowercase__ : int = previous lowercase__ : Union[str, Any] = node lowercase__ : Optional[int] = self.rear def snake_case_ ( self , a): if node.prev is None or node.next is None: return None lowercase__ : Union[str, Any] = node.next lowercase__ : Tuple = node.prev lowercase__ : Union[str, Any] = None lowercase__ : List[Any] = None return node class SCREAMING_SNAKE_CASE__ (Generic[T, U] ): __lowerCamelCase : dict[Callable[[T], U], LRUCache[T, U]] = {} def __init__( self , a): lowercase__ : DoubleLinkedList[T, U] = DoubleLinkedList() lowercase__ : Optional[Any] = capacity lowercase__ : Union[str, Any] = 0 lowercase__ : Tuple = 0 lowercase__ : int = 0 lowercase__ : dict[T, DoubleLinkedListNode[T, U]] = {} def __repr__( self): return ( f"""CacheInfo(hits={self.hits}, misses={self.miss}, """ f"""capacity={self.capacity}, current size={self.num_keys})""" ) def __contains__( self , a): return key in self.cache def snake_case_ ( self , a): # Note: pythonic interface would throw KeyError rather than return None if key in self.cache: self.hits += 1 lowercase__ : DoubleLinkedListNode[T, U] = self.cache[key] lowercase__ : str = self.list.remove(self.cache[key]) assert node == value_node # node is guaranteed not None because it is in self.cache assert node is not None self.list.add(a) return node.val self.miss += 1 return None def snake_case_ ( self , a , a): if key not in self.cache: if self.num_keys >= self.capacity: # delete first node (oldest) when over capacity lowercase__ : Optional[int] = self.list.head.next # guaranteed to have a non-None first node when num_keys > 0 # explain to type checker via assertions assert first_node is not None assert first_node.key is not None assert ( self.list.remove(a) is not None ) # node guaranteed to be in list assert node.key is not None del self.cache[first_node.key] self.num_keys -= 1 lowercase__ : Optional[Any] = DoubleLinkedListNode(a , a) self.list.add(self.cache[key]) self.num_keys += 1 else: # bump node to the end of the list, update value lowercase__ : Any = self.list.remove(self.cache[key]) assert node is not None # node guaranteed to be in list lowercase__ : Union[str, Any] = value self.list.add(a) @classmethod def snake_case_ ( cls , a = 128): def cache_decorator_inner(a) -> Callable[..., U]: def cache_decorator_wrapper(*a) -> U: if func not in cls.decorator_function_to_instance_map: lowercase__ : Dict = LRUCache(a) lowercase__ : str = cls.decorator_function_to_instance_map[func].get(args[0]) if result is None: lowercase__ : str = func(*a) cls.decorator_function_to_instance_map[func].put(args[0] , a) return result def cache_info() -> LRUCache[T, U]: return cls.decorator_function_to_instance_map[func] setattr(a , 'cache_info' , a) # noqa: B010 return cache_decorator_wrapper return cache_decorator_inner if __name__ == "__main__": import doctest doctest.testmod()
214
0
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class __lowercase (__SCREAMING_SNAKE_CASE , unittest.TestCase ): """simple docstring""" _UpperCAmelCase = ShapEPipeline _UpperCAmelCase = ["""prompt"""] _UpperCAmelCase = ["""prompt"""] _UpperCAmelCase = [ """num_images_per_prompt""", """num_inference_steps""", """generator""", """latents""", """guidance_scale""", """frame_size""", """output_type""", """return_dict""", ] _UpperCAmelCase = False @property def UpperCamelCase__ ( self ): """simple docstring""" return 3_2 @property def UpperCamelCase__ ( self ): """simple docstring""" return 3_2 @property def UpperCamelCase__ ( self ): """simple docstring""" return self.time_input_dim * 4 @property def UpperCamelCase__ ( self ): """simple docstring""" return 8 @property def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) return tokenizer @property def UpperCamelCase__ ( self ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ : int = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) return CLIPTextModelWithProjection(lowerCAmelCase__ ) @property def UpperCamelCase__ ( self ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ : Dict = { 'num_attention_heads': 2, 'attention_head_dim': 1_6, 'embedding_dim': self.time_input_dim, 'num_embeddings': 3_2, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } SCREAMING_SNAKE_CASE_ : List[str] = PriorTransformer(**lowerCAmelCase__ ) return model @property def UpperCamelCase__ ( self ): """simple docstring""" torch.manual_seed(0 ) SCREAMING_SNAKE_CASE_ : List[str] = { 'param_shapes': ( (self.renderer_dim, 9_3), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 1_2, 'background': ( 0.1, 0.1, 0.1, ), } SCREAMING_SNAKE_CASE_ : List[Any] = ShapERenderer(**lowerCAmelCase__ ) return model def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.dummy_prior SCREAMING_SNAKE_CASE_ : Optional[Any] = self.dummy_text_encoder SCREAMING_SNAKE_CASE_ : Optional[Any] = self.dummy_tokenizer SCREAMING_SNAKE_CASE_ : str = self.dummy_renderer SCREAMING_SNAKE_CASE_ : Union[str, Any] = HeunDiscreteScheduler( beta_schedule='exp' , num_train_timesteps=1_0_2_4 , prediction_type='sample' , use_karras_sigmas=lowerCAmelCase__ , clip_sample=lowerCAmelCase__ , clip_sample_range=1.0 , ) SCREAMING_SNAKE_CASE_ : Dict = { 'prior': prior, 'text_encoder': text_encoder, 'tokenizer': tokenizer, 'renderer': renderer, 'scheduler': scheduler, } return components def UpperCamelCase__ ( self , lowerCAmelCase__ , lowerCAmelCase__=0 ): """simple docstring""" if str(lowerCAmelCase__ ).startswith('mps' ): SCREAMING_SNAKE_CASE_ : Tuple = torch.manual_seed(lowerCAmelCase__ ) else: SCREAMING_SNAKE_CASE_ : List[Any] = torch.Generator(device=lowerCAmelCase__ ).manual_seed(lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : str = { 'prompt': 'horse', 'generator': generator, 'num_inference_steps': 1, 'frame_size': 3_2, 'output_type': 'np', } return inputs def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = 'cpu' SCREAMING_SNAKE_CASE_ : int = self.get_dummy_components() SCREAMING_SNAKE_CASE_ : Dict = self.pipeline_class(**lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : List[str] = pipe(**self.get_dummy_inputs(lowerCAmelCase__ ) ) SCREAMING_SNAKE_CASE_ : List[Any] = output.images[0] SCREAMING_SNAKE_CASE_ : Tuple = image[0, -3:, -3:, -1] assert image.shape == (2_0, 3_2, 3_2, 3) SCREAMING_SNAKE_CASE_ : Optional[Any] = np.array( [ 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, 0.00_039_216, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase__ ( self ): """simple docstring""" self._test_inference_batch_consistent(batch_sizes=[1, 2] ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = torch_device == 'cpu' SCREAMING_SNAKE_CASE_ : List[str] = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=lowerCAmelCase__ , relax_max_difference=lowerCAmelCase__ , ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = self.get_dummy_components() SCREAMING_SNAKE_CASE_ : Tuple = self.pipeline_class(**lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : Dict = pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : int = 1 SCREAMING_SNAKE_CASE_ : str = 2 SCREAMING_SNAKE_CASE_ : Tuple = self.get_dummy_inputs(lowerCAmelCase__ ) for key in inputs.keys(): if key in self.batch_params: SCREAMING_SNAKE_CASE_ : Optional[Any] = batch_size * [inputs[key]] SCREAMING_SNAKE_CASE_ : int = pipe(**lowerCAmelCase__ , num_images_per_prompt=lowerCAmelCase__ )[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class __lowercase (unittest.TestCase ): """simple docstring""" def UpperCamelCase__ ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : str = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_np_out.npy' ) SCREAMING_SNAKE_CASE_ : Optional[Any] = ShapEPipeline.from_pretrained('openai/shap-e' ) SCREAMING_SNAKE_CASE_ : Dict = pipe.to(lowerCAmelCase__ ) pipe.set_progress_bar_config(disable=lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : int = torch.Generator(device=lowerCAmelCase__ ).manual_seed(0 ) SCREAMING_SNAKE_CASE_ : int = pipe( 'a shark' , generator=lowerCAmelCase__ , guidance_scale=15.0 , num_inference_steps=6_4 , frame_size=6_4 , output_type='np' , ).images[0] assert images.shape == (2_0, 6_4, 6_4, 3) assert_mean_pixel_difference(lowerCAmelCase__ , lowerCAmelCase__ )
162
import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def a__ ( ): raise RuntimeError('CUDA out of memory.' ) class __lowercase (nn.Module ): """simple docstring""" def __init__( self ): """simple docstring""" super().__init__() SCREAMING_SNAKE_CASE_ : int = nn.Linear(3 , 4 ) SCREAMING_SNAKE_CASE_ : Tuple = nn.BatchNormad(4 ) SCREAMING_SNAKE_CASE_ : str = nn.Linear(4 , 5 ) def UpperCamelCase__ ( self , lowerCAmelCase__ ): """simple docstring""" return self.lineara(self.batchnorm(self.lineara(lowerCAmelCase__ ) ) ) class __lowercase (unittest.TestCase ): """simple docstring""" def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = [] @find_executable_batch_size(starting_batch_size=1_2_8 ) def mock_training_loop_function(lowerCAmelCase__ ): nonlocal batch_sizes batch_sizes.append(lowerCAmelCase__ ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(lowerCAmelCase__ , [1_2_8, 6_4, 3_2, 1_6, 8] ) def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple = [] @find_executable_batch_size(starting_batch_size=1_2_8 ) def mock_training_loop_function(lowerCAmelCase__ , lowerCAmelCase__ ): nonlocal batch_sizes batch_sizes.append(lowerCAmelCase__ ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ : Dict = mock_training_loop_function('hello' ) self.assertListEqual(lowerCAmelCase__ , [1_2_8, 6_4, 3_2, 1_6, 8] ) self.assertListEqual([bs, arga] , [8, 'hello'] ) def UpperCamelCase__ ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(lowerCAmelCase__ ): pass with self.assertRaises(lowerCAmelCase__ ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.' , cm.exception.args[0] ) def UpperCamelCase__ ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=1_6 ) def mock_training_loop_function(lowerCAmelCase__ ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(lowerCAmelCase__ ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.' , cm.exception.args[0] ) def UpperCamelCase__ ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=1_2_8 ) def mock_training_loop_function(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(lowerCAmelCase__ ) as cm: mock_training_loop_function(1_2_8 , 'hello' , 'world' ) self.assertIn('Batch size was passed into `f`' , cm.exception.args[0] ) self.assertIn('`f(arg1=\'hello\', arg2=\'world\')' , cm.exception.args[0] ) def UpperCamelCase__ ( self ): """simple docstring""" @find_executable_batch_size(starting_batch_size=1_6 ) def mock_training_loop_function(lowerCAmelCase__ ): raise ValueError('Oops, we had an error!' ) with self.assertRaises(lowerCAmelCase__ ) as cm: mock_training_loop_function() self.assertIn('Oops, we had an error!' , cm.exception.args[0] ) @require_cuda def UpperCamelCase__ ( self ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = torch.cuda.memory_allocated() SCREAMING_SNAKE_CASE_ : Optional[int] = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , lowerCAmelCase__ ) SCREAMING_SNAKE_CASE_ : List[Any] = release_memory(lowerCAmelCase__ ) self.assertEqual(torch.cuda.memory_allocated() , lowerCAmelCase__ )
162
1
def SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase ) -> Dict: '''simple docstring''' assert ( isinstance(_UpperCAmelCase, _UpperCAmelCase ) and number_of_steps > 0 ), f"number_of_steps needs to be positive integer, your input {number_of_steps}" if number_of_steps == 1: return 1 lowerCAmelCase : List[Any] = 1, 1 for _ in range(number_of_steps - 1 ): lowerCAmelCase : List[str] = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
138
'''simple docstring''' def A_ ( snake_case ): if not all(x.isalpha() for x in string ): raise ValueError("String must only contain alphabetic characters." ) SCREAMING_SNAKE_CASE:Optional[int] = sorted(string.lower() ) return len(snake_case ) == len(set(snake_case ) ) if __name__ == "__main__": A_ = input("Enter a string ").strip() A_ = is_isogram(input_str) print(f'''{input_str} is {"an" if isogram else "not an"} isogram.''')
139
0
import numpy as np import torch from ..models.clipseg import CLIPSegForImageSegmentation from ..utils import is_vision_available, requires_backends from .base import PipelineTool if is_vision_available(): from PIL import Image class UpperCamelCase__ ( a__ ): _SCREAMING_SNAKE_CASE : Optional[Any] = ( "This is a tool that creates a segmentation mask of an image according to a label. It cannot create an image." "It takes two arguments named `image` which should be the original image, and `label` which should be a text " "describing the elements what should be identified in the segmentation mask. The tool returns the mask." ) _SCREAMING_SNAKE_CASE : List[str] = "CIDAS/clipseg-rd64-refined" _SCREAMING_SNAKE_CASE : Any = "image_segmenter" _SCREAMING_SNAKE_CASE : str = CLIPSegForImageSegmentation _SCREAMING_SNAKE_CASE : Optional[Any] = ["image", "text"] _SCREAMING_SNAKE_CASE : Any = ["image"] def __init__(self : Optional[int] , *snake_case_ : Optional[int] , **snake_case_ : Union[str, Any] ): requires_backends(self , ['''vision'''] ) super().__init__(*_lowerCamelCase , **_lowerCamelCase ) def lowerCAmelCase (self : List[str] , snake_case_ : "Image" , snake_case_ : str ): return self.pre_processor(text=[label] , images=[image] , padding=_lowerCamelCase , return_tensors='''pt''' ) def lowerCAmelCase (self : Union[str, Any] , snake_case_ : Optional[int] ): with torch.no_grad(): __a : Optional[int] = self.model(**_lowerCamelCase ).logits return logits def lowerCAmelCase (self : List[str] , snake_case_ : Optional[int] ): __a : int = outputs.cpu().detach().numpy() __a : Tuple = 0 __a : List[str] = 1 return Image.fromarray((array * 2_5_5).astype(np.uinta ) )
354
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowercase__ ={ 'configuration_efficientnet': [ 'EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP', 'EfficientNetConfig', 'EfficientNetOnnxConfig', ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ =['EfficientNetImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ =[ 'EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST', 'EfficientNetForImageClassification', 'EfficientNetModel', 'EfficientNetPreTrainedModel', ] if TYPE_CHECKING: from .configuration_efficientnet import ( EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientNetConfig, EfficientNetOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientnet import EfficientNetImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientnet import ( EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientNetForImageClassification, EfficientNetModel, EfficientNetPreTrainedModel, ) else: import sys lowercase__ =_LazyModule(__name__, globals()['__file__'], _import_structure)
90
0
import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def lowerCAmelCase__ ( lowerCamelCase_ : int=None ,lowerCamelCase_ : Union[str, Any]=None): '''simple docstring''' return field(default_factory=lambda: default ,metadata=lowerCamelCase_) @dataclass class lowerCamelCase__ : '''simple docstring''' snake_case_ =field( metadata={"""help""": """The csv file to plot."""} , ) snake_case_ =field( default=lowerCamelCase__ , metadata={"""help""": """Whether to plot along batch size or sequence length. Defaults to sequence length."""} , ) snake_case_ =field( default=lowerCamelCase__ , metadata={"""help""": """Whether the csv file has time results or memory results. Defaults to memory results."""} , ) snake_case_ =field( default=lowerCamelCase__ , metadata={"""help""": """Disable logarithmic scale when plotting"""} , ) snake_case_ =field( default=lowerCamelCase__ , metadata={ """help""": """Whether the csv file has training results or inference results. Defaults to inference results.""" } , ) snake_case_ =field( default=lowerCamelCase__ , metadata={"""help""": """Filename under which the plot will be saved. If unused no plot is saved."""} , ) snake_case_ =list_field( default=lowerCamelCase__ , metadata={"""help""": """List of model names that are used instead of the ones in the csv file."""}) def lowerCAmelCase__ ( lowerCamelCase_ : Optional[Any]): '''simple docstring''' try: int(lowerCamelCase_) return True except ValueError: return False def lowerCAmelCase__ ( lowerCamelCase_ : Optional[int]): '''simple docstring''' try: float(lowerCamelCase_) return True except ValueError: return False class lowerCamelCase__ : '''simple docstring''' def __init__(self ,__lowerCamelCase ) -> Optional[Any]: """simple docstring""" lowerCAmelCase__ : Tuple = args lowerCAmelCase__ : List[str] = defaultdict(lambda: {"bsz": [], "seq_len": [], "result": {}} ) with open(self.args.csv_file ,newline='''''' ) as csv_file: lowerCAmelCase__ : Union[str, Any] = csv.DictReader(__lowerCamelCase ) for row in reader: lowerCAmelCase__ : str = row['''model'''] self.result_dict[model_name]["bsz"].append(int(row['''batch_size'''] ) ) self.result_dict[model_name]["seq_len"].append(int(row['''sequence_length'''] ) ) if can_convert_to_int(row['''result'''] ): # value is not None lowerCAmelCase__ : Tuple = int(row['''result'''] ) elif can_convert_to_float(row['''result'''] ): # value is not None lowerCAmelCase__ : Union[str, Any] = float(row['''result'''] ) def lowerCAmelCase__ (self ) -> Any: """simple docstring""" lowerCAmelCase__ , lowerCAmelCase__ : str = plt.subplots() lowerCAmelCase__ : List[Any] = '''Time usage''' if self.args.is_time else '''Memory usage''' lowerCAmelCase__ : Union[str, Any] = title_str + ''' for training''' if self.args.is_train else title_str + ''' for inference''' if not self.args.no_log_scale: # set logarithm scales ax.set_xscale('''log''' ) ax.set_yscale('''log''' ) for axis in [ax.xaxis, ax.yaxis]: axis.set_major_formatter(ScalarFormatter() ) for model_name_idx, model_name in enumerate(self.result_dict.keys() ): lowerCAmelCase__ : Tuple = sorted(set(self.result_dict[model_name]['''bsz'''] ) ) lowerCAmelCase__ : List[str] = sorted(set(self.result_dict[model_name]['''seq_len'''] ) ) lowerCAmelCase__ : int = self.result_dict[model_name]['''result'''] ((lowerCAmelCase__) , (lowerCAmelCase__)) : List[Any] = ( (batch_sizes, sequence_lengths) if self.args.plot_along_batch else (sequence_lengths, batch_sizes) ) lowerCAmelCase__ : List[str] = ( model_name if self.args.short_model_names is None else self.args.short_model_names[model_name_idx] ) for inner_loop_value in inner_loop_array: if self.args.plot_along_batch: lowerCAmelCase__ : List[Any] = np.asarray( [results[(x, inner_loop_value)] for x in x_axis_array if (x, inner_loop_value) in results] ,dtype=__lowerCamelCase ,) else: lowerCAmelCase__ : List[Any] = np.asarray( [results[(inner_loop_value, x)] for x in x_axis_array if (inner_loop_value, x) in results] ,dtype=np.floataa ,) ((lowerCAmelCase__) , (lowerCAmelCase__)) : Union[str, Any] = ( ('''batch_size''', '''len''') if self.args.plot_along_batch else ('''in #tokens''', '''bsz''') ) lowerCAmelCase__ : Tuple = np.asarray(__lowerCamelCase ,__lowerCamelCase )[: len(__lowerCamelCase )] plt.scatter( __lowerCamelCase ,__lowerCamelCase ,label=f"""{label_model_name} - {inner_loop_label}: {inner_loop_value}""" ) plt.plot(__lowerCamelCase ,__lowerCamelCase ,'''--''' ) title_str += f""" {label_model_name} vs.""" lowerCAmelCase__ : Optional[int] = title_str[:-4] lowerCAmelCase__ : int = '''Time in s''' if self.args.is_time else '''Memory in MB''' # plot plt.title(__lowerCamelCase ) plt.xlabel(__lowerCamelCase ) plt.ylabel(__lowerCamelCase ) plt.legend() if self.args.figure_png_file is not None: plt.savefig(self.args.figure_png_file ) else: plt.show() def lowerCAmelCase__ ( ): '''simple docstring''' lowerCAmelCase__ : Tuple = HfArgumentParser(lowerCamelCase_) lowerCAmelCase__ : Optional[Any] = parser.parse_args_into_dataclasses()[0] lowerCAmelCase__ : Tuple = Plot(args=lowerCamelCase_) plot.plot() if __name__ == "__main__": main()
129
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("""TEST_SAGEMAKER""" , """False""")) is not True , reason="""Skipping test because should only be run when releasing minor transformers version""" , ) @pytest.mark.usefixtures("""sm_env""") @parameterized_class( [ { """framework""": """pytorch""", """script""": """run_glue.py""", """model_name_or_path""": """distilbert-base-cased""", """instance_type""": """ml.p3.16xlarge""", """results""": {"""train_runtime""": 650, """eval_accuracy""": 0.7, """eval_loss""": 0.6}, }, { """framework""": """pytorch""", """script""": """run_ddp.py""", """model_name_or_path""": """distilbert-base-cased""", """instance_type""": """ml.p3.16xlarge""", """results""": {"""train_runtime""": 600, """eval_accuracy""": 0.7, """eval_loss""": 0.6}, }, { """framework""": """tensorflow""", """script""": """run_tf_dist.py""", """model_name_or_path""": """distilbert-base-cased""", """instance_type""": """ml.p3.16xlarge""", """results""": {"""train_runtime""": 600, """eval_accuracy""": 0.6, """eval_loss""": 0.7}, }, ]) class lowerCamelCase__ ( unittest.TestCase): '''simple docstring''' def lowerCAmelCase__ (self ) -> List[str]: """simple docstring""" if self.framework == "pytorch": subprocess.run( f"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() ,encoding='''utf-8''' ,check=__lowerCamelCase ,) assert hasattr(self ,'''env''' ) def lowerCAmelCase__ (self ,__lowerCamelCase ) -> str: """simple docstring""" lowerCAmelCase__ : Optional[int] = f"""{self.env.base_job_name}-{instance_count}-{'ddp' if 'ddp' in self.script else 'smd'}""" # distributed data settings lowerCAmelCase__ : Optional[Any] = {'''smdistributed''': {'''dataparallel''': {'''enabled''': True}}} if self.script != '''run_ddp.py''' else None # creates estimator return HuggingFace( entry_point=self.script ,source_dir=self.env.test_path ,role=self.env.role ,image_uri=self.env.image_uri ,base_job_name=__lowerCamelCase ,instance_count=__lowerCamelCase ,instance_type=self.instance_type ,debugger_hook_config=__lowerCamelCase ,hyperparameters={**self.env.distributed_hyperparameters, '''model_name_or_path''': self.model_name_or_path} ,metric_definitions=self.env.metric_definitions ,distribution=__lowerCamelCase ,py_version='''py36''' ,) def lowerCAmelCase__ (self ,__lowerCamelCase ) -> str: """simple docstring""" TrainingJobAnalytics(__lowerCamelCase ).export_csv(f"""{self.env.test_path}/{job_name}_metrics.csv""" ) @parameterized.expand([(2,)] ) def lowerCAmelCase__ (self ,__lowerCamelCase ) -> Optional[Any]: """simple docstring""" lowerCAmelCase__ : Optional[Any] = self.create_estimator(__lowerCamelCase ) # run training estimator.fit() # result dataframe lowerCAmelCase__ : Dict = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis lowerCAmelCase__ : List[Any] = list(result_metrics_df[result_metrics_df.metric_name == '''eval_accuracy''']['''value'''] ) lowerCAmelCase__ : Union[str, Any] = list(result_metrics_df[result_metrics_df.metric_name == '''eval_loss''']['''value'''] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping lowerCAmelCase__ : Optional[Any] = ( Session().describe_training_job(estimator.latest_training_job.name ).get('''TrainingTimeInSeconds''' ,99_99_99 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results['''eval_accuracy'''] for t in eval_accuracy ) assert all(t <= self.results['''eval_loss'''] for t in eval_loss ) # dump tests result into json file to share in PR with open(f"""{estimator.latest_training_job.name}.json""" ,'''w''' ) as outfile: json.dump({'''train_time''': train_runtime, '''eval_accuracy''': eval_accuracy, '''eval_loss''': eval_loss} ,__lowerCamelCase )
129
1
__magic_name__ = {0: [2, 3], 1: [0], 2: [1], 3: [4], 4: []} __magic_name__ = {0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]} def _lowerCAmelCase ( A__: dict[int, list[int]] , A__: int , A__: list[bool] ): '''simple docstring''' UpperCAmelCase = True UpperCAmelCase = [] for neighbour in graph[vert]: if not visited[neighbour]: order += topology_sort(A__ , A__ , A__ ) order.append(A__ ) return order def _lowerCAmelCase ( A__: dict[int, list[int]] , A__: int , A__: list[bool] ): '''simple docstring''' UpperCAmelCase = True UpperCAmelCase = [vert] for neighbour in reversed_graph[vert]: if not visited[neighbour]: component += find_components(A__ , A__ , A__ ) return component def _lowerCAmelCase ( A__: dict[int, list[int]] ): '''simple docstring''' UpperCAmelCase = len(A__ ) * [False] UpperCAmelCase = {vert: [] for vert in range(len(A__ ) )} for vert, neighbours in graph.items(): for neighbour in neighbours: reversed_graph[neighbour].append(A__ ) UpperCAmelCase = [] for i, was_visited in enumerate(A__ ): if not was_visited: order += topology_sort(A__ , A__ , A__ ) UpperCAmelCase = [] UpperCAmelCase = len(A__ ) * [False] for i in range(len(A__ ) ): UpperCAmelCase = order[len(A__ ) - i - 1] if not visited[vert]: UpperCAmelCase = find_components(A__ , A__ , A__ ) components_list.append(A__ ) return components_list
152
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging __magic_name__ = logging.get_logger(__name__) def _lowerCAmelCase ( A__: nn.ModuleList , A__: nn.ModuleList , A__: List[int] ): '''simple docstring''' UpperCAmelCase = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(A__ ) == len(A__ ), F"""{len(A__ )} != {len(A__ )}""" dest_layers.load_state_dict(layers_to_copy.state_dict() ) __magic_name__ = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } __magic_name__ = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def _lowerCAmelCase ( A__: List[str] , A__: Optional[int] ): '''simple docstring''' try: UpperCAmelCase = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F"""no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first""" F""" {n_student}""" ) return list(range(A__ ) ) def _lowerCAmelCase ( A__: Optional[int] , A__: Tuple ): '''simple docstring''' if n_student > n_teacher: raise ValueError(F"""Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}""" ) elif n_teacher == n_student: return list(range(A__ ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def _lowerCAmelCase ( A__: Union[str, PreTrainedModel] , A__: Union[str, Path] = "student" , A__: Union[int, None] = None , A__: Union[int, None] = None , A__: Optional[int]=False , A__: Tuple=None , A__: Any=None , **A__: List[str] , ): '''simple docstring''' UpperCAmelCase = '''encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher.''' assert (e is not None) or (d is not None), _msg if isinstance(A__ , A__ ): AutoTokenizer.from_pretrained(A__ ).save_pretrained(A__ ) # purely for convenience UpperCAmelCase = AutoModelForSeqaSeqLM.from_pretrained(A__ ).eval() else: assert isinstance(A__ , A__ ), F"""teacher must be a model or string got type {type(A__ )}""" UpperCAmelCase = teacher.config.to_diff_dict() try: UpperCAmelCase , UpperCAmelCase = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: UpperCAmelCase = teacher_e if d is None: UpperCAmelCase = teacher_d init_kwargs.update({'''encoder_layers''': e, '''decoder_layers''': d} ) except AttributeError: # T5 if hasattr(teacher.config , '''num_encoder_layers''' ): UpperCAmelCase , UpperCAmelCase = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: UpperCAmelCase , UpperCAmelCase = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: UpperCAmelCase = teacher_e if d is None: UpperCAmelCase = teacher_d if hasattr(teacher.config , '''num_encoder_layers''' ): init_kwargs.update({'''num_encoder_layers''': e, '''num_decoder_layers''': d} ) else: init_kwargs.update({'''num_layers''': e, '''num_decoder_layers''': d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(A__ ) # Copy weights UpperCAmelCase = teacher.config_class(**A__ ) UpperCAmelCase = AutoModelForSeqaSeqLM.from_config(A__ ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. UpperCAmelCase = student.load_state_dict(teacher.state_dict() , strict=A__ ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save UpperCAmelCase , UpperCAmelCase = list(range(A__ ) ), list(range(A__ ) ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to""" F""" {save_path}""" ) student.save_pretrained(A__ ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: UpperCAmelCase = pick_layers_to_copy(A__ , A__ ) if d_layers_to_copy is None: UpperCAmelCase = pick_layers_to_copy(A__ , A__ ) try: if hasattr( A__ , '''prophetnet''' ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , A__ ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , A__ ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , A__ ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , A__ ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , A__ ) copy_layers(teacher.decoder.block , student.decoder.block , A__ ) logger.info( F"""Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}""" ) UpperCAmelCase = { '''teacher_type''': teacher.config.model_type, '''copied_encoder_layers''': e_layers_to_copy, '''copied_decoder_layers''': d_layers_to_copy, } student.save_pretrained(A__ ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
152
1
import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor __snake_case :Tuple = logging.get_logger(__name__) class _A ( __UpperCAmelCase ): def __init__( self : Optional[int] , *__SCREAMING_SNAKE_CASE : Union[str, Any] , **__SCREAMING_SNAKE_CASE : List[Any]): '''simple docstring''' warnings.warn( '''The class MobileViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use MobileViTImageProcessor instead.''' , __SCREAMING_SNAKE_CASE , ) super().__init__(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE)
49
class a__ ( snake_case__ ): pass class a__ ( snake_case__ ): pass class a__ : def __init__( self ): """simple docstring""" __lowerCAmelCase = [ [], [], [], ] def __SCREAMING_SNAKE_CASE( self , _A , _A ): """simple docstring""" try: if len(self.queues[priority] ) >= 1_0_0: raise OverflowError("Maximum queue size is 100" ) self.queues[priority].append(_A ) except IndexError: raise ValueError("Valid priorities are 0, 1, and 2" ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" for queue in self.queues: if queue: return queue.pop(0 ) raise UnderFlowError("All queues are empty" ) def __str__( self ): """simple docstring""" return "\n".join(f"""Priority {i}: {q}""" for i, q in enumerate(self.queues ) ) class a__ : def __init__( self ): """simple docstring""" __lowerCAmelCase = [] def __SCREAMING_SNAKE_CASE( self , _A ): """simple docstring""" if len(self.queue ) == 1_0_0: raise OverFlowError("Maximum queue size is 100" ) self.queue.append(_A ) def __SCREAMING_SNAKE_CASE( self ): """simple docstring""" if not self.queue: raise UnderFlowError("The queue is empty" ) else: __lowerCAmelCase = min(self.queue ) self.queue.remove(_A ) return data def __str__( self ): """simple docstring""" return str(self.queue ) def _a ( ): __lowerCAmelCase = FixedPriorityQueue() fpq.enqueue(0 , 10 ) fpq.enqueue(1 , 70 ) fpq.enqueue(0 , 1_00 ) fpq.enqueue(2 , 1 ) fpq.enqueue(2 , 5 ) fpq.enqueue(1 , 7 ) fpq.enqueue(2 , 4 ) fpq.enqueue(1 , 64 ) fpq.enqueue(0 , 1_28 ) print(SCREAMING_SNAKE_CASE_ ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(SCREAMING_SNAKE_CASE_ ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) print(fpq.dequeue() ) def _a ( ): __lowerCAmelCase = ElementPriorityQueue() epq.enqueue(10 ) epq.enqueue(70 ) epq.enqueue(1_00 ) epq.enqueue(1 ) epq.enqueue(5 ) epq.enqueue(7 ) epq.enqueue(4 ) epq.enqueue(64 ) epq.enqueue(1_28 ) print(SCREAMING_SNAKE_CASE_ ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(SCREAMING_SNAKE_CASE_ ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) print(epq.dequeue() ) if __name__ == "__main__": fixed_priority_queue() element_priority_queue()
92
0
def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ = 1_0_0_0 ) -> int: __lowerCamelCase : List[Any] = 2**power __lowerCamelCase : Dict = 0 while n: __lowerCamelCase , __lowerCamelCase : Union[str, Any] = r + n % 1_0, n // 1_0 return r if __name__ == "__main__": print(solution(int(str(input()).strip())))
113
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed a ="""true""" def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__=8_2 , lowerCamelCase__=1_6 ) -> List[Any]: set_seed(4_2 ) __lowerCamelCase : Tuple = RegressionModel() __lowerCamelCase : str = deepcopy(lowerCamelCase__ ) __lowerCamelCase : Optional[int] = RegressionDataset(length=lowerCamelCase__ ) __lowerCamelCase : Union[str, Any] = DataLoader(lowerCamelCase__ , batch_size=lowerCamelCase__ ) model.to(accelerator.device ) __lowerCamelCase , __lowerCamelCase : Tuple = accelerator.prepare(lowerCamelCase__ , lowerCamelCase__ ) return model, ddp_model, dataloader def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__=False ) -> List[Any]: __lowerCamelCase : Tuple = AutoTokenizer.from_pretrained('hf-internal-testing/mrpc-bert-base-cased' ) __lowerCamelCase : Any = load_dataset('glue' , 'mrpc' , split='validation' ) def tokenize_function(lowerCamelCase__ ): __lowerCamelCase : Union[str, Any] = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=lowerCamelCase__ , max_length=lowerCamelCase__ ) return outputs with accelerator.main_process_first(): __lowerCamelCase : Union[str, Any] = dataset.map( lowerCamelCase__ , batched=lowerCamelCase__ , remove_columns=['idx', 'sentence1', 'sentence2'] , ) __lowerCamelCase : Tuple = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(lowerCamelCase__ ): if use_longest: return tokenizer.pad(lowerCamelCase__ , padding='longest' , return_tensors='pt' ) return tokenizer.pad(lowerCamelCase__ , padding='max_length' , max_length=1_2_8 , return_tensors='pt' ) return DataLoader(lowerCamelCase__ , shuffle=lowerCamelCase__ , collate_fn=lowerCamelCase__ , batch_size=1_6 ) def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ ) -> Optional[int]: __lowerCamelCase : Optional[int] = Accelerator(dispatch_batches=lowerCamelCase__ , split_batches=lowerCamelCase__ ) __lowerCamelCase : Union[str, Any] = get_dataloader(lowerCamelCase__ , not dispatch_batches ) __lowerCamelCase : Union[str, Any] = AutoModelForSequenceClassification.from_pretrained( 'hf-internal-testing/mrpc-bert-base-cased' , return_dict=lowerCamelCase__ ) __lowerCamelCase , __lowerCamelCase : Tuple = accelerator.prepare(lowerCamelCase__ , lowerCamelCase__ ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) -> List[Any]: __lowerCamelCase : str = [] for batch in dataloader: __lowerCamelCase , __lowerCamelCase : Union[str, Any] = batch.values() with torch.no_grad(): __lowerCamelCase : Tuple = model(lowerCamelCase__ ) __lowerCamelCase , __lowerCamelCase : Optional[Any] = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) __lowerCamelCase , __lowerCamelCase : Dict = [], [] for logit, targ in logits_and_targets: logits.append(lowerCamelCase__ ) targs.append(lowerCamelCase__ ) __lowerCamelCase , __lowerCamelCase : Union[str, Any] = torch.cat(lowerCamelCase__ ), torch.cat(lowerCamelCase__ ) return logits, targs def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ , lowerCamelCase__=8_2 , lowerCamelCase__=False , lowerCamelCase__=False , lowerCamelCase__=1_6 ) -> Dict: __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : Optional[Any] = get_basic_setup(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) __lowerCamelCase , __lowerCamelCase : Dict = generate_predictions(lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) assert ( len(lowerCamelCase__ ) == num_samples ), F"Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(lowerCamelCase__ )}" def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ = False , lowerCamelCase__ = False ) -> Dict: __lowerCamelCase : Dict = evaluate.load('glue' , 'mrpc' ) __lowerCamelCase , __lowerCamelCase : Optional[int] = get_mrpc_setup(lowerCamelCase__ , lowerCamelCase__ ) # First do baseline __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : int = setup['no'] model.to(lowerCamelCase__ ) model.eval() for batch in dataloader: batch.to(lowerCamelCase__ ) with torch.inference_mode(): __lowerCamelCase : Dict = model(**lowerCamelCase__ ) __lowerCamelCase : Any = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=lowerCamelCase__ , references=batch['labels'] ) __lowerCamelCase : str = metric.compute() # Then do distributed __lowerCamelCase , __lowerCamelCase , __lowerCamelCase : Union[str, Any] = setup['ddp'] model.eval() for batch in dataloader: with torch.inference_mode(): __lowerCamelCase : List[str] = model(**lowerCamelCase__ ) __lowerCamelCase : List[Any] = outputs.logits.argmax(dim=-1 ) __lowerCamelCase : List[str] = batch['labels'] __lowerCamelCase , __lowerCamelCase : Union[str, Any] = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=lowerCamelCase__ , references=lowerCamelCase__ ) __lowerCamelCase : Dict = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), F"Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n" def SCREAMING_SNAKE_CASE__ ( ) -> List[str]: __lowerCamelCase : int = Accelerator(split_batches=lowerCamelCase__ , dispatch_batches=lowerCamelCase__ ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print('**Testing gather_for_metrics**' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(F"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`" ) test_mrpc(lowerCamelCase__ , lowerCamelCase__ ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('**Test torch metrics**' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: __lowerCamelCase : Optional[Any] = Accelerator(split_batches=lowerCamelCase__ , dispatch_batches=lowerCamelCase__ ) if accelerator.is_local_main_process: print(F"With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99" ) test_torch_metrics(lowerCamelCase__ , 9_9 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('**Test last batch is not dropped when perfectly divisible**' ) __lowerCamelCase : Dict = Accelerator() test_torch_metrics(lowerCamelCase__ , 5_1_2 ) accelerator.state._reset_state() def SCREAMING_SNAKE_CASE__ ( lowerCamelCase__ ) -> Optional[Any]: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
113
1
import numpy as np from matplotlib import pyplot as plt from sklearn.datasets import load_iris from sklearn.metrics import ConfusionMatrixDisplay from sklearn.model_selection import train_test_split from xgboost import XGBClassifier def lowerCamelCase_ ( lowerCamelCase__ ): return (data["data"], data["target"]) def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ ): lowerCamelCase_ = XGBClassifier() classifier.fit(lowerCamelCase__ , lowerCamelCase__ ) return classifier def lowerCamelCase_ ( ): lowerCamelCase_ = load_iris() lowerCamelCase_ , lowerCamelCase_ = data_handling(lowerCamelCase__ ) lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = train_test_split( lowerCamelCase__ , lowerCamelCase__ , test_size=0.25 ) lowerCamelCase_ = iris["target_names"] # Create an XGBoost Classifier from the training data lowerCamelCase_ = xgboost(lowerCamelCase__ , lowerCamelCase__ ) # Display the confusion matrix of the classifier with both training and test sets ConfusionMatrixDisplay.from_estimator( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , display_labels=lowerCamelCase__ , cmap="Blues" , normalize="true" , ) plt.title("Normalized Confusion Matrix - IRIS Dataset" ) plt.show() if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
19
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase__ : str = { '''configuration_x_clip''': [ '''XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XCLIPConfig''', '''XCLIPTextConfig''', '''XCLIPVisionConfig''', ], '''processing_x_clip''': ['''XCLIPProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase__ : Tuple = [ '''XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XCLIPModel''', '''XCLIPPreTrainedModel''', '''XCLIPTextModel''', '''XCLIPVisionModel''', ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys lowercase__ : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
264
0
'''simple docstring''' from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
67
'''simple docstring''' from __future__ import annotations _lowerCamelCase = [ [-1, 0], # left [0, -1], # down [1, 0], # right [0, 1], # up ] def a__ ( _SCREAMING_SNAKE_CASE : list[list[int]] , _SCREAMING_SNAKE_CASE : list[int] , _SCREAMING_SNAKE_CASE : list[int] , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : list[list[int]] , ) -> tuple[list[list[int]], list[list[int]]]: """simple docstring""" UpperCAmelCase_ : int = [ [0 for col in range(len(grid[0] ) )] for row in range(len(_SCREAMING_SNAKE_CASE ) ) ] # the reference grid UpperCAmelCase_ : Tuple = 1 UpperCAmelCase_ : str = [ [0 for col in range(len(grid[0] ) )] for row in range(len(_SCREAMING_SNAKE_CASE ) ) ] # the action grid UpperCAmelCase_ : Tuple = init[0] UpperCAmelCase_ : List[Any] = init[1] UpperCAmelCase_ : Optional[Any] = 0 UpperCAmelCase_ : Optional[Any] = g + heuristic[x][y] # cost from starting cell to destination cell UpperCAmelCase_ : List[str] = [[f, g, x, y]] UpperCAmelCase_ : Tuple = False # flag that is set when search is complete UpperCAmelCase_ : Union[str, Any] = False # flag set if we can't find expand while not found and not resign: if len(_SCREAMING_SNAKE_CASE ) == 0: raise ValueError("Algorithm is unable to find solution" ) else: # to choose the least costliest action so as to move closer to the goal cell.sort() cell.reverse() UpperCAmelCase_ : Dict = cell.pop() UpperCAmelCase_ : Tuple = next_cell[2] UpperCAmelCase_ : str = next_cell[3] UpperCAmelCase_ : List[Any] = next_cell[1] if x == goal[0] and y == goal[1]: UpperCAmelCase_ : Optional[Any] = True else: for i in range(len(_SCREAMING_SNAKE_CASE ) ): # to try out different valid actions UpperCAmelCase_ : Union[str, Any] = x + DIRECTIONS[i][0] UpperCAmelCase_ : Optional[Any] = y + DIRECTIONS[i][1] if xa >= 0 and xa < len(_SCREAMING_SNAKE_CASE ) and ya >= 0 and ya < len(grid[0] ): if closed[xa][ya] == 0 and grid[xa][ya] == 0: UpperCAmelCase_ : Any = g + cost UpperCAmelCase_ : Optional[Any] = ga + heuristic[xa][ya] cell.append([fa, ga, xa, ya] ) UpperCAmelCase_ : List[Any] = 1 UpperCAmelCase_ : List[str] = i UpperCAmelCase_ : Union[str, Any] = [] UpperCAmelCase_ : List[Any] = goal[0] UpperCAmelCase_ : str = goal[1] invpath.append([x, y] ) # we get the reverse path from here while x != init[0] or y != init[1]: UpperCAmelCase_ : Optional[int] = x - DIRECTIONS[action[x][y]][0] UpperCAmelCase_ : Optional[int] = y - DIRECTIONS[action[x][y]][1] UpperCAmelCase_ : Optional[Any] = xa UpperCAmelCase_ : List[str] = ya invpath.append([x, y] ) UpperCAmelCase_ : Tuple = [] for i in range(len(_SCREAMING_SNAKE_CASE ) ): path.append(invpath[len(_SCREAMING_SNAKE_CASE ) - 1 - i] ) return path, action if __name__ == "__main__": _lowerCamelCase = [ [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 1, 0, 0, 0, 0], [0, 1, 0, 0, 1, 0], [0, 0, 0, 0, 1, 0], ] _lowerCamelCase = [0, 0] # all coordinates are given in format [y,x] _lowerCamelCase = [len(grid) - 1, len(grid[0]) - 1] _lowerCamelCase = 1 # the cost map which pushes the path closer to the goal _lowerCamelCase = [[0 for row in range(len(grid[0]))] for col in range(len(grid))] for i in range(len(grid)): for j in range(len(grid[0])): _lowerCamelCase = abs(i - goal[0]) + abs(j - goal[1]) if grid[i][j] == 1: # added extra penalty in the heuristic map _lowerCamelCase = 99 _lowerCamelCase , _lowerCamelCase = search(grid, init, goal, cost, heuristic) print("""ACTION MAP""") for i in range(len(action)): print(action[i]) for i in range(len(path)): print(path[i])
67
1
from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING __A = logging.get_logger(__name__) @add_end_docstrings(__magic_name__ ) class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" def __init__( self , **lowerCamelCase__ ) -> Any: '''simple docstring''' super().__init__(**lowerCamelCase__ ) requires_backends(self , 'vision' ) requires_backends(self , 'torch' ) if self.framework != "pt": raise ValueError(f"""The {self.__class__} is only available in PyTorch.""" ) self.check_model_type(lowerCamelCase__ ) def lowercase_ ( self , **lowerCamelCase__ ) -> List[Any]: '''simple docstring''' __lowerCamelCase = {} __lowerCamelCase = {} __lowerCamelCase = {} # preprocess args if "points_per_batch" in kwargs: __lowerCamelCase = kwargs['points_per_batch'] if "points_per_crop" in kwargs: __lowerCamelCase = kwargs['points_per_crop'] if "crops_n_layers" in kwargs: __lowerCamelCase = kwargs['crops_n_layers'] if "crop_overlap_ratio" in kwargs: __lowerCamelCase = kwargs['crop_overlap_ratio'] if "crop_n_points_downscale_factor" in kwargs: __lowerCamelCase = kwargs['crop_n_points_downscale_factor'] # postprocess args if "pred_iou_thresh" in kwargs: __lowerCamelCase = kwargs['pred_iou_thresh'] if "stability_score_offset" in kwargs: __lowerCamelCase = kwargs['stability_score_offset'] if "mask_threshold" in kwargs: __lowerCamelCase = kwargs['mask_threshold'] if "stability_score_thresh" in kwargs: __lowerCamelCase = kwargs['stability_score_thresh'] if "crops_nms_thresh" in kwargs: __lowerCamelCase = kwargs['crops_nms_thresh'] if "output_rle_mask" in kwargs: __lowerCamelCase = kwargs['output_rle_mask'] if "output_bboxes_mask" in kwargs: __lowerCamelCase = kwargs['output_bboxes_mask'] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__( self , lowerCamelCase__ , *lowerCamelCase__ , lowerCamelCase__=None , lowerCamelCase__=None , **lowerCamelCase__ ) -> List[Any]: '''simple docstring''' return super().__call__(lowerCamelCase__ , *lowerCamelCase__ , num_workers=lowerCamelCase__ , batch_size=lowerCamelCase__ , **lowerCamelCase__ ) def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__=64 , lowerCamelCase__ = 0 , lowerCamelCase__ = 512 / 1_500 , lowerCamelCase__ = 32 , lowerCamelCase__ = 1 , ) -> Optional[int]: '''simple docstring''' __lowerCamelCase = load_image(lowerCamelCase__ ) __lowerCamelCase = self.image_processor.size['longest_edge'] __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = self.image_processor.generate_crop_boxes( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) __lowerCamelCase = self.image_processor(images=lowerCamelCase__ , return_tensors='pt' ) with self.device_placement(): if self.framework == "pt": __lowerCamelCase = self.get_inference_context() with inference_context(): __lowerCamelCase = self._ensure_tensor_on_device(lowerCamelCase__ , device=self.device ) __lowerCamelCase = self.model.get_image_embeddings(model_inputs.pop('pixel_values' ) ) __lowerCamelCase = image_embeddings __lowerCamelCase = grid_points.shape[1] __lowerCamelCase = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( 'Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. ' 'To return all points at once, set points_per_batch to None' ) for i in range(0 , lowerCamelCase__ , lowerCamelCase__ ): __lowerCamelCase = grid_points[:, i : i + points_per_batch, :, :] __lowerCamelCase = input_labels[:, i : i + points_per_batch] __lowerCamelCase = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__=0.88 , lowerCamelCase__=0.95 , lowerCamelCase__=0 , lowerCamelCase__=1 , ) -> Any: '''simple docstring''' __lowerCamelCase = model_inputs.pop('input_boxes' ) __lowerCamelCase = model_inputs.pop('is_last' ) __lowerCamelCase = model_inputs.pop('original_sizes' ).tolist() __lowerCamelCase = model_inputs.pop('reshaped_input_sizes' ).tolist() __lowerCamelCase = self.model(**lowerCamelCase__ ) # post processing happens here in order to avoid CPU GPU copies of ALL the masks __lowerCamelCase = model_outputs['pred_masks'] __lowerCamelCase = self.image_processor.post_process_masks( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , binarize=lowerCamelCase__ ) __lowerCamelCase = model_outputs['iou_scores'] __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = self.image_processor.filter_masks( masks[0] , iou_scores[0] , original_sizes[0] , input_boxes[0] , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def lowercase_ ( self , lowerCamelCase__ , lowerCamelCase__=False , lowerCamelCase__=False , lowerCamelCase__=0.7 , ) -> Any: '''simple docstring''' __lowerCamelCase = [] __lowerCamelCase = [] __lowerCamelCase = [] for model_output in model_outputs: all_scores.append(model_output.pop('iou_scores' ) ) all_masks.extend(model_output.pop('masks' ) ) all_boxes.append(model_output.pop('boxes' ) ) __lowerCamelCase = torch.cat(lowerCamelCase__ ) __lowerCamelCase = torch.cat(lowerCamelCase__ ) __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = self.image_processor.post_process_for_mask_generation( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ) __lowerCamelCase = defaultdict(lowerCamelCase__ ) for output in model_outputs: for k, v in output.items(): extra[k].append(lowerCamelCase__ ) __lowerCamelCase = {} if output_rle_mask: __lowerCamelCase = rle_mask if output_bboxes_mask: __lowerCamelCase = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
90
class __lowercase : """simple docstring""" def __init__( self : List[Any] , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : List[Any]): SCREAMING_SNAKE_CASE_: List[str] = name SCREAMING_SNAKE_CASE_: Union[str, Any] = val def __str__( self : Dict): return F"{self.__class__.__name__}({self.name}, {self.val})" def __lt__( self : List[str] , lowerCAmelCase__ : Any): return self.val < other.val class __lowercase : """simple docstring""" def __init__( self : Tuple , lowerCAmelCase__ : Dict): SCREAMING_SNAKE_CASE_: str = {} SCREAMING_SNAKE_CASE_: int = {} SCREAMING_SNAKE_CASE_: Any = self.build_heap(lowerCAmelCase__) def __getitem__( self : List[Any] , lowerCAmelCase__ : Dict): return self.get_value(lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowerCAmelCase__ : Dict): return (idx - 1) // 2 def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase__ : Optional[Any]): return idx * 2 + 1 def _SCREAMING_SNAKE_CASE ( self : Optional[int] , lowerCAmelCase__ : Tuple): return idx * 2 + 2 def _SCREAMING_SNAKE_CASE ( self : Optional[int] , lowerCAmelCase__ : Optional[int]): return self.heap_dict[key] def _SCREAMING_SNAKE_CASE ( self : List[Any] , lowerCAmelCase__ : Union[str, Any]): SCREAMING_SNAKE_CASE_: Tuple = len(lowerCAmelCase__) - 1 SCREAMING_SNAKE_CASE_: List[str] = self.get_parent_idx(lowerCAmelCase__) for idx, i in enumerate(lowerCAmelCase__): SCREAMING_SNAKE_CASE_: Union[str, Any] = idx SCREAMING_SNAKE_CASE_: str = i.val for i in range(lowerCAmelCase__ , -1 , -1): self.sift_down(lowerCAmelCase__ , lowerCAmelCase__) return array def _SCREAMING_SNAKE_CASE ( self : str , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : List[str]): while True: SCREAMING_SNAKE_CASE_: Optional[Any] = self.get_left_child_idx(lowerCAmelCase__) # noqa: E741 SCREAMING_SNAKE_CASE_: Dict = self.get_right_child_idx(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: int = idx if l < len(lowerCAmelCase__) and array[l] < array[idx]: SCREAMING_SNAKE_CASE_: List[str] = l if r < len(lowerCAmelCase__) and array[r] < array[smallest]: SCREAMING_SNAKE_CASE_: str = r if smallest != idx: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Any = array[smallest], array[idx] ( ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ): Optional[Any] = ( self.idx_of_element[array[smallest]], self.idx_of_element[array[idx]], ) SCREAMING_SNAKE_CASE_: Optional[int] = smallest else: break def _SCREAMING_SNAKE_CASE ( self : Dict , lowerCAmelCase__ : str): SCREAMING_SNAKE_CASE_: Any = self.get_parent_idx(lowerCAmelCase__) while p >= 0 and self.heap[p] > self.heap[idx]: SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: List[Any] = self.heap[idx], self.heap[p] SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Tuple = ( self.idx_of_element[self.heap[idx]], self.idx_of_element[self.heap[p]], ) SCREAMING_SNAKE_CASE_: Union[str, Any] = p SCREAMING_SNAKE_CASE_: Optional[int] = self.get_parent_idx(lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : List[Any]): return self.heap[0] def _SCREAMING_SNAKE_CASE ( self : Dict): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Tuple = self.heap[-1], self.heap[0] SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: List[str] = ( self.idx_of_element[self.heap[-1]], self.idx_of_element[self.heap[0]], ) SCREAMING_SNAKE_CASE_: int = self.heap.pop() del self.idx_of_element[x] self.sift_down(0 , self.heap) return x def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase__ : Tuple): self.heap.append(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[str] = len(self.heap) - 1 SCREAMING_SNAKE_CASE_: List[str] = node.val self.sift_up(len(self.heap) - 1) def _SCREAMING_SNAKE_CASE ( self : List[Any]): return len(self.heap) == 0 def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowerCAmelCase__ : Any , lowerCAmelCase__ : Optional[int]): assert ( self.heap[self.idx_of_element[node]].val > new_value ), "newValue must be less that current value" SCREAMING_SNAKE_CASE_: Any = new_value SCREAMING_SNAKE_CASE_: Tuple = new_value self.sift_up(self.idx_of_element[node]) lowerCAmelCase : int = Node("""R""", -1) lowerCAmelCase : str = Node("""B""", 6) lowerCAmelCase : str = Node("""A""", 3) lowerCAmelCase : List[str] = Node("""X""", 1) lowerCAmelCase : Union[str, Any] = Node("""E""", 4) # Use one of these two ways to generate Min-Heap # Generating Min-Heap from array lowerCAmelCase : Optional[Any] = MinHeap([r, b, a, x, e]) # Generating Min-Heap by Insert method # myMinHeap.insert(a) # myMinHeap.insert(b) # myMinHeap.insert(x) # myMinHeap.insert(r) # myMinHeap.insert(e) # Before print("""Min Heap - before decrease key""") for i in my_min_heap.heap: print(i) print("""Min Heap - After decrease key of node [B -> -17]""") my_min_heap.decrease_key(b, -17) # After for i in my_min_heap.heap: print(i) if __name__ == "__main__": import doctest doctest.testmod()
13
0
from argparse import ArgumentParser, Namespace from ..utils import logging from . import BaseTransformersCLICommand def __SCREAMING_SNAKE_CASE ( __UpperCAmelCase ): return ConvertCommand( args.model_type , args.tf_checkpoint , args.pytorch_dump_output , args.config , args.finetuning_task_name ) UpperCAmelCase: Optional[Any] = '\ntransformers can only be used from the commandline to convert TensorFlow models in PyTorch, In that case, it requires\nTensorFlow to be installed. Please see https://www.tensorflow.org/install/ for installation instructions.\n' class UpperCamelCase ( lowercase__ ): """simple docstring""" @staticmethod def lowerCamelCase__ ( UpperCAmelCase_ ): _lowercase : Optional[int] = parser.add_parser( """convert""" ,help="""CLI tool to run convert model from original author checkpoints to Transformers PyTorch checkpoints.""" ,) train_parser.add_argument("""--model_type""" ,type=_UpperCamelCase ,required=_UpperCamelCase ,help="""Model's type.""" ) train_parser.add_argument( """--tf_checkpoint""" ,type=_UpperCamelCase ,required=_UpperCamelCase ,help="""TensorFlow checkpoint path or folder.""" ) train_parser.add_argument( """--pytorch_dump_output""" ,type=_UpperCamelCase ,required=_UpperCamelCase ,help="""Path to the PyTorch saved model output.""" ) train_parser.add_argument("""--config""" ,type=_UpperCamelCase ,default="""""" ,help="""Configuration file path or folder.""" ) train_parser.add_argument( """--finetuning_task_name""" ,type=_UpperCamelCase ,default=_UpperCamelCase ,help="""Optional fine-tuning task name if the TF model was a finetuned model.""" ,) train_parser.set_defaults(func=_UpperCamelCase ) def __init__( self ,UpperCAmelCase_ ,UpperCAmelCase_ ,UpperCAmelCase_ ,UpperCAmelCase_ ,UpperCAmelCase_ ,*UpperCAmelCase_ ,): _lowercase : Dict = logging.get_logger("""transformers-cli/converting""" ) self._logger.info(f"""Loading model {model_type}""" ) _lowercase : str = model_type _lowercase : Dict = tf_checkpoint _lowercase : List[Any] = pytorch_dump_output _lowercase : int = config _lowercase : int = finetuning_task_name def lowerCamelCase__ ( self ): if self._model_type == "albert": try: from ..models.albert.convert_albert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_UpperCamelCase ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint ,self._config ,self._pytorch_dump_output ) elif self._model_type == "bert": try: from ..models.bert.convert_bert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_UpperCamelCase ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint ,self._config ,self._pytorch_dump_output ) elif self._model_type == "funnel": try: from ..models.funnel.convert_funnel_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_UpperCamelCase ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint ,self._config ,self._pytorch_dump_output ) elif self._model_type == "t5": try: from ..models.ta.convert_ta_original_tf_checkpoint_to_pytorch import convert_tf_checkpoint_to_pytorch except ImportError: raise ImportError(_UpperCamelCase ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint ,self._config ,self._pytorch_dump_output ) elif self._model_type == "gpt": from ..models.openai.convert_openai_original_tf_checkpoint_to_pytorch import ( convert_openai_checkpoint_to_pytorch, ) convert_openai_checkpoint_to_pytorch(self._tf_checkpoint ,self._config ,self._pytorch_dump_output ) elif self._model_type == "transfo_xl": try: from ..models.transfo_xl.convert_transfo_xl_original_tf_checkpoint_to_pytorch import ( convert_transfo_xl_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_UpperCamelCase ) if "ckpt" in self._tf_checkpoint.lower(): _lowercase : Optional[int] = self._tf_checkpoint _lowercase : List[str] = """""" else: _lowercase : Any = self._tf_checkpoint _lowercase : Tuple = """""" convert_transfo_xl_checkpoint_to_pytorch( _UpperCamelCase ,self._config ,self._pytorch_dump_output ,_UpperCamelCase ) elif self._model_type == "gpt2": try: from ..models.gpta.convert_gpta_original_tf_checkpoint_to_pytorch import ( convert_gpta_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_UpperCamelCase ) convert_gpta_checkpoint_to_pytorch(self._tf_checkpoint ,self._config ,self._pytorch_dump_output ) elif self._model_type == "xlnet": try: from ..models.xlnet.convert_xlnet_original_tf_checkpoint_to_pytorch import ( convert_xlnet_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_UpperCamelCase ) convert_xlnet_checkpoint_to_pytorch( self._tf_checkpoint ,self._config ,self._pytorch_dump_output ,self._finetuning_task_name ) elif self._model_type == "xlm": from ..models.xlm.convert_xlm_original_pytorch_checkpoint_to_pytorch import ( convert_xlm_checkpoint_to_pytorch, ) convert_xlm_checkpoint_to_pytorch(self._tf_checkpoint ,self._pytorch_dump_output ) elif self._model_type == "lxmert": from ..models.lxmert.convert_lxmert_original_tf_checkpoint_to_pytorch import ( convert_lxmert_checkpoint_to_pytorch, ) convert_lxmert_checkpoint_to_pytorch(self._tf_checkpoint ,self._pytorch_dump_output ) elif self._model_type == "rembert": from ..models.rembert.convert_rembert_tf_checkpoint_to_pytorch import ( convert_rembert_tf_checkpoint_to_pytorch, ) convert_rembert_tf_checkpoint_to_pytorch(self._tf_checkpoint ,self._config ,self._pytorch_dump_output ) else: raise ValueError( """--model_type should be selected in the list [bert, gpt, gpt2, t5, transfo_xl, xlnet, xlm, lxmert]""" )
358
"""simple docstring""" import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING UpperCAmelCase: Any = logging.get_logger(__name__) UpperCAmelCase: List[str] = { """Salesforce/instruct-blip-flan-t5""": """https://huggingface.co/Salesforce/instruct-blip-flan-t5/resolve/main/config.json""", } class UpperCamelCase ( snake_case ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Optional[Any] = "instructblip_vision_model" def __init__( self ,UpperCAmelCase_=14_08 ,UpperCAmelCase_=61_44 ,UpperCAmelCase_=39 ,UpperCAmelCase_=16 ,UpperCAmelCase_=2_24 ,UpperCAmelCase_=14 ,UpperCAmelCase_="gelu" ,UpperCAmelCase_=1E-6 ,UpperCAmelCase_=0.0 ,UpperCAmelCase_=1E-10 ,UpperCAmelCase_=True ,**UpperCAmelCase_ ,): super().__init__(**UpperCAmelCase_ ) _lowercase : Optional[Any] = hidden_size _lowercase : Tuple = intermediate_size _lowercase : List[Any] = num_hidden_layers _lowercase : Tuple = num_attention_heads _lowercase : Optional[Any] = patch_size _lowercase : Optional[Any] = image_size _lowercase : Union[str, Any] = initializer_range _lowercase : Optional[Any] = attention_dropout _lowercase : List[Any] = layer_norm_eps _lowercase : Optional[int] = hidden_act _lowercase : Tuple = qkv_bias @classmethod def lowerCamelCase__ ( cls ,UpperCAmelCase_ ,**UpperCAmelCase_ ): cls._set_token_in_kwargs(UpperCAmelCase_ ) _lowercase , _lowercase : List[Any] = cls.get_config_dict(UpperCAmelCase_ ,**UpperCAmelCase_ ) # get the vision config dict if we are loading from InstructBlipConfig if config_dict.get("""model_type""" ) == "instructblip": _lowercase : int = config_dict["""vision_config"""] if "model_type" in config_dict and hasattr(cls ,"""model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(UpperCAmelCase_ ,**UpperCAmelCase_ ) class UpperCamelCase ( snake_case ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = "instructblip_qformer" def __init__( self ,UpperCAmelCase_=3_05_22 ,UpperCAmelCase_=7_68 ,UpperCAmelCase_=12 ,UpperCAmelCase_=12 ,UpperCAmelCase_=30_72 ,UpperCAmelCase_="gelu" ,UpperCAmelCase_=0.1 ,UpperCAmelCase_=0.1 ,UpperCAmelCase_=5_12 ,UpperCAmelCase_=0.02 ,UpperCAmelCase_=1E-12 ,UpperCAmelCase_=0 ,UpperCAmelCase_="absolute" ,UpperCAmelCase_=2 ,UpperCAmelCase_=14_08 ,**UpperCAmelCase_ ,): super().__init__(pad_token_id=UpperCAmelCase_ ,**UpperCAmelCase_ ) _lowercase : List[Any] = vocab_size _lowercase : List[Any] = hidden_size _lowercase : str = num_hidden_layers _lowercase : List[str] = num_attention_heads _lowercase : Optional[Any] = hidden_act _lowercase : int = intermediate_size _lowercase : Union[str, Any] = hidden_dropout_prob _lowercase : Optional[Any] = attention_probs_dropout_prob _lowercase : List[Any] = max_position_embeddings _lowercase : Tuple = initializer_range _lowercase : Optional[int] = layer_norm_eps _lowercase : Any = position_embedding_type _lowercase : Dict = cross_attention_frequency _lowercase : Optional[Any] = encoder_hidden_size @classmethod def lowerCamelCase__ ( cls ,UpperCAmelCase_ ,**UpperCAmelCase_ ): cls._set_token_in_kwargs(UpperCAmelCase_ ) _lowercase , _lowercase : Dict = cls.get_config_dict(UpperCAmelCase_ ,**UpperCAmelCase_ ) # get the qformer config dict if we are loading from InstructBlipConfig if config_dict.get("""model_type""" ) == "instructblip": _lowercase : str = config_dict["""qformer_config"""] if "model_type" in config_dict and hasattr(cls ,"""model_type""" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(UpperCAmelCase_ ,**UpperCAmelCase_ ) class UpperCamelCase ( snake_case ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Any = "instructblip" SCREAMING_SNAKE_CASE_ : List[str] = True def __init__( self ,UpperCAmelCase_=None ,UpperCAmelCase_=None ,UpperCAmelCase_=None ,UpperCAmelCase_=32 ,**UpperCAmelCase_ ): super().__init__(**UpperCAmelCase_ ) if vision_config is None: _lowercase : str = {} logger.info("""vision_config is None. initializing the InstructBlipVisionConfig with default values.""" ) if qformer_config is None: _lowercase : Any = {} logger.info("""qformer_config is None. Initializing the InstructBlipQFormerConfig with default values.""" ) if text_config is None: _lowercase : Optional[int] = {} logger.info("""text_config is None. Initializing the text config with default values (`OPTConfig`).""" ) _lowercase : int = InstructBlipVisionConfig(**UpperCAmelCase_ ) _lowercase : Optional[int] = InstructBlipQFormerConfig(**UpperCAmelCase_ ) _lowercase : Dict = text_config["""model_type"""] if """model_type""" in text_config else """opt""" _lowercase : str = CONFIG_MAPPING[text_model_type](**UpperCAmelCase_ ) _lowercase : str = self.text_config.tie_word_embeddings _lowercase : Union[str, Any] = self.text_config.is_encoder_decoder _lowercase : List[str] = num_query_tokens _lowercase : List[str] = self.vision_config.hidden_size _lowercase : Dict = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES _lowercase : Union[str, Any] = 1.0 _lowercase : Dict = 0.02 @classmethod def lowerCamelCase__ ( cls ,UpperCAmelCase_ ,UpperCAmelCase_ ,UpperCAmelCase_ ,**UpperCAmelCase_ ,): return cls( vision_config=vision_config.to_dict() ,qformer_config=qformer_config.to_dict() ,text_config=text_config.to_dict() ,**UpperCAmelCase_ ,) def lowerCamelCase__ ( self ): _lowercase : Union[str, Any] = copy.deepcopy(self.__dict__ ) _lowercase : int = self.vision_config.to_dict() _lowercase : Any = self.qformer_config.to_dict() _lowercase : Any = self.text_config.to_dict() _lowercase : Optional[int] = self.__class__.model_type return output
336
0
import argparse import logging import os from pathlib import Path from typing import Any, Dict import pytorch_lightning as pl from pytorch_lightning.utilities import rank_zero_info from transformers import ( AdamW, AutoConfig, AutoModel, AutoModelForPreTraining, AutoModelForQuestionAnswering, AutoModelForSeqaSeqLM, AutoModelForSequenceClassification, AutoModelForTokenClassification, AutoModelWithLMHead, AutoTokenizer, PretrainedConfig, PreTrainedTokenizer, ) from transformers.optimization import ( Adafactor, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.utils.versions import require_version lowerCAmelCase__ = logging.getLogger(__name__) require_version('pytorch_lightning>=1.0.4') lowerCAmelCase__ = { 'base': AutoModel, 'sequence-classification': AutoModelForSequenceClassification, 'question-answering': AutoModelForQuestionAnswering, 'pretraining': AutoModelForPreTraining, 'token-classification': AutoModelForTokenClassification, 'language-modeling': AutoModelWithLMHead, 'summarization': AutoModelForSeqaSeqLM, 'translation': AutoModelForSeqaSeqLM, } # update this and the import above to support new schedulers from transformers.optimization lowerCAmelCase__ = { 'linear': get_linear_schedule_with_warmup, 'cosine': get_cosine_schedule_with_warmup, 'cosine_w_restarts': get_cosine_with_hard_restarts_schedule_with_warmup, 'polynomial': get_polynomial_decay_schedule_with_warmup, # '': get_constant_schedule, # not supported for now # '': get_constant_schedule_with_warmup, # not supported for now } lowerCAmelCase__ = sorted(arg_to_scheduler.keys()) lowerCAmelCase__ = '{' + ', '.join(arg_to_scheduler_choices) + '}' class lowerCAmelCase__ ( pl.LightningModule): '''simple docstring''' def __init__( self , __lowerCamelCase , __lowerCamelCase=None , __lowerCamelCase="base" , __lowerCamelCase=None , __lowerCamelCase=None , __lowerCamelCase=None , **__lowerCamelCase , ) -> int: super().__init__() # TODO: move to self.save_hyperparameters() # self.save_hyperparameters() # can also expand arguments into trainer signature for easier reading self.save_hyperparameters(__lowerCamelCase) _A : Tuple = 0 _A : Union[str, Any] = Path(self.hparams.output_dir) _A : str = self.hparams.cache_dir if self.hparams.cache_dir else None if config is None: _A : List[Any] = AutoConfig.from_pretrained( self.hparams.config_name if self.hparams.config_name else self.hparams.model_name_or_path , **({"num_labels": num_labels} if num_labels is not None else {}) , cache_dir=__lowerCamelCase , **__lowerCamelCase , ) else: _A : PretrainedConfig = config _A : Optional[int] = ("encoder_layerdrop", "decoder_layerdrop", "dropout", "attention_dropout") for p in extra_model_params: if getattr(self.hparams , __lowerCamelCase , __lowerCamelCase): assert hasattr(self.config , __lowerCamelCase), F"model config doesn't have a `{p}` attribute" setattr(self.config , __lowerCamelCase , getattr(self.hparams , __lowerCamelCase)) if tokenizer is None: _A : Dict = AutoTokenizer.from_pretrained( self.hparams.tokenizer_name if self.hparams.tokenizer_name else self.hparams.model_name_or_path , cache_dir=__lowerCamelCase , ) else: _A : PreTrainedTokenizer = tokenizer _A : List[Any] = MODEL_MODES[mode] if model is None: _A : Optional[Any] = self.model_type.from_pretrained( self.hparams.model_name_or_path , from_tf=bool(".ckpt" in self.hparams.model_name_or_path) , config=self.config , cache_dir=__lowerCamelCase , ) else: _A : Any = model def _lowerCamelCase ( self , *__lowerCamelCase , **__lowerCamelCase) -> Union[str, Any]: _A : List[str] = self.model_type.from_pretrained(*__lowerCamelCase , **__lowerCamelCase) def _lowerCamelCase ( self) -> List[str]: _A : Tuple = arg_to_scheduler[self.hparams.lr_scheduler] _A : int = get_schedule_func( self.opt , num_warmup_steps=self.hparams.warmup_steps , num_training_steps=self.total_steps()) _A : Union[str, Any] = {"scheduler": scheduler, "interval": "step", "frequency": 1} return scheduler def _lowerCamelCase ( self) -> Union[str, Any]: _A : Any = self.model _A : Optional[int] = ["bias", "LayerNorm.weight"] _A : int = [ { "params": [ p for n, p in model.named_parameters() if not any(nd in n for nd in no_decay) ], # check this named paramters "weight_decay": self.hparams.weight_decay, }, { "params": [p for n, p in model.named_parameters() if any(nd in n for nd in no_decay)], "weight_decay": 0.0, }, ] if self.hparams.adafactor: _A : Optional[int] = Adafactor( __lowerCamelCase , lr=self.hparams.learning_rate , scale_parameter=__lowerCamelCase , relative_step=__lowerCamelCase) else: _A : List[Any] = AdamW( __lowerCamelCase , lr=self.hparams.learning_rate , eps=self.hparams.adam_epsilon) _A : Union[str, Any] = optimizer _A : Optional[Any] = self.get_lr_scheduler() return [optimizer], [scheduler] def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase) -> Dict: return self.validation_step(__lowerCamelCase , __lowerCamelCase) def _lowerCamelCase ( self , __lowerCamelCase) -> List[Any]: return self.validation_end(__lowerCamelCase) def _lowerCamelCase ( self) -> int: _A : int = max(1 , self.hparams.gpus) # TODO: consider num_tpu_cores _A : Union[str, Any] = self.hparams.train_batch_size * self.hparams.accumulate_grad_batches * num_devices return (self.dataset_size / effective_batch_size) * self.hparams.max_epochs def _lowerCamelCase ( self , __lowerCamelCase) -> Dict: if stage == "test": _A : Tuple = len(self.test_dataloader().dataset) else: _A : List[str] = self.get_dataloader("train" , self.hparams.train_batch_size , shuffle=__lowerCamelCase) _A : List[str] = len(self.train_dataloader().dataset) def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase = False) -> Any: raise NotImplementedError("You must implement this for your task") def _lowerCamelCase ( self) -> Any: return self.train_loader def _lowerCamelCase ( self) -> List[Any]: return self.get_dataloader("dev" , self.hparams.eval_batch_size , shuffle=__lowerCamelCase) def _lowerCamelCase ( self) -> List[Any]: return self.get_dataloader("test" , self.hparams.eval_batch_size , shuffle=__lowerCamelCase) def _lowerCamelCase ( self , __lowerCamelCase) -> int: return os.path.join( self.hparams.data_dir , "cached_{}_{}_{}".format( __lowerCamelCase , list(filter(__lowerCamelCase , self.hparams.model_name_or_path.split("/"))).pop() , str(self.hparams.max_seq_length) , ) , ) @pl.utilities.rank_zero_only def _lowerCamelCase ( self , __lowerCamelCase) -> None: _A : int = self.output_dir.joinpath("best_tfmr") _A : Optional[int] = self.step_count self.model.save_pretrained(__lowerCamelCase) self.tokenizer.save_pretrained(__lowerCamelCase) @staticmethod def _lowerCamelCase ( __lowerCamelCase , __lowerCamelCase) -> List[str]: parser.add_argument( "--model_name_or_path" , default=__lowerCamelCase , type=__lowerCamelCase , required=__lowerCamelCase , help="Path to pretrained model or model identifier from huggingface.co/models" , ) parser.add_argument( "--config_name" , default="" , type=__lowerCamelCase , help="Pretrained config name or path if not the same as model_name") parser.add_argument( "--tokenizer_name" , default=__lowerCamelCase , type=__lowerCamelCase , help="Pretrained tokenizer name or path if not the same as model_name" , ) parser.add_argument( "--cache_dir" , default=str(Path(__lowerCamelCase).parent / "test_run" / "cache") , type=__lowerCamelCase , help="Where do you want to store the pre-trained models downloaded from huggingface.co" , ) parser.add_argument( "--encoder_layerdrop" , type=__lowerCamelCase , help="Encoder layer dropout probability (Optional). Goes into model.config" , ) parser.add_argument( "--decoder_layerdrop" , type=__lowerCamelCase , help="Decoder layer dropout probability (Optional). Goes into model.config" , ) parser.add_argument( "--dropout" , type=__lowerCamelCase , help="Dropout probability (Optional). Goes into model.config" , ) parser.add_argument( "--attention_dropout" , type=__lowerCamelCase , help="Attention dropout probability (Optional). Goes into model.config" , ) parser.add_argument("--learning_rate" , default=5e-5 , type=__lowerCamelCase , help="The initial learning rate for Adam.") parser.add_argument( "--lr_scheduler" , default="linear" , choices=__lowerCamelCase , metavar=__lowerCamelCase , type=__lowerCamelCase , help="Learning rate scheduler" , ) parser.add_argument("--weight_decay" , default=0.0 , type=__lowerCamelCase , help="Weight decay if we apply some.") parser.add_argument("--adam_epsilon" , default=1e-8 , type=__lowerCamelCase , help="Epsilon for Adam optimizer.") parser.add_argument("--warmup_steps" , default=0 , type=__lowerCamelCase , help="Linear warmup over warmup_steps.") parser.add_argument("--num_workers" , default=4 , type=__lowerCamelCase , help="kwarg passed to DataLoader") parser.add_argument("--num_train_epochs" , dest="max_epochs" , default=3 , type=__lowerCamelCase) parser.add_argument("--train_batch_size" , default=3_2 , type=__lowerCamelCase) parser.add_argument("--eval_batch_size" , default=3_2 , type=__lowerCamelCase) parser.add_argument("--adafactor" , action="store_true") class lowerCAmelCase__ ( pl.Callback): '''simple docstring''' def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase) -> Optional[Any]: if ( trainer.is_global_zero and trainer.global_rank == 0 ): # we initialize the retriever only on master worker with RAY. In new pytorch-lightning accelorators are removed. pl_module.model.rag.retriever.init_retrieval() # better to use hook functions. class lowerCAmelCase__ ( pl.Callback): '''simple docstring''' def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase) -> Union[str, Any]: # print(pl_module.model.rag) for name, param in pl_module.model.rag.named_parameters(): if param.grad is None: print(__lowerCamelCase) class lowerCAmelCase__ ( pl.Callback): '''simple docstring''' def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase) -> Any: _A : int = trainer.lr_schedulers[0]["scheduler"] _A : Union[str, Any] = {F"lr_group_{i}": lr for i, lr in enumerate(lr_scheduler.get_lr())} pl_module.logger.log_metrics(__lowerCamelCase) def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase) -> Optional[Any]: rank_zero_info("***** Validation results *****") _A : Union[str, Any] = trainer.callback_metrics # Log results for key in sorted(__lowerCamelCase): if key not in ["log", "progress_bar"]: rank_zero_info("{} = {}\n".format(__lowerCamelCase , str(metrics[key]))) def _lowerCamelCase ( self , __lowerCamelCase , __lowerCamelCase) -> List[str]: rank_zero_info("***** Test results *****") _A : int = trainer.callback_metrics # Log and save results to file _A : List[Any] = os.path.join(pl_module.hparams.output_dir , "test_results.txt") with open(__lowerCamelCase , "w") as writer: for key in sorted(__lowerCamelCase): if key not in ["log", "progress_bar"]: rank_zero_info("{} = {}\n".format(__lowerCamelCase , str(metrics[key]))) writer.write("{} = {}\n".format(__lowerCamelCase , str(metrics[key]))) def _UpperCAmelCase (UpperCamelCase__ : List[str] , UpperCamelCase__ : int ): # To allow all pl args uncomment the following line # parser = pl.Trainer.add_argparse_args(parser) parser.add_argument( "--output_dir" , default=str(Path(UpperCamelCase__ ).parent / "test_run" / "model_checkpoints" ) , type=UpperCamelCase__ , help="The output directory where the model predictions and checkpoints will be written." , ) parser.add_argument( "--fp16" , action="store_true" , help="Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit" , ) parser.add_argument( "--fp16_opt_level" , type=UpperCamelCase__ , default="O2" , help=( "For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3']." "See details at https://nvidia.github.io/apex/amp.html" ) , ) parser.add_argument("--n_tpu_cores" , dest="tpu_cores" , type=UpperCamelCase__ ) parser.add_argument("--max_grad_norm" , dest="gradient_clip_val" , default=1.0 , type=UpperCamelCase__ , help="Max gradient norm" ) parser.add_argument("--do_train" , action="store_true" , help="Whether to run training." ) parser.add_argument("--do_predict" , action="store_true" , help="Whether to run predictions on the test set." ) parser.add_argument( "--gradient_accumulation_steps" , dest="accumulate_grad_batches" , type=UpperCamelCase__ , default=1 , help="Number of updates steps to accumulate before performing a backward/update pass." , ) parser.add_argument("--seed" , type=UpperCamelCase__ , default=42 , help="random seed for initialization" ) parser.add_argument( "--data_dir" , default=str(Path(UpperCamelCase__ ).parent / "test_run" / "dummy-train-data" ) , type=UpperCamelCase__ , help="The input data dir. Should contain the training files for the CoNLL-2003 NER task." , ) def _UpperCAmelCase (UpperCamelCase__ : BaseTransformer , UpperCamelCase__ : argparse.Namespace , UpperCamelCase__ : str=None , UpperCamelCase__ : Dict=True , UpperCamelCase__ : List[Any]=[] , UpperCamelCase__ : Any=None , UpperCamelCase__ : Optional[Any]=None , **UpperCamelCase__ : Union[str, Any] , ): pl.seed_everything(args.seed ) # init model _A : str = Path(model.hparams.output_dir ) odir.mkdir(exist_ok=UpperCamelCase__ ) # add custom checkpoints if checkpoint_callback is None: _A : Optional[Any] = pl.callbacks.ModelCheckpoint( filepath=args.output_dir , prefix="checkpoint" , monitor="val_loss" , mode="min" , save_top_k=1 ) if early_stopping_callback: extra_callbacks.append(UpperCamelCase__ ) if logging_callback is None: _A : List[Any] = LoggingCallback() _A : int = {} if args.fpaa: _A : int = 16 if args.gpus > 1: _A : str = "auto" _A : Optional[int] = "ddp" _A : List[str] = args.accumulate_grad_batches _A : str = None _A : Union[str, Any] = "auto" _A : List[str] = pl.Trainer.from_argparse_args( UpperCamelCase__ , weights_summary=UpperCamelCase__ , callbacks=[logging_callback] + extra_callbacks + [InitCallback()] + [checkpoint_callback] , logger=UpperCamelCase__ , val_check_interval=1 , num_sanity_val_steps=2 , **UpperCamelCase__ , ) if args.do_train: trainer.fit(UpperCamelCase__ ) else: print("RAG modeling tests with new set functions successfuly executed!" ) return trainer
11
'''simple docstring''' # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCamelCase__ = { '''configuration_vivit''': ['''VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''VivitConfig'''], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = ['''VivitImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ '''VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''VivitModel''', '''VivitPreTrainedModel''', '''VivitForVideoClassification''', ] if TYPE_CHECKING: from .configuration_vivit import VIVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, VivitConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_vivit import VivitImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vivit import ( VIVIT_PRETRAINED_MODEL_ARCHIVE_LIST, VivitForVideoClassification, VivitModel, VivitPreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
181
0
'''simple docstring''' import unittest from transformers import DebertaVaTokenizer, DebertaVaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin A : str = get_tests_dir('''fixtures/spiece.model''') @require_sentencepiece @require_tokenizers class __lowerCamelCase ( a_ , unittest.TestCase ): """simple docstring""" a = DebertaVaTokenizer a = DebertaVaTokenizerFast a = True a = True def A ( self : Dict): super().setUp() # We have a SentencePiece fixture for testing _A : Tuple = DebertaVaTokenizer(SCREAMING_SNAKE_CASE , unk_token='<unk>') tokenizer.save_pretrained(self.tmpdirname) def A ( self : str , SCREAMING_SNAKE_CASE : Optional[int]): _A : Optional[int] = 'this is a test' _A : int = 'this is a test' return input_text, output_text def A ( self : Union[str, Any]): _A : Union[str, Any] = '<pad>' _A : Dict = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(SCREAMING_SNAKE_CASE) , SCREAMING_SNAKE_CASE) self.assertEqual(self.get_tokenizer()._convert_id_to_token(SCREAMING_SNAKE_CASE) , SCREAMING_SNAKE_CASE) def A ( self : List[str]): _A : Optional[Any] = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , '<pad>') self.assertEqual(vocab_keys[1] , '<unk>') self.assertEqual(vocab_keys[-1] , '[PAD]') self.assertEqual(len(SCREAMING_SNAKE_CASE) , 30001) def A ( self : str): self.assertEqual(self.get_tokenizer().vocab_size , 30000) def A ( self : Dict): # fmt: off _A : List[str] = ' \tHeLLo!how \n Are yoU? ' _A : List[Any] = ['▁hello', '!', 'how', '▁are', '▁you', '?'] # fmt: on _A : Any = DebertaVaTokenizer(SCREAMING_SNAKE_CASE , do_lower_case=SCREAMING_SNAKE_CASE) _A : Dict = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE)) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : str = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE , do_lower_case=SCREAMING_SNAKE_CASE) _A : int = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE)) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) @unittest.skip('There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.') def A ( self : List[Any]): pass @unittest.skip('There is an inconsistency between slow and fast tokenizer due to a bug in the fast one.') def A ( self : List[Any]): pass def A ( self : str): # fmt: off _A : Union[str, Any] = 'I was born in 92000, and this is falsé.' _A : Tuple = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on _A : List[str] = DebertaVaTokenizer(SCREAMING_SNAKE_CASE , split_by_punct=SCREAMING_SNAKE_CASE) _A : Dict = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE)) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : List[str] = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE , split_by_punct=SCREAMING_SNAKE_CASE) _A : int = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE)) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) def A ( self : List[str]): # fmt: off _A : Union[str, Any] = 'I was born in 92000, and this is falsé.' _A : List[Any] = ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on _A : int = DebertaVaTokenizer(SCREAMING_SNAKE_CASE , do_lower_case=SCREAMING_SNAKE_CASE , split_by_punct=SCREAMING_SNAKE_CASE) _A : List[Any] = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE)) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : List[str] = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE , do_lower_case=SCREAMING_SNAKE_CASE , split_by_punct=SCREAMING_SNAKE_CASE) _A : Dict = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE)) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) def A ( self : Optional[Any]): # fmt: off _A : Any = 'I was born in 92000, and this is falsé.' _A : Dict = ['▁i', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.', ] # fmt: on _A : Union[str, Any] = DebertaVaTokenizer(SCREAMING_SNAKE_CASE , do_lower_case=SCREAMING_SNAKE_CASE , split_by_punct=SCREAMING_SNAKE_CASE) _A : str = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE)) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : List[str] = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE , do_lower_case=SCREAMING_SNAKE_CASE , split_by_punct=SCREAMING_SNAKE_CASE) _A : Union[str, Any] = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE)) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) def A ( self : List[str]): # fmt: off _A : Union[str, Any] = 'I was born in 92000, and this is falsé.' _A : Optional[Any] = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', '▁', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '▁', '.', ] # fmt: on _A : Tuple = DebertaVaTokenizer(SCREAMING_SNAKE_CASE , do_lower_case=SCREAMING_SNAKE_CASE , split_by_punct=SCREAMING_SNAKE_CASE) _A : List[str] = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE)) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : Tuple = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE , do_lower_case=SCREAMING_SNAKE_CASE , split_by_punct=SCREAMING_SNAKE_CASE) _A : List[Any] = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE)) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) def A ( self : Any): # fmt: off _A : Any = ' \tHeLLo!how \n Are yoU? ' _A : List[str] = ['▁', '<unk>', 'e', '<unk>', 'o', '!', 'how', '▁', '<unk>', 're', '▁yo', '<unk>', '?'] # fmt: on _A : str = DebertaVaTokenizer(SCREAMING_SNAKE_CASE , do_lower_case=SCREAMING_SNAKE_CASE , split_by_punct=SCREAMING_SNAKE_CASE) _A : Union[str, Any] = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE)) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : str = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE , do_lower_case=SCREAMING_SNAKE_CASE , split_by_punct=SCREAMING_SNAKE_CASE) _A : Union[str, Any] = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE)) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) def A ( self : Tuple): _A : Any = self.get_tokenizer() _A : Any = self.get_rust_tokenizer() _A : Optional[Any] = 'I was born in 92000, and this is falsé.' _A : Union[str, Any] = tokenizer.convert_ids_to_tokens(tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE)) _A : Tuple = rust_tokenizer.convert_ids_to_tokens(rust_tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE)) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : Optional[Any] = tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE) _A : int = rust_tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : Tuple = self.get_rust_tokenizer() _A : int = tokenizer.encode(SCREAMING_SNAKE_CASE) _A : str = rust_tokenizer.encode(SCREAMING_SNAKE_CASE) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) def A ( self : str): _A : str = 'This is a test' _A : str = [13, 1, 4398, 25, 21, 1289] _A : List[Any] = ['▁', 'T', 'his', '▁is', '▁a', '▁test'] _A : Optional[int] = ['▁', '<unk>', 'his', '▁is', '▁a', '▁test'] _A : str = DebertaVaTokenizer(SCREAMING_SNAKE_CASE , keep_accents=SCREAMING_SNAKE_CASE) _A : Any = DebertaVaTokenizerFast(SCREAMING_SNAKE_CASE , keep_accents=SCREAMING_SNAKE_CASE) _A : Any = tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : List[Any] = tokenizer.tokenize(SCREAMING_SNAKE_CASE) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : Any = tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : str = rust_tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : Optional[int] = rust_tokenizer.tokenize(SCREAMING_SNAKE_CASE) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : Union[str, Any] = rust_tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) # fmt: off _A : Union[str, Any] = 'I was born in 92000, and this is falsé.' _A : Optional[Any] = [13, 1, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9] _A : Tuple = ['▁', 'I', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', 'é', '.', ] _A : List[Any] = ['▁', '<unk>', '▁was', '▁born', '▁in', '▁9', '2000', ',', '▁and', '▁this', '▁is', '▁fal', 's', '<unk>', '.', ] # fmt: on _A : int = tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : Tuple = tokenizer.tokenize(SCREAMING_SNAKE_CASE) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : Optional[Any] = tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : int = rust_tokenizer.encode(SCREAMING_SNAKE_CASE , add_special_tokens=SCREAMING_SNAKE_CASE) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : Dict = rust_tokenizer.tokenize(SCREAMING_SNAKE_CASE) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) _A : Union[str, Any] = rust_tokenizer.convert_ids_to_tokens(SCREAMING_SNAKE_CASE) self.assertListEqual(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) def A ( self : str): _A : Any = DebertaVaTokenizer(SCREAMING_SNAKE_CASE) _A : Optional[int] = tokenizer.encode('sequence builders') _A : Tuple = tokenizer.encode('multi-sequence build') _A : Dict = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE) _A : Optional[Any] = tokenizer.build_inputs_with_special_tokens(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE) self.assertEqual([tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] , SCREAMING_SNAKE_CASE) self.assertEqual( [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [tokenizer.sep_token_id] , SCREAMING_SNAKE_CASE , ) @slow def A ( self : Optional[int]): # fmt: off _A : Optional[int] = {'input_ids': [[1, 39867, 36, 19390, 486, 27, 35052, 81436, 18, 60685, 1225, 7, 35052, 81436, 18, 9367, 16899, 18, 15937, 53, 594, 773, 18, 16287, 30465, 36, 15937, 6, 41139, 38, 36979, 60763, 191, 6, 34132, 99, 6, 50538, 390, 43230, 6, 34132, 2779, 20850, 14, 699, 1072, 1194, 36, 382, 10901, 53, 7, 699, 1072, 2084, 36, 20422, 630, 53, 19, 105, 3049, 1896, 1053, 16899, 1506, 11, 37978, 4243, 7, 1237, 31869, 200, 16566, 654, 6, 35052, 81436, 7, 55630, 13593, 4, 2], [1, 26, 15011, 13, 667, 8, 1053, 18, 23611, 1237, 72356, 12820, 34, 104134, 1209, 35, 13313, 6627, 21, 202, 347, 7, 164, 2399, 11, 46, 4485, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 5, 1232, 2864, 15785, 14951, 105, 5, 8581, 1250, 4, 2, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'token_type_ids': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=SCREAMING_SNAKE_CASE , model_name='microsoft/deberta-v2-xlarge' , revision='ad6e42c1532ddf3a15c39246b63f5559d558b670' , )
227
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A : List[str] = { '''configuration_electra''': ['''ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ElectraConfig''', '''ElectraOnnxConfig'''], '''tokenization_electra''': ['''ElectraTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : int = ['''ElectraTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : List[str] = [ '''ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ElectraForCausalLM''', '''ElectraForMaskedLM''', '''ElectraForMultipleChoice''', '''ElectraForPreTraining''', '''ElectraForQuestionAnswering''', '''ElectraForSequenceClassification''', '''ElectraForTokenClassification''', '''ElectraModel''', '''ElectraPreTrainedModel''', '''load_tf_weights_in_electra''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Dict = [ '''TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFElectraForMaskedLM''', '''TFElectraForMultipleChoice''', '''TFElectraForPreTraining''', '''TFElectraForQuestionAnswering''', '''TFElectraForSequenceClassification''', '''TFElectraForTokenClassification''', '''TFElectraModel''', '''TFElectraPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : List[Any] = [ '''FlaxElectraForCausalLM''', '''FlaxElectraForMaskedLM''', '''FlaxElectraForMultipleChoice''', '''FlaxElectraForPreTraining''', '''FlaxElectraForQuestionAnswering''', '''FlaxElectraForSequenceClassification''', '''FlaxElectraForTokenClassification''', '''FlaxElectraModel''', '''FlaxElectraPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_electra import ELECTRA_PRETRAINED_CONFIG_ARCHIVE_MAP, ElectraConfig, ElectraOnnxConfig from .tokenization_electra import ElectraTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_electra_fast import ElectraTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_electra import ( ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, ElectraForCausalLM, ElectraForMaskedLM, ElectraForMultipleChoice, ElectraForPreTraining, ElectraForQuestionAnswering, ElectraForSequenceClassification, ElectraForTokenClassification, ElectraModel, ElectraPreTrainedModel, load_tf_weights_in_electra, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_electra import ( TF_ELECTRA_PRETRAINED_MODEL_ARCHIVE_LIST, TFElectraForMaskedLM, TFElectraForMultipleChoice, TFElectraForPreTraining, TFElectraForQuestionAnswering, TFElectraForSequenceClassification, TFElectraForTokenClassification, TFElectraModel, TFElectraPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_electra import ( FlaxElectraForCausalLM, FlaxElectraForMaskedLM, FlaxElectraForMultipleChoice, FlaxElectraForPreTraining, FlaxElectraForQuestionAnswering, FlaxElectraForSequenceClassification, FlaxElectraForTokenClassification, FlaxElectraModel, FlaxElectraPreTrainedModel, ) else: import sys A : Tuple = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
227
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = { '''facebook/data2vec-vision-base-ft''': ( '''https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json''' ), } class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = "data2vec-vision" def __init__( self, lowerCAmelCase__=768, lowerCAmelCase__=12, lowerCAmelCase__=12, lowerCAmelCase__=3072, lowerCAmelCase__="gelu", lowerCAmelCase__=0.0, lowerCAmelCase__=0.0, lowerCAmelCase__=0.02, lowerCAmelCase__=1e-12, lowerCAmelCase__=224, lowerCAmelCase__=16, lowerCAmelCase__=3, lowerCAmelCase__=False, lowerCAmelCase__=False, lowerCAmelCase__=False, lowerCAmelCase__=False, lowerCAmelCase__=0.1, lowerCAmelCase__=0.1, lowerCAmelCase__=True, lowerCAmelCase__=[3, 5, 7, 11], lowerCAmelCase__=[1, 2, 3, 6], lowerCAmelCase__=True, lowerCAmelCase__=0.4, lowerCAmelCase__=256, lowerCAmelCase__=1, lowerCAmelCase__=False, lowerCAmelCase__=255, **lowerCAmelCase__, ) -> Optional[int]: super().__init__(**lowerCAmelCase__) snake_case_ = hidden_size snake_case_ = num_hidden_layers snake_case_ = num_attention_heads snake_case_ = intermediate_size snake_case_ = hidden_act snake_case_ = hidden_dropout_prob snake_case_ = attention_probs_dropout_prob snake_case_ = initializer_range snake_case_ = layer_norm_eps snake_case_ = image_size snake_case_ = patch_size snake_case_ = num_channels snake_case_ = use_mask_token snake_case_ = use_absolute_position_embeddings snake_case_ = use_relative_position_bias snake_case_ = use_shared_relative_position_bias snake_case_ = layer_scale_init_value snake_case_ = drop_path_rate snake_case_ = use_mean_pooling # decode head attributes (semantic segmentation) snake_case_ = out_indices snake_case_ = pool_scales # auxiliary head attributes (semantic segmentation) snake_case_ = use_auxiliary_head snake_case_ = auxiliary_loss_weight snake_case_ = auxiliary_channels snake_case_ = auxiliary_num_convs snake_case_ = auxiliary_concat_input snake_case_ = semantic_loss_ignore_index class UpperCamelCase ( lowerCAmelCase__ ): SCREAMING_SNAKE_CASE_ = version.parse("1.11" ) @property def a_ ( self) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ]) @property def a_ ( self) -> float: return 1e-4
69
"""simple docstring""" from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) # TODO Update this lowercase__ = { """facebook/esm-1b""": """https://huggingface.co/facebook/esm-1b/resolve/main/config.json""", # See all ESM models at https://huggingface.co/models?filter=esm } class __lowerCamelCase ( A__ ): '''simple docstring''' a_ : str = """esm""" def __init__( self : Union[str, Any] , a_ : int=None , a_ : List[str]=None , a_ : Optional[int]=None , a_ : Optional[int]=7_68 , a_ : List[Any]=12 , a_ : List[str]=12 , a_ : Optional[Any]=30_72 , a_ : Optional[Any]=0.1 , a_ : Tuple=0.1 , a_ : Union[str, Any]=10_26 , a_ : List[str]=0.02 , a_ : Optional[int]=1e-1_2 , a_ : int="absolute" , a_ : Union[str, Any]=True , a_ : int=None , a_ : int=False , a_ : Optional[Any]=False , a_ : Any=None , a_ : List[str]=None , **a_ : int , ): super().__init__(pad_token_id=a_ , mask_token_id=a_ , **a_ ) lowerCAmelCase_ : str = vocab_size lowerCAmelCase_ : List[str] = hidden_size lowerCAmelCase_ : str = num_hidden_layers lowerCAmelCase_ : str = num_attention_heads lowerCAmelCase_ : Tuple = intermediate_size lowerCAmelCase_ : List[str] = hidden_dropout_prob lowerCAmelCase_ : Dict = attention_probs_dropout_prob lowerCAmelCase_ : Dict = max_position_embeddings lowerCAmelCase_ : Optional[Any] = initializer_range lowerCAmelCase_ : Tuple = layer_norm_eps lowerCAmelCase_ : Dict = position_embedding_type lowerCAmelCase_ : str = use_cache lowerCAmelCase_ : str = emb_layer_norm_before lowerCAmelCase_ : Any = token_dropout lowerCAmelCase_ : str = is_folding_model if is_folding_model: if esmfold_config is None: logger.info("No esmfold_config supplied for folding model, using default values." ) lowerCAmelCase_ : int = EsmFoldConfig() elif isinstance(a_ , a_ ): lowerCAmelCase_ : int = EsmFoldConfig(**a_ ) lowerCAmelCase_ : Tuple = esmfold_config if vocab_list is None: logger.warning("No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!" ) lowerCAmelCase_ : Any = get_default_vocab_list() else: lowerCAmelCase_ : Optional[Any] = vocab_list else: lowerCAmelCase_ : List[str] = None lowerCAmelCase_ : Tuple = None if self.esmfold_config is not None and getattr(self.esmfold_config , "use_esm_attn_map" , a_ ): raise ValueError("The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!" ) def lowerCamelCase ( self : Dict ): lowerCAmelCase_ : List[str] = super().to_dict() if isinstance(self.esmfold_config , a_ ): lowerCAmelCase_ : int = self.esmfold_config.to_dict() return output @dataclass class __lowerCamelCase : '''simple docstring''' a_ : str = None a_ : bool = True a_ : bool = False a_ : bool = False a_ : bool = False a_ : float = 0 a_ : bool = True a_ : bool = False a_ : int = 128 a_ : "TrunkConfig" = None def lowerCamelCase ( self : str ): if self.trunk is None: lowerCAmelCase_ : List[Any] = TrunkConfig() elif isinstance(self.trunk , a_ ): lowerCAmelCase_ : str = TrunkConfig(**self.trunk ) def lowerCamelCase ( self : Any ): lowerCAmelCase_ : Any = asdict(self ) lowerCAmelCase_ : int = self.trunk.to_dict() return output @dataclass class __lowerCamelCase : '''simple docstring''' a_ : int = 48 a_ : int = 1024 a_ : int = 128 a_ : int = 32 a_ : int = 32 a_ : int = 32 a_ : float = 0 a_ : float = 0 a_ : bool = False a_ : int = 4 a_ : Optional[int] = 128 a_ : "StructureModuleConfig" = None def lowerCamelCase ( self : Optional[Any] ): if self.structure_module is None: lowerCAmelCase_ : Any = StructureModuleConfig() elif isinstance(self.structure_module , a_ ): lowerCAmelCase_ : Union[str, Any] = StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(f'''`max_recycles` should be positive, got {self.max_recycles}.''' ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( "`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got" f''' {self.sequence_state_dim} and {self.sequence_state_dim}.''' ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( "`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got" f''' {self.pairwise_state_dim} and {self.pairwise_state_dim}.''' ) lowerCAmelCase_ : List[str] = self.sequence_state_dim // self.sequence_head_width lowerCAmelCase_ : str = self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( "`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got" f''' {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}.''' ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( "`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got" f''' {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}.''' ) if self.pairwise_state_dim % 2 != 0: raise ValueError(f'''`pairwise_state_dim` should be even, got {self.pairwise_state_dim}.''' ) if self.dropout >= 0.4: raise ValueError(f'''`dropout` should not be greater than 0.4, got {self.dropout}.''' ) def lowerCamelCase ( self : Optional[int] ): lowerCAmelCase_ : Union[str, Any] = asdict(self ) lowerCAmelCase_ : str = self.structure_module.to_dict() return output @dataclass class __lowerCamelCase : '''simple docstring''' a_ : int = 384 a_ : int = 128 a_ : int = 16 a_ : int = 128 a_ : int = 12 a_ : int = 4 a_ : int = 8 a_ : float = 0.1 a_ : int = 8 a_ : int = 1 a_ : int = 2 a_ : int = 7 a_ : int = 10 a_ : float = 1E-8 a_ : float = 1E5 def lowerCamelCase ( self : Optional[int] ): return asdict(self ) def __lowerCamelCase ( ) -> Tuple: """simple docstring""" return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
241
0
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roberta import RobertaTokenizer lowerCAmelCase = logging.get_logger(__name__) lowerCAmelCase = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} lowerCAmelCase = { 'vocab_file': { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/vocab.json', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/vocab.json', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/vocab.json', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/vocab.json', 'roberta-base-openai-detector': 'https://huggingface.co/roberta-base-openai-detector/resolve/main/vocab.json', 'roberta-large-openai-detector': ( 'https://huggingface.co/roberta-large-openai-detector/resolve/main/vocab.json' ), }, 'merges_file': { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/merges.txt', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/merges.txt', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/merges.txt', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/merges.txt', 'roberta-base-openai-detector': 'https://huggingface.co/roberta-base-openai-detector/resolve/main/merges.txt', 'roberta-large-openai-detector': ( 'https://huggingface.co/roberta-large-openai-detector/resolve/main/merges.txt' ), }, 'tokenizer_file': { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/tokenizer.json', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/tokenizer.json', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/tokenizer.json', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/tokenizer.json', 'roberta-base-openai-detector': ( 'https://huggingface.co/roberta-base-openai-detector/resolve/main/tokenizer.json' ), 'roberta-large-openai-detector': ( 'https://huggingface.co/roberta-large-openai-detector/resolve/main/tokenizer.json' ), }, } lowerCAmelCase = { 'roberta-base': 512, 'roberta-large': 512, 'roberta-large-mnli': 512, 'distilroberta-base': 512, 'roberta-base-openai-detector': 512, 'roberta-large-openai-detector': 512, } class _a ( UpperCamelCase__ ): _lowercase : Union[str, Any] = VOCAB_FILES_NAMES _lowercase : Any = PRETRAINED_VOCAB_FILES_MAP _lowercase : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowercase : Optional[int] = ['''input_ids''', '''attention_mask'''] _lowercase : Dict = RobertaTokenizer def __init__( self: int , UpperCamelCase_: Any=None , UpperCamelCase_: Any=None , UpperCamelCase_: List[str]=None , UpperCamelCase_: Optional[int]="replace" , UpperCamelCase_: Any="<s>" , UpperCamelCase_: List[Any]="</s>" , UpperCamelCase_: Optional[Any]="</s>" , UpperCamelCase_: Dict="<s>" , UpperCamelCase_: List[Any]="<unk>" , UpperCamelCase_: Tuple="<pad>" , UpperCamelCase_: Any="<mask>" , UpperCamelCase_: Optional[Any]=False , UpperCamelCase_: int=True , **UpperCamelCase_: Optional[Any] , ) -> Optional[Any]: """simple docstring""" super().__init__( UpperCamelCase_ , UpperCamelCase_ , tokenizer_file=UpperCamelCase_ , errors=UpperCamelCase_ , bos_token=UpperCamelCase_ , eos_token=UpperCamelCase_ , sep_token=UpperCamelCase_ , cls_token=UpperCamelCase_ , unk_token=UpperCamelCase_ , pad_token=UpperCamelCase_ , mask_token=UpperCamelCase_ , add_prefix_space=UpperCamelCase_ , trim_offsets=UpperCamelCase_ , **UpperCamelCase_ , ) lowercase__ = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , UpperCamelCase_ ) != add_prefix_space: lowercase__ = getattr(UpperCamelCase_ , pre_tok_state.pop('''type''' ) ) lowercase__ = add_prefix_space lowercase__ = pre_tok_class(**UpperCamelCase_ ) lowercase__ = add_prefix_space lowercase__ = '''post_processor''' lowercase__ = getattr(self.backend_tokenizer , UpperCamelCase_ , UpperCamelCase_ ) if tokenizer_component_instance: lowercase__ = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: lowercase__ = tuple(state['''sep'''] ) if "cls" in state: lowercase__ = tuple(state['''cls'''] ) lowercase__ = False if state.get('''add_prefix_space''' , UpperCamelCase_ ) != add_prefix_space: lowercase__ = add_prefix_space lowercase__ = True if state.get('''trim_offsets''' , UpperCamelCase_ ) != trim_offsets: lowercase__ = trim_offsets lowercase__ = True if changes_to_apply: lowercase__ = getattr(UpperCamelCase_ , state.pop('''type''' ) ) lowercase__ = component_class(**UpperCamelCase_ ) setattr(self.backend_tokenizer , UpperCamelCase_ , UpperCamelCase_ ) @property def lowerCamelCase_ ( self: List[str] ) -> str: """simple docstring""" if self._mask_token is None: if self.verbose: logger.error('''Using mask_token, but it is not set yet.''' ) return None return str(self._mask_token ) @mask_token.setter def lowerCamelCase_ ( self: Dict , UpperCamelCase_: Tuple ) -> Tuple: """simple docstring""" lowercase__ = AddedToken(UpperCamelCase_ , lstrip=UpperCamelCase_ , rstrip=UpperCamelCase_ ) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) else value lowercase__ = value def lowerCamelCase_ ( self: str , *UpperCamelCase_: Any , **UpperCamelCase_: List[Any] ) -> BatchEncoding: """simple docstring""" lowercase__ = kwargs.get('''is_split_into_words''' , UpperCamelCase_ ) assert self.add_prefix_space or not is_split_into_words, ( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*UpperCamelCase_ , **UpperCamelCase_ ) def lowerCamelCase_ ( self: Any , *UpperCamelCase_: List[Any] , **UpperCamelCase_: int ) -> BatchEncoding: """simple docstring""" lowercase__ = kwargs.get('''is_split_into_words''' , UpperCamelCase_ ) assert self.add_prefix_space or not is_split_into_words, ( f'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs." ) return super()._encode_plus(*UpperCamelCase_ , **UpperCamelCase_ ) def lowerCamelCase_ ( self: Optional[int] , UpperCamelCase_: str , UpperCamelCase_: Optional[str] = None ) -> Tuple[str]: """simple docstring""" lowercase__ = self._tokenizer.model.save(UpperCamelCase_ , name=UpperCamelCase_ ) return tuple(UpperCamelCase_ ) def lowerCamelCase_ ( self: Union[str, Any] , UpperCamelCase_: Optional[Any] , UpperCamelCase_: str=None ) -> List[str]: """simple docstring""" lowercase__ = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def lowerCamelCase_ ( self: Optional[Any] , UpperCamelCase_: List[int] , UpperCamelCase_: Optional[List[int]] = None ) -> List[int]: """simple docstring""" lowercase__ = [self.sep_token_id] lowercase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
93
from collections.abc import Sequence from queue import Queue class _a : def __init__( self: Tuple , UpperCamelCase_: Optional[int] , UpperCamelCase_: int , UpperCamelCase_: Optional[Any] , UpperCamelCase_: Union[str, Any]=None , UpperCamelCase_: Dict=None ) -> Tuple: """simple docstring""" lowercase__ = start lowercase__ = end lowercase__ = val lowercase__ = (start + end) // 2 lowercase__ = left lowercase__ = right def __repr__( self: Optional[int] ) -> Optional[Any]: """simple docstring""" return f'SegmentTreeNode(start={self.start}, end={self.end}, val={self.val})' class _a : def __init__( self: Any , UpperCamelCase_: Sequence , UpperCamelCase_: Any ) -> List[str]: """simple docstring""" lowercase__ = collection lowercase__ = function if self.collection: lowercase__ = self._build_tree(0 , len(UpperCamelCase_ ) - 1 ) def lowerCamelCase_ ( self: int , UpperCamelCase_: Union[str, Any] , UpperCamelCase_: Union[str, Any] ) -> Union[str, Any]: """simple docstring""" self._update_tree(self.root , UpperCamelCase_ , UpperCamelCase_ ) def lowerCamelCase_ ( self: str , UpperCamelCase_: int , UpperCamelCase_: List[str] ) -> Optional[Any]: """simple docstring""" return self._query_range(self.root , UpperCamelCase_ , UpperCamelCase_ ) def lowerCamelCase_ ( self: List[str] , UpperCamelCase_: Tuple , UpperCamelCase_: Dict ) -> str: """simple docstring""" if start == end: return SegmentTreeNode(UpperCamelCase_ , UpperCamelCase_ , self.collection[start] ) lowercase__ = (start + end) // 2 lowercase__ = self._build_tree(UpperCamelCase_ , UpperCamelCase_ ) lowercase__ = self._build_tree(mid + 1 , UpperCamelCase_ ) return SegmentTreeNode(UpperCamelCase_ , UpperCamelCase_ , self.fn(left.val , right.val ) , UpperCamelCase_ , UpperCamelCase_ ) def lowerCamelCase_ ( self: List[Any] , UpperCamelCase_: Dict , UpperCamelCase_: Tuple , UpperCamelCase_: Union[str, Any] ) -> Dict: """simple docstring""" if node.start == i and node.end == i: lowercase__ = val return if i <= node.mid: self._update_tree(node.left , UpperCamelCase_ , UpperCamelCase_ ) else: self._update_tree(node.right , UpperCamelCase_ , UpperCamelCase_ ) lowercase__ = self.fn(node.left.val , node.right.val ) def lowerCamelCase_ ( self: List[Any] , UpperCamelCase_: str , UpperCamelCase_: Dict , UpperCamelCase_: Dict ) -> List[Any]: """simple docstring""" if node.start == i and node.end == j: return node.val if i <= node.mid: if j <= node.mid: # range in left child tree return self._query_range(node.left , UpperCamelCase_ , UpperCamelCase_ ) else: # range in left child tree and right child tree return self.fn( self._query_range(node.left , UpperCamelCase_ , node.mid ) , self._query_range(node.right , node.mid + 1 , UpperCamelCase_ ) , ) else: # range in right child tree return self._query_range(node.right , UpperCamelCase_ , UpperCamelCase_ ) def lowerCamelCase_ ( self: Optional[int] ) -> str: """simple docstring""" if self.root is not None: lowercase__ = Queue() queue.put(self.root ) while not queue.empty(): lowercase__ = queue.get() yield node if node.left is not None: queue.put(node.left ) if node.right is not None: queue.put(node.right ) if __name__ == "__main__": import operator for fn in [operator.add, max, min]: print('*' * 50) lowerCAmelCase = SegmentTree([2, 1, 5, 3, 4], fn) for node in arr.traverse(): print(node) print() arr.update(1, 5) for node in arr.traverse(): print(node) print() print(arr.query_range(3, 4)) # 7 print(arr.query_range(2, 2)) # 5 print(arr.query_range(1, 3)) # 13 print()
93
1
'''simple docstring''' import argparse import requests import torch from PIL import Image from torchvision.transforms import Compose, Normalize, Resize, ToTensor from transformers import SwinaSRConfig, SwinaSRForImageSuperResolution, SwinaSRImageProcessor def _SCREAMING_SNAKE_CASE (A ) -> Tuple: """simple docstring""" lowercase__ = SwinaSRConfig() if "Swin2SR_ClassicalSR_X4_64" in checkpoint_url: lowercase__ = 4 elif "Swin2SR_CompressedSR_X4_48" in checkpoint_url: lowercase__ = 4 lowercase__ = 48 lowercase__ = '''pixelshuffle_aux''' elif "Swin2SR_Lightweight_X2_64" in checkpoint_url: lowercase__ = [6, 6, 6, 6] lowercase__ = 60 lowercase__ = [6, 6, 6, 6] lowercase__ = '''pixelshuffledirect''' elif "Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR" in checkpoint_url: lowercase__ = 4 lowercase__ = '''nearest+conv''' elif "Swin2SR_Jpeg_dynamic" in checkpoint_url: lowercase__ = 1 lowercase__ = 1 lowercase__ = 126 lowercase__ = 7 lowercase__ = 255.0 lowercase__ = '''''' return config def _SCREAMING_SNAKE_CASE (A , A ) -> Dict: """simple docstring""" if "patch_embed.proj" in name and "layers" not in name: lowercase__ = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: lowercase__ = name.replace('''patch_embed.norm''' , '''embeddings.patch_embeddings.layernorm''' ) if "layers" in name: lowercase__ = name.replace('''layers''' , '''encoder.stages''' ) if "residual_group.blocks" in name: lowercase__ = name.replace('''residual_group.blocks''' , '''layers''' ) if "attn.proj" in name: lowercase__ = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: lowercase__ = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: lowercase__ = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: lowercase__ = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: lowercase__ = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: lowercase__ = name.replace('''mlp.fc2''' , '''output.dense''' ) if "q_bias" in name: lowercase__ = name.replace('''q_bias''' , '''query.bias''' ) if "k_bias" in name: lowercase__ = name.replace('''k_bias''' , '''key.bias''' ) if "v_bias" in name: lowercase__ = name.replace('''v_bias''' , '''value.bias''' ) if "cpb_mlp" in name: lowercase__ = name.replace('''cpb_mlp''' , '''continuous_position_bias_mlp''' ) if "patch_embed.proj" in name: lowercase__ = name.replace('''patch_embed.proj''' , '''patch_embed.projection''' ) if name == "norm.weight": lowercase__ = '''layernorm.weight''' if name == "norm.bias": lowercase__ = '''layernorm.bias''' if "conv_first" in name: lowercase__ = name.replace('''conv_first''' , '''first_convolution''' ) if ( "upsample" in name or "conv_before_upsample" in name or "conv_bicubic" in name or "conv_up" in name or "conv_hr" in name or "conv_last" in name or "aux" in name ): # heads if "conv_last" in name: lowercase__ = name.replace('''conv_last''' , '''final_convolution''' ) if config.upsampler in ["pixelshuffle", "pixelshuffle_aux", "nearest+conv"]: if "conv_before_upsample.0" in name: lowercase__ = name.replace('''conv_before_upsample.0''' , '''conv_before_upsample''' ) if "upsample.0" in name: lowercase__ = name.replace('''upsample.0''' , '''upsample.convolution_0''' ) if "upsample.2" in name: lowercase__ = name.replace('''upsample.2''' , '''upsample.convolution_1''' ) lowercase__ = '''upsample.''' + name elif config.upsampler == "pixelshuffledirect": lowercase__ = name.replace('''upsample.0.weight''' , '''upsample.conv.weight''' ) lowercase__ = name.replace('''upsample.0.bias''' , '''upsample.conv.bias''' ) else: pass else: lowercase__ = '''swin2sr.''' + name return name def _SCREAMING_SNAKE_CASE (A , A ) -> Dict: """simple docstring""" for key in orig_state_dict.copy().keys(): lowercase__ = orig_state_dict.pop(A ) if "qkv" in key: lowercase__ = key.split('''.''' ) lowercase__ = int(key_split[1] ) lowercase__ = int(key_split[4] ) lowercase__ = config.embed_dim if "weight" in key: lowercase__ = val[:dim, :] lowercase__ = val[dim : dim * 2, :] lowercase__ = val[-dim:, :] else: lowercase__ = val[:dim] lowercase__ = val[dim : dim * 2] lowercase__ = val[-dim:] pass else: lowercase__ = val return orig_state_dict def _SCREAMING_SNAKE_CASE (A , A , A ) -> Tuple: """simple docstring""" lowercase__ = get_config(A ) lowercase__ = SwinaSRForImageSuperResolution(A ) model.eval() lowercase__ = torch.hub.load_state_dict_from_url(A , map_location='''cpu''' ) lowercase__ = convert_state_dict(A , A ) lowercase__ ,lowercase__ = model.load_state_dict(A , strict=A ) if len(A ) > 0: raise ValueError('''Missing keys when converting: {}'''.format(A ) ) for key in unexpected_keys: if not ("relative_position_index" in key or "relative_coords_table" in key or "self_mask" in key): raise ValueError(f"Unexpected key {key} in state_dict" ) # verify values lowercase__ = '''https://github.com/mv-lab/swin2sr/blob/main/testsets/real-inputs/shanghai.jpg?raw=true''' lowercase__ = Image.open(requests.get(A , stream=A ).raw ).convert('''RGB''' ) lowercase__ = SwinaSRImageProcessor() # pixel_values = processor(image, return_tensors="pt").pixel_values lowercase__ = 126 if '''Jpeg''' in checkpoint_url else 256 lowercase__ = Compose( [ Resize((image_size, image_size) ), ToTensor(), Normalize(mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ), ] ) lowercase__ = transforms(A ).unsqueeze(0 ) if config.num_channels == 1: lowercase__ = pixel_values[:, 0, :, :].unsqueeze(1 ) lowercase__ = model(A ) # assert values if "Swin2SR_ClassicalSR_X2_64" in checkpoint_url: lowercase__ = torch.Size([1, 3, 512, 512] ) lowercase__ = torch.tensor( [[-0.7_087, -0.7_138, -0.6_721], [-0.8_340, -0.8_095, -0.7_298], [-0.9_149, -0.8_414, -0.7_940]] ) elif "Swin2SR_ClassicalSR_X4_64" in checkpoint_url: lowercase__ = torch.Size([1, 3, 1_024, 1_024] ) lowercase__ = torch.tensor( [[-0.7_775, -0.8_105, -0.8_933], [-0.7_764, -0.8_356, -0.9_225], [-0.7_976, -0.8_686, -0.9_579]] ) elif "Swin2SR_CompressedSR_X4_48" in checkpoint_url: # TODO values didn't match exactly here lowercase__ = torch.Size([1, 3, 1_024, 1_024] ) lowercase__ = torch.tensor( [[-0.8_035, -0.7_504, -0.7_491], [-0.8_538, -0.8_124, -0.7_782], [-0.8_804, -0.8_651, -0.8_493]] ) elif "Swin2SR_Lightweight_X2_64" in checkpoint_url: lowercase__ = torch.Size([1, 3, 512, 512] ) lowercase__ = torch.tensor( [[-0.7_669, -0.8_662, -0.8_767], [-0.8_810, -0.9_962, -0.9_820], [-0.9_340, -1.0_322, -1.1_149]] ) elif "Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR" in checkpoint_url: lowercase__ = torch.Size([1, 3, 1_024, 1_024] ) lowercase__ = torch.tensor( [[-0.5_238, -0.5_557, -0.6_321], [-0.6_016, -0.5_903, -0.6_391], [-0.6_244, -0.6_334, -0.6_889]] ) assert ( outputs.reconstruction.shape == expected_shape ), f"Shape of reconstruction should be {expected_shape}, but is {outputs.reconstruction.shape}" assert torch.allclose(outputs.reconstruction[0, 0, :3, :3] , A , atol=1E-3 ) print('''Looks ok!''' ) lowercase__ = { '''https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X2_64.pth''': ( '''swin2SR-classical-sr-x2-64''' ), '''https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X4_64.pth''': ( '''swin2SR-classical-sr-x4-64''' ), '''https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_CompressedSR_X4_48.pth''': ( '''swin2SR-compressed-sr-x4-48''' ), '''https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_Lightweight_X2_64.pth''': ( '''swin2SR-lightweight-x2-64''' ), '''https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_RealworldSR_X4_64_BSRGAN_PSNR.pth''': ( '''swin2SR-realworld-sr-x4-64-bsrgan-psnr''' ), } lowercase__ = url_to_name[checkpoint_url] if pytorch_dump_folder_path is not None: print(f"Saving model {model_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(A ) print(f"Saving image processor to {pytorch_dump_folder_path}" ) processor.save_pretrained(A ) if push_to_hub: model.push_to_hub(f"caidas/{model_name}" ) processor.push_to_hub(f"caidas/{model_name}" ) if __name__ == "__main__": lowerCamelCase : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://github.com/mv-lab/swin2sr/releases/download/v0.0.1/Swin2SR_ClassicalSR_X2_64.pth', type=str, help='URL of the original Swin2SR checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument('--push_to_hub', action='store_true', help='Whether to push the converted model to the hub.') lowerCamelCase : Dict = parser.parse_args() convert_swinasr_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
2
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase : str = logging.get_logger(__name__) lowerCamelCase : int = { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json', } class __lowerCAmelCase (lowercase_ ): '''simple docstring''' lowerCAmelCase__ : Union[str, Any] = """mvp""" lowerCAmelCase__ : Optional[Any] = ["""past_key_values"""] lowerCAmelCase__ : List[str] = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__(self : Any , UpperCamelCase : Optional[int]=50267 , UpperCamelCase : Tuple=1024 , UpperCamelCase : int=12 , UpperCamelCase : Tuple=4096 , UpperCamelCase : Dict=16 , UpperCamelCase : int=12 , UpperCamelCase : Optional[int]=4096 , UpperCamelCase : Optional[int]=16 , UpperCamelCase : Tuple=0.0 , UpperCamelCase : Tuple=0.0 , UpperCamelCase : List[Any]="gelu" , UpperCamelCase : Union[str, Any]=1024 , UpperCamelCase : Optional[Any]=0.1 , UpperCamelCase : str=0.0 , UpperCamelCase : str=0.0 , UpperCamelCase : Optional[Any]=0.02 , UpperCamelCase : List[str]=0.0 , UpperCamelCase : List[str]=False , UpperCamelCase : Optional[int]=True , UpperCamelCase : Any=1 , UpperCamelCase : int=0 , UpperCamelCase : int=2 , UpperCamelCase : Any=True , UpperCamelCase : Optional[Any]=2 , UpperCamelCase : Optional[Any]=2 , UpperCamelCase : Tuple=False , UpperCamelCase : int=100 , UpperCamelCase : Optional[Any]=800 , **UpperCamelCase : str , ): '''simple docstring''' lowercase__ = vocab_size lowercase__ = max_position_embeddings lowercase__ = d_model lowercase__ = encoder_ffn_dim lowercase__ = encoder_layers lowercase__ = encoder_attention_heads lowercase__ = decoder_ffn_dim lowercase__ = decoder_layers lowercase__ = decoder_attention_heads lowercase__ = dropout lowercase__ = attention_dropout lowercase__ = activation_dropout lowercase__ = activation_function lowercase__ = init_std lowercase__ = encoder_layerdrop lowercase__ = decoder_layerdrop lowercase__ = classifier_dropout lowercase__ = use_cache lowercase__ = encoder_layers lowercase__ = scale_embedding # scale factor will be sqrt(d_model) if True lowercase__ = use_prompt lowercase__ = prompt_length lowercase__ = prompt_mid_dim super().__init__( pad_token_id=UpperCamelCase , bos_token_id=UpperCamelCase , eos_token_id=UpperCamelCase , is_encoder_decoder=UpperCamelCase , decoder_start_token_id=UpperCamelCase , forced_eos_token_id=UpperCamelCase , **UpperCamelCase , ) if self.forced_bos_token_id is None and kwargs.get('''force_bos_token_to_be_generated''' , UpperCamelCase ): lowercase__ = self.bos_token_id warnings.warn( f"Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. " '''The config can simply be saved and uploaded again to be fixed.''' )
2
1
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPanoramaPipeline, UNetaDConditionModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() @skip_mps class UpperCamelCase_ (_UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): __magic_name__ = StableDiffusionPanoramaPipeline __magic_name__ = TEXT_TO_IMAGE_PARAMS __magic_name__ = TEXT_TO_IMAGE_BATCH_PARAMS __magic_name__ = TEXT_TO_IMAGE_IMAGE_PARAMS __magic_name__ = TEXT_TO_IMAGE_IMAGE_PARAMS def _SCREAMING_SNAKE_CASE ( self : List[str] ) -> Optional[Any]: torch.manual_seed(0 ) UpperCAmelCase_ : Optional[int] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=1 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) UpperCAmelCase_ : List[Any] = DDIMScheduler() torch.manual_seed(0 ) UpperCAmelCase_ : Tuple = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) torch.manual_seed(0 ) UpperCAmelCase_ : Tuple = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) UpperCAmelCase_ : List[str] = CLIPTextModel(_UpperCAmelCase ) UpperCAmelCase_ : int = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) UpperCAmelCase_ : int = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def _SCREAMING_SNAKE_CASE ( self : Tuple , lowerCAmelCase_ : str , lowerCAmelCase_ : str=0 ) -> List[Any]: UpperCAmelCase_ : int = torch.manual_seed(_UpperCAmelCase ) UpperCAmelCase_ : List[Any] = { '''prompt''': '''a photo of the dolomites''', '''generator''': generator, # Setting height and width to None to prevent OOMs on CPU. '''height''': None, '''width''': None, '''num_inference_steps''': 1, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> List[Any]: UpperCAmelCase_ : Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ : List[str] = self.get_dummy_components() UpperCAmelCase_ : Union[str, Any] = StableDiffusionPanoramaPipeline(**_UpperCAmelCase ) UpperCAmelCase_ : int = sd_pipe.to(_UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=_UpperCAmelCase ) UpperCAmelCase_ : str = self.get_dummy_inputs(_UpperCAmelCase ) UpperCAmelCase_ : Any = sd_pipe(**_UpperCAmelCase ).images UpperCAmelCase_ : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase_ : List[str] = np.array([0.6_1_8_6, 0.5_3_7_4, 0.4_9_1_5, 0.4_1_3_5, 0.4_1_1_4, 0.4_5_6_3, 0.5_1_2_8, 0.4_9_7_7, 0.4_7_5_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self : int ) -> int: super().test_inference_batch_consistent(batch_sizes=[1, 2] ) def _SCREAMING_SNAKE_CASE ( self : Any ) -> int: super().test_inference_batch_single_identical(batch_size=2 , expected_max_diff=3.25e-3 ) def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> str: UpperCAmelCase_ : Optional[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ : Union[str, Any] = self.get_dummy_components() UpperCAmelCase_ : str = StableDiffusionPanoramaPipeline(**_UpperCAmelCase ) UpperCAmelCase_ : str = sd_pipe.to(_UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=_UpperCAmelCase ) UpperCAmelCase_ : str = self.get_dummy_inputs(_UpperCAmelCase ) UpperCAmelCase_ : Union[str, Any] = '''french fries''' UpperCAmelCase_ : Union[str, Any] = sd_pipe(**_UpperCAmelCase , negative_prompt=_UpperCAmelCase ) UpperCAmelCase_ : Optional[Any] = output.images UpperCAmelCase_ : str = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase_ : Optional[int] = np.array([0.6_1_8_7, 0.5_3_7_5, 0.4_9_1_5, 0.4_1_3_6, 0.4_1_1_4, 0.4_5_6_3, 0.5_1_2_8, 0.4_9_7_6, 0.4_7_5_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Optional[Any]: UpperCAmelCase_ : Optional[Any] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ : Union[str, Any] = self.get_dummy_components() UpperCAmelCase_ : Optional[Any] = StableDiffusionPanoramaPipeline(**_UpperCAmelCase ) UpperCAmelCase_ : str = sd_pipe.to(_UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=_UpperCAmelCase ) UpperCAmelCase_ : Optional[int] = self.get_dummy_inputs(_UpperCAmelCase ) UpperCAmelCase_ : Union[str, Any] = sd_pipe(**_UpperCAmelCase , view_batch_size=2 ) UpperCAmelCase_ : List[str] = output.images UpperCAmelCase_ : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase_ : List[Any] = np.array([0.6_1_8_7, 0.5_3_7_5, 0.4_9_1_5, 0.4_1_3_6, 0.4_1_1_4, 0.4_5_6_3, 0.5_1_2_8, 0.4_9_7_6, 0.4_7_5_7] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self : Any ) -> Any: UpperCAmelCase_ : Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ : int = self.get_dummy_components() UpperCAmelCase_ : List[str] = EulerAncestralDiscreteScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="scaled_linear" ) UpperCAmelCase_ : Any = StableDiffusionPanoramaPipeline(**_UpperCAmelCase ) UpperCAmelCase_ : Any = sd_pipe.to(_UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=_UpperCAmelCase ) UpperCAmelCase_ : int = self.get_dummy_inputs(_UpperCAmelCase ) UpperCAmelCase_ : Dict = sd_pipe(**_UpperCAmelCase ).images UpperCAmelCase_ : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase_ : List[Any] = np.array([0.4_0_2_4, 0.6_5_1_0, 0.4_9_0_1, 0.5_3_7_8, 0.5_8_1_3, 0.5_6_2_2, 0.4_7_9_5, 0.4_4_6_7, 0.4_9_5_2] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self : str ) -> str: UpperCAmelCase_ : int = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ : List[Any] = self.get_dummy_components() UpperCAmelCase_ : Any = PNDMScheduler( beta_start=0.0_0_0_8_5 , beta_end=0.0_1_2 , beta_schedule="scaled_linear" , skip_prk_steps=_UpperCAmelCase ) UpperCAmelCase_ : Dict = StableDiffusionPanoramaPipeline(**_UpperCAmelCase ) UpperCAmelCase_ : int = sd_pipe.to(_UpperCAmelCase ) sd_pipe.set_progress_bar_config(disable=_UpperCAmelCase ) UpperCAmelCase_ : Optional[int] = self.get_dummy_inputs(_UpperCAmelCase ) UpperCAmelCase_ : Dict = sd_pipe(**_UpperCAmelCase ).images UpperCAmelCase_ : str = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase_ : List[Any] = np.array([0.6_3_9_1, 0.6_2_9_1, 0.4_8_6_1, 0.5_1_3_4, 0.5_5_5_2, 0.4_5_7_8, 0.5_0_3_2, 0.5_0_2_3, 0.4_5_3_9] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch_gpu class UpperCamelCase_ (unittest.TestCase ): def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Optional[Any]: super().tearDown() gc.collect() torch.cuda.empty_cache() def _SCREAMING_SNAKE_CASE ( self : str , lowerCAmelCase_ : Tuple=0 ) -> str: UpperCAmelCase_ : Union[str, Any] = torch.manual_seed(_UpperCAmelCase ) UpperCAmelCase_ : int = { '''prompt''': '''a photo of the dolomites''', '''generator''': generator, '''num_inference_steps''': 3, '''guidance_scale''': 7.5, '''output_type''': '''numpy''', } return inputs def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> Any: UpperCAmelCase_ : Any = '''stabilityai/stable-diffusion-2-base''' UpperCAmelCase_ : str = DDIMScheduler.from_pretrained(_UpperCAmelCase , subfolder="scheduler" ) UpperCAmelCase_ : Dict = StableDiffusionPanoramaPipeline.from_pretrained(_UpperCAmelCase , scheduler=_UpperCAmelCase , safety_checker=_UpperCAmelCase ) pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) pipe.enable_attention_slicing() UpperCAmelCase_ : Tuple = self.get_inputs() UpperCAmelCase_ : Optional[Any] = pipe(**_UpperCAmelCase ).images UpperCAmelCase_ : Optional[Any] = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 2_048, 3) UpperCAmelCase_ : List[Any] = np.array( [ 0.3_6_9_6_8_3_9_2, 0.2_7_0_2_5_3_7_2, 0.3_2_4_4_6_7_6_6, 0.2_8_3_7_9_3_8_7, 0.3_6_3_6_3_2_7_4, 0.3_0_7_3_3_3_4_7, 0.2_7_1_0_0_0_2_7, 0.2_7_0_5_4_1_2_5, 0.2_5_5_3_6_0_9_6, ] ) assert np.abs(expected_slice - image_slice ).max() < 1e-2 def _SCREAMING_SNAKE_CASE ( self : int ) -> int: UpperCAmelCase_ : int = StableDiffusionPanoramaPipeline.from_pretrained( "stabilityai/stable-diffusion-2-base" , safety_checker=_UpperCAmelCase ) UpperCAmelCase_ : Tuple = LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) pipe.enable_attention_slicing() UpperCAmelCase_ : List[str] = self.get_inputs() UpperCAmelCase_ : Dict = pipe(**_UpperCAmelCase ).images UpperCAmelCase_ : Tuple = image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 512, 2_048, 3) UpperCAmelCase_ : List[Any] = np.array( [ [ 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, ] ] ) assert np.abs(expected_slice - image_slice ).max() < 1e-3 def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> Union[str, Any]: UpperCAmelCase_ : int = 0 def callback_fn(lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : int , lowerCAmelCase_ : int ) -> None: UpperCAmelCase_ : List[str] = True nonlocal number_of_steps number_of_steps += 1 if step == 1: UpperCAmelCase_ : Dict = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 256) UpperCAmelCase_ : Any = latents[0, -3:, -3:, -1] UpperCAmelCase_ : List[Any] = np.array( [ 0.1_8_6_8_1_8_6_9, 0.3_3_9_0_7_8_1_6, 0.5_3_6_1_2_7_6, 0.1_4_4_3_2_8_6_5, -0.0_2_8_5_6_6_1_1, -0.7_3_9_4_1_1_2_3, 0.2_3_3_9_7_9_8_7, 0.4_7_3_2_2_6_8_2, -0.3_7_8_2_3_1_6_4, ] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 elif step == 2: UpperCAmelCase_ : Tuple = latents.detach().cpu().numpy() assert latents.shape == (1, 4, 64, 256) UpperCAmelCase_ : Optional[Any] = latents[0, -3:, -3:, -1] UpperCAmelCase_ : Any = np.array( [ 0.1_8_5_3_9_6_4_5, 0.3_3_9_8_7_2_4_8, 0.5_3_7_8_5_5_9, 0.1_4_4_3_7_1_4_2, -0.0_2_4_5_5_2_6_1, -0.7_3_3_8_3_1_7, 0.2_3_9_9_0_7_5_5, 0.4_7_3_5_6_2_7_2, -0.3_7_8_6_5_0_5, ] ) assert np.abs(latents_slice.flatten() - expected_slice ).max() < 5e-2 UpperCAmelCase_ : int = False UpperCAmelCase_ : str = '''stabilityai/stable-diffusion-2-base''' UpperCAmelCase_ : Union[str, Any] = DDIMScheduler.from_pretrained(_UpperCAmelCase , subfolder="scheduler" ) UpperCAmelCase_ : Tuple = StableDiffusionPanoramaPipeline.from_pretrained(_UpperCAmelCase , scheduler=_UpperCAmelCase , safety_checker=_UpperCAmelCase ) UpperCAmelCase_ : Optional[Any] = pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) pipe.enable_attention_slicing() UpperCAmelCase_ : Tuple = self.get_inputs() pipe(**_UpperCAmelCase , callback=_UpperCAmelCase , callback_steps=1 ) assert callback_fn.has_been_called assert number_of_steps == 3 def _SCREAMING_SNAKE_CASE ( self : Dict ) -> List[Any]: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() UpperCAmelCase_ : List[Any] = '''stabilityai/stable-diffusion-2-base''' UpperCAmelCase_ : Any = DDIMScheduler.from_pretrained(_UpperCAmelCase , subfolder="scheduler" ) UpperCAmelCase_ : int = StableDiffusionPanoramaPipeline.from_pretrained(_UpperCAmelCase , scheduler=_UpperCAmelCase , safety_checker=_UpperCAmelCase ) UpperCAmelCase_ : List[Any] = pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() UpperCAmelCase_ : Any = self.get_inputs() UpperCAmelCase_ : List[str] = pipe(**_UpperCAmelCase ) UpperCAmelCase_ : Optional[int] = torch.cuda.max_memory_allocated() # make sure that less than 5.2 GB is allocated assert mem_bytes < 5.5 * 10**9
352
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import TensorType, logging if TYPE_CHECKING: from ...onnx.config import PatchingSpec from ...tokenization_utils_base import PreTrainedTokenizerBase lowerCamelCase_ = logging.get_logger(__name__) lowerCamelCase_ = { '''allenai/longformer-base-4096''': '''https://huggingface.co/allenai/longformer-base-4096/resolve/main/config.json''', '''allenai/longformer-large-4096''': '''https://huggingface.co/allenai/longformer-large-4096/resolve/main/config.json''', '''allenai/longformer-large-4096-finetuned-triviaqa''': ( '''https://huggingface.co/allenai/longformer-large-4096-finetuned-triviaqa/resolve/main/config.json''' ), '''allenai/longformer-base-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-base-4096-extra.pos.embd.only/resolve/main/config.json''' ), '''allenai/longformer-large-4096-extra.pos.embd.only''': ( '''https://huggingface.co/allenai/longformer-large-4096-extra.pos.embd.only/resolve/main/config.json''' ), } class UpperCamelCase_ (__A ): __magic_name__ = '''longformer''' def __init__( self : List[str] , lowerCAmelCase_ : Union[List[int], int] = 512 , lowerCAmelCase_ : int = 2 , lowerCAmelCase_ : int = 1 , lowerCAmelCase_ : int = 0 , lowerCAmelCase_ : int = 2 , lowerCAmelCase_ : int = 30_522 , lowerCAmelCase_ : int = 768 , lowerCAmelCase_ : int = 12 , lowerCAmelCase_ : int = 12 , lowerCAmelCase_ : int = 3_072 , lowerCAmelCase_ : str = "gelu" , lowerCAmelCase_ : float = 0.1 , lowerCAmelCase_ : float = 0.1 , lowerCAmelCase_ : int = 512 , lowerCAmelCase_ : int = 2 , lowerCAmelCase_ : float = 0.0_2 , lowerCAmelCase_ : float = 1e-12 , lowerCAmelCase_ : bool = False , **lowerCAmelCase_ : Optional[int] , ) -> Dict: super().__init__(pad_token_id=lowerCAmelCase_ , **lowerCAmelCase_ ) UpperCAmelCase_ : List[Any] = attention_window UpperCAmelCase_ : Dict = sep_token_id UpperCAmelCase_ : Any = bos_token_id UpperCAmelCase_ : Dict = eos_token_id UpperCAmelCase_ : List[str] = vocab_size UpperCAmelCase_ : Any = hidden_size UpperCAmelCase_ : List[Any] = num_hidden_layers UpperCAmelCase_ : Tuple = num_attention_heads UpperCAmelCase_ : int = hidden_act UpperCAmelCase_ : Union[str, Any] = intermediate_size UpperCAmelCase_ : Tuple = hidden_dropout_prob UpperCAmelCase_ : Any = attention_probs_dropout_prob UpperCAmelCase_ : Union[str, Any] = max_position_embeddings UpperCAmelCase_ : List[str] = type_vocab_size UpperCAmelCase_ : Optional[int] = initializer_range UpperCAmelCase_ : Optional[Any] = layer_norm_eps UpperCAmelCase_ : Optional[Any] = onnx_export class UpperCamelCase_ (__A ): def __init__( self : List[Any] , lowerCAmelCase_ : "PretrainedConfig" , lowerCAmelCase_ : str = "default" , lowerCAmelCase_ : "List[PatchingSpec]" = None ) -> Union[str, Any]: super().__init__(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) UpperCAmelCase_ : int = True @property def _SCREAMING_SNAKE_CASE ( self : Dict ) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": UpperCAmelCase_ : Tuple = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCAmelCase_ : Optional[Any] = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("global_attention_mask", dynamic_axis), ] ) @property def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Mapping[str, Mapping[int, str]]: UpperCAmelCase_ : Dict = super().outputs if self.task == "default": UpperCAmelCase_ : List[str] = {0: "batch"} return outputs @property def _SCREAMING_SNAKE_CASE ( self : int ) -> float: return 1e-4 @property def _SCREAMING_SNAKE_CASE ( self : Tuple ) -> int: # needs to be >= 14 to support tril operator return max(super().default_onnx_opset , 14 ) def _SCREAMING_SNAKE_CASE ( self : int , lowerCAmelCase_ : "PreTrainedTokenizerBase" , lowerCAmelCase_ : int = -1 , lowerCAmelCase_ : int = -1 , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : Optional[TensorType] = None , ) -> Mapping[str, Any]: UpperCAmelCase_ : Tuple = super().generate_dummy_inputs( preprocessor=lowerCAmelCase_ , batch_size=lowerCAmelCase_ , seq_length=lowerCAmelCase_ , is_pair=lowerCAmelCase_ , framework=lowerCAmelCase_ ) import torch # for some reason, replacing this code by inputs["global_attention_mask"] = torch.randint(2, inputs["input_ids"].shape, dtype=torch.int64) # makes the export fail randomly UpperCAmelCase_ : str = torch.zeros_like(inputs["input_ids"] ) # make every second token global UpperCAmelCase_ : Union[str, Any] = 1 return inputs
253
0
import warnings from contextlib import contextmanager from ....processing_utils import ProcessorMixin class __lowerCAmelCase ( SCREAMING_SNAKE_CASE ): _a = """MCTCTFeatureExtractor""" _a = """AutoTokenizer""" def __init__( self , lowerCAmelCase , lowerCAmelCase ) -> Dict: '''simple docstring''' super().__init__(lowerCAmelCase , lowerCAmelCase ) _lowercase =self.feature_extractor _lowercase =False def __call__( self , *lowerCAmelCase , **lowerCAmelCase ) -> Any: '''simple docstring''' if self._in_target_context_manager: return self.current_processor(*lowerCAmelCase , **lowerCAmelCase ) if "raw_speech" in kwargs: warnings.warn('Using `raw_speech` as a keyword argument is deprecated. Use `audio` instead.' ) _lowercase =kwargs.pop('raw_speech' ) else: _lowercase =kwargs.pop('audio' , lowerCAmelCase ) _lowercase =kwargs.pop('sampling_rate' , lowerCAmelCase ) _lowercase =kwargs.pop('text' , lowerCAmelCase ) if len(lowerCAmelCase ) > 0: _lowercase =args[0] _lowercase =args[1:] if audio is None and text is None: raise ValueError('You need to specify either an `audio` or `text` input to process.' ) if audio is not None: _lowercase =self.feature_extractor(lowerCAmelCase , *lowerCAmelCase , sampling_rate=lowerCAmelCase , **lowerCAmelCase ) if text is not None: _lowercase =self.tokenizer(lowerCAmelCase , **lowerCAmelCase ) if text is None: return inputs elif audio is None: return encodings else: _lowercase =encodings['input_ids'] return inputs def A__ ( self , *lowerCAmelCase , **lowerCAmelCase ) -> Any: '''simple docstring''' return self.tokenizer.batch_decode(*lowerCAmelCase , **lowerCAmelCase ) def A__ ( self , *lowerCAmelCase , **lowerCAmelCase ) -> Union[str, Any]: '''simple docstring''' if self._in_target_context_manager: return self.current_processor.pad(*lowerCAmelCase , **lowerCAmelCase ) _lowercase =kwargs.pop('input_features' , lowerCAmelCase ) _lowercase =kwargs.pop('labels' , lowerCAmelCase ) if len(lowerCAmelCase ) > 0: _lowercase =args[0] _lowercase =args[1:] if input_features is not None: _lowercase =self.feature_extractor.pad(lowerCAmelCase , *lowerCAmelCase , **lowerCAmelCase ) if labels is not None: _lowercase =self.tokenizer.pad(lowerCAmelCase , **lowerCAmelCase ) if labels is None: return input_features elif input_features is None: return labels else: _lowercase =labels['input_ids'] return input_features def A__ ( self , *lowerCAmelCase , **lowerCAmelCase ) -> Any: '''simple docstring''' return self.tokenizer.decode(*lowerCAmelCase , **lowerCAmelCase ) @contextmanager def A__ ( self ) -> int: '''simple docstring''' warnings.warn( '`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ' 'labels by using the argument `text` of the regular `__call__` method (either in the same call as ' 'your audio inputs, or in a separate call.' ) _lowercase =True _lowercase =self.tokenizer yield _lowercase =self.feature_extractor _lowercase =False
205
import numpy as np from cva import COLOR_BGR2GRAY, CV_8UC3, cvtColor, filteraD, imread, imshow, waitKey def a ( A__ : int , A__ : int , A__ : int , A__ : int , A__ : int , A__ : int ) -> np.ndarray: """simple docstring""" if (ksize % 2) == 0: _lowercase =ksize + 1 _lowercase =np.zeros((ksize, ksize) , dtype=np.floataa ) # each value for y in range(A__ ): for x in range(A__ ): # distance from center _lowercase =x - ksize // 2 _lowercase =y - ksize // 2 # degree to radiant _lowercase =theta / 180 * np.pi _lowercase =np.cos(_theta ) _lowercase =np.sin(_theta ) # get kernel x _lowercase =cos_theta * px + sin_theta * py # get kernel y _lowercase =-sin_theta * px + cos_theta * py # fill kernel _lowercase =np.exp( -(_x**2 + gamma**2 * _y**2) / (2 * sigma**2) ) * np.cos(2 * np.pi * _x / lambd + psi ) return gabor if __name__ == "__main__": import doctest doctest.testmod() # read original image lowercase_ = imread('../image_data/lena.jpg') # turn image in gray scale value lowercase_ = cvtColor(img, COLOR_BGR2GRAY) # Apply multiple Kernel to detect edges lowercase_ = np.zeros(gray.shape[:2]) for theta in [0, 3_0, 6_0, 9_0, 1_2_0, 1_5_0]: lowercase_ = gabor_filter_kernel(1_0, 8, theta, 1_0, 0, 0) out += filteraD(gray, CV_8UC3, kernel_aa) lowercase_ = out / out.max() * 2_5_5 lowercase_ = out.astype(np.uinta) imshow('Original', gray) imshow('Gabor filter with 20x20 mask and 6 directions', out) waitKey(0)
205
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool class lowercase ( snake_case__): """simple docstring""" a__ : str = "philschmid/bart-large-cnn-samsum" a__ : List[Any] = ( "This is a tool that summarizes an English text. It takes an input `text` containing the text to summarize, " "and returns a summary of the text." ) a__ : List[str] = "summarizer" a__ : List[Any] = AutoTokenizer a__ : Tuple = AutoModelForSeqaSeqLM a__ : str = ["text"] a__ : Optional[int] = ["text"] def _SCREAMING_SNAKE_CASE ( self : List[str] , __UpperCAmelCase : List[str] ) -> Any: return self.pre_processor(__UpperCAmelCase , return_tensors="""pt""" , truncation=__UpperCAmelCase ) def _SCREAMING_SNAKE_CASE ( self : List[Any] , __UpperCAmelCase : Tuple ) -> List[str]: return self.model.generate(**__UpperCAmelCase )[0] def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , __UpperCAmelCase : Union[str, Any] ) -> int: return self.pre_processor.decode(__UpperCAmelCase , skip_special_tokens=__UpperCAmelCase , clean_up_tokenization_spaces=__UpperCAmelCase )
369
import shutil import tempfile import unittest import numpy as np import pytest from transformers.testing_utils import require_vision from transformers.utils import is_vision_available if is_vision_available(): from PIL import Image from transformers import AutoProcessor, BlipaProcessor, BlipImageProcessor, GPTaTokenizer, PreTrainedTokenizerFast @require_vision class lowercase ( unittest.TestCase): """simple docstring""" def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> List[Any]: UpperCAmelCase_= tempfile.mkdtemp() UpperCAmelCase_= BlipImageProcessor() UpperCAmelCase_= GPTaTokenizer.from_pretrained("""hf-internal-testing/tiny-random-GPT2Model""" ) UpperCAmelCase_= BlipaProcessor(__UpperCAmelCase , __UpperCAmelCase ) processor.save_pretrained(self.tmpdirname ) def _SCREAMING_SNAKE_CASE ( self : List[Any] , **__UpperCAmelCase : Union[str, Any] ) -> int: return AutoProcessor.from_pretrained(self.tmpdirname , **__UpperCAmelCase ).tokenizer def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , **__UpperCAmelCase : str ) -> Optional[int]: return AutoProcessor.from_pretrained(self.tmpdirname , **__UpperCAmelCase ).image_processor def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> List[Any]: shutil.rmtree(self.tmpdirname ) def _SCREAMING_SNAKE_CASE ( self : List[Any] ) -> Any: UpperCAmelCase_= [np.random.randint(255 , size=(3, 30, 400) , dtype=np.uinta )] UpperCAmelCase_= [Image.fromarray(np.moveaxis(__UpperCAmelCase , 0 , -1 ) ) for x in image_inputs] return image_inputs def _SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Dict: UpperCAmelCase_= BlipaProcessor(tokenizer=self.get_tokenizer() , image_processor=self.get_image_processor() ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase_= self.get_tokenizer(bos_token="""(BOS)""" , eos_token="""(EOS)""" ) UpperCAmelCase_= self.get_image_processor(do_normalize=__UpperCAmelCase , padding_value=1.0 ) UpperCAmelCase_= BlipaProcessor.from_pretrained( self.tmpdirname , bos_token="""(BOS)""" , eos_token="""(EOS)""" , do_normalize=__UpperCAmelCase , padding_value=1.0 ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) self.assertIsInstance(processor.tokenizer , __UpperCAmelCase ) self.assertEqual(processor.image_processor.to_json_string() , image_processor_add_kwargs.to_json_string() ) self.assertIsInstance(processor.image_processor , __UpperCAmelCase ) def _SCREAMING_SNAKE_CASE ( self : Dict ) -> Optional[int]: UpperCAmelCase_= self.get_image_processor() UpperCAmelCase_= self.get_tokenizer() UpperCAmelCase_= BlipaProcessor(tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) UpperCAmelCase_= self.prepare_image_inputs() UpperCAmelCase_= image_processor(__UpperCAmelCase , return_tensors="""np""" ) UpperCAmelCase_= processor(images=__UpperCAmelCase , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1E-2 ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Any: UpperCAmelCase_= self.get_image_processor() UpperCAmelCase_= self.get_tokenizer() UpperCAmelCase_= BlipaProcessor(tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) UpperCAmelCase_= """lower newer""" UpperCAmelCase_= processor(text=__UpperCAmelCase ) UpperCAmelCase_= tokenizer(__UpperCAmelCase , return_token_type_ids=__UpperCAmelCase ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> Dict: UpperCAmelCase_= self.get_image_processor() UpperCAmelCase_= self.get_tokenizer() UpperCAmelCase_= BlipaProcessor(tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) UpperCAmelCase_= """lower newer""" UpperCAmelCase_= self.prepare_image_inputs() UpperCAmelCase_= processor(text=__UpperCAmelCase , images=__UpperCAmelCase ) self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """input_ids""", """attention_mask"""] ) # test if it raises when no input is passed with pytest.raises(__UpperCAmelCase ): processor() def _SCREAMING_SNAKE_CASE ( self : str ) -> Any: UpperCAmelCase_= self.get_image_processor() UpperCAmelCase_= self.get_tokenizer() UpperCAmelCase_= BlipaProcessor(tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) UpperCAmelCase_= [[1, 4, 5, 8, 1, 0, 8], [3, 4, 3, 1, 1, 8, 9]] UpperCAmelCase_= processor.batch_decode(__UpperCAmelCase ) UpperCAmelCase_= tokenizer.batch_decode(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) def _SCREAMING_SNAKE_CASE ( self : Dict ) -> int: UpperCAmelCase_= self.get_image_processor() UpperCAmelCase_= self.get_tokenizer() UpperCAmelCase_= BlipaProcessor(tokenizer=__UpperCAmelCase , image_processor=__UpperCAmelCase ) UpperCAmelCase_= """lower newer""" UpperCAmelCase_= self.prepare_image_inputs() UpperCAmelCase_= processor(text=__UpperCAmelCase , images=__UpperCAmelCase ) # For now the processor supports only ['pixel_values', 'input_ids', 'attention_mask'] self.assertListEqual(list(inputs.keys() ) , ["""pixel_values""", """input_ids""", """attention_mask"""] )
277
0
from sklearn.metrics import recall_score import datasets _A = "\nRecall is the fraction of the positive examples that were correctly labeled by the model as positive. It can be computed with the equation:\nRecall = TP / (TP + FN)\nWhere TP is the true positives and FN is the false negatives.\n" _A = "\nArgs:\n- **predictions** (`list` of `int`): The predicted labels.\n- **references** (`list` of `int`): The ground truth labels.\n- **labels** (`list` of `int`): The set of labels to include when `average` is not set to `binary`, and their order when average is `None`. Labels present in the data can be excluded in this input, for example to calculate a multiclass average ignoring a majority negative class, while labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in y_true and y_pred are used in sorted order. Defaults to None.\n- **pos_label** (`int`): The class label to use as the 'positive class' when calculating the recall. Defaults to `1`.\n- **average** (`string`): This parameter is required for multiclass/multilabel targets. If None, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `'binary'`.\n - `'binary'`: Only report results for the class specified by `pos_label`. This is applicable only if the target labels and predictions are binary.\n - `'micro'`: Calculate metrics globally by counting the total true positives, false negatives, and false positives.\n - `'macro'`: Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account.\n - `'weighted'`: Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `'macro'` to account for label imbalance. Note that it can result in an F-score that is not between precision and recall.\n - `'samples'`: Calculate metrics for each instance, and find their average (only meaningful for multilabel classification).\n- **sample_weight** (`list` of `float`): Sample weights Defaults to `None`.\n- **zero_division** (): Sets the value to return when there is a zero division. Defaults to .\n - `'warn'`: If there is a zero division, the return value is `0`, but warnings are also raised.\n - `0`: If there is a zero division, the return value is `0`.\n - `1`: If there is a zero division, the return value is `1`.\n\nReturns:\n- **recall** (`float`, or `array` of `float`): Either the general recall score, or the recall scores for individual classes, depending on the values input to `labels` and `average`. Minimum possible value is 0. Maximum possible value is 1. A higher recall means that more of the positive examples have been labeled correctly. Therefore, a higher recall is generally considered better.\n\nExamples:\n\n Example 1-A simple example with some errors\n >>> recall_metric = datasets.load_metric('recall')\n >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1])\n >>> print(results)\n {'recall': 0.6666666666666666}\n\n Example 2-The same example as Example 1, but with `pos_label=0` instead of the default `pos_label=1`.\n >>> recall_metric = datasets.load_metric('recall')\n >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1], pos_label=0)\n >>> print(results)\n {'recall': 0.5}\n\n Example 3-The same example as Example 1, but with `sample_weight` included.\n >>> recall_metric = datasets.load_metric('recall')\n >>> sample_weight = [0.9, 0.2, 0.9, 0.3, 0.8]\n >>> results = recall_metric.compute(references=[0, 0, 1, 1, 1], predictions=[0, 1, 0, 1, 1], sample_weight=sample_weight)\n >>> print(results)\n {'recall': 0.55}\n\n Example 4-A multiclass example, using different averages.\n >>> recall_metric = datasets.load_metric('recall')\n >>> predictions = [0, 2, 1, 0, 0, 1]\n >>> references = [0, 1, 2, 0, 1, 2]\n >>> results = recall_metric.compute(predictions=predictions, references=references, average='macro')\n >>> print(results)\n {'recall': 0.3333333333333333}\n >>> results = recall_metric.compute(predictions=predictions, references=references, average='micro')\n >>> print(results)\n {'recall': 0.3333333333333333}\n >>> results = recall_metric.compute(predictions=predictions, references=references, average='weighted')\n >>> print(results)\n {'recall': 0.3333333333333333}\n >>> results = recall_metric.compute(predictions=predictions, references=references, average=None)\n >>> print(results)\n {'recall': array([1., 0., 0.])}\n" _A = "\n@article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011}\n" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCAmelCase ( datasets.Metric ): def __a ( self ) -> Union[str, Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Sequence(datasets.Value("int32" ) ), "references": datasets.Sequence(datasets.Value("int32" ) ), } if self.config_name == "multilabel" else { "predictions": datasets.Value("int32" ), "references": datasets.Value("int32" ), } ) , reference_urls=["https://scikit-learn.org/stable/modules/generated/sklearn.metrics.recall_score.html"] , ) def __a ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase=None , _UpperCamelCase=1 , _UpperCamelCase="binary" , _UpperCamelCase=None , _UpperCamelCase="warn" , ) -> str: lowerCAmelCase_ = recall_score( _UpperCamelCase , _UpperCamelCase , labels=_UpperCamelCase , pos_label=_UpperCamelCase , average=_UpperCamelCase , sample_weight=_UpperCamelCase , zero_division=_UpperCamelCase , ) return {"recall": float(_UpperCamelCase ) if score.size == 1 else score}
231
from __future__ import annotations import csv import requests from bsa import BeautifulSoup def lowerCamelCase__ ( __lowerCAmelCase : str = "" ): """simple docstring""" lowerCAmelCase_ = url or "https://www.imdb.com/chart/top/?ref_=nv_mv_250" lowerCAmelCase_ = BeautifulSoup(requests.get(__lowerCAmelCase ).text , "html.parser" ) lowerCAmelCase_ = soup.find_all("td" , attrs="titleColumn" ) lowerCAmelCase_ = soup.find_all("td" , class_="ratingColumn imdbRating" ) return { title.a.text: float(rating.strong.text ) for title, rating in zip(__lowerCAmelCase , __lowerCAmelCase ) } def lowerCamelCase__ ( __lowerCAmelCase : str = "IMDb_Top_250_Movies.csv" ): """simple docstring""" lowerCAmelCase_ = get_imdb_top_aaa_movies() with open(__lowerCAmelCase , "w" , newline="" ) as out_file: lowerCAmelCase_ = csv.writer(__lowerCAmelCase ) writer.writerow(["Movie title", "IMDb rating"] ) for title, rating in movies.items(): writer.writerow([title, rating] ) if __name__ == "__main__": write_movies()
231
1
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case : Optional[Any] = logging.get_logger(__name__) _snake_case : List[Any] = { "facebook/wav2vec2-base-960h": "https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/config.json", # See all Wav2Vec2 models at https://huggingface.co/models?filter=wav2vec2 } class a (_lowerCAmelCase ): """simple docstring""" __UpperCAmelCase : Tuple = "wav2vec2" def __init__( self : str , lowerCamelCase : List[Any]=32 , lowerCamelCase : Any=768 , lowerCamelCase : Dict=12 , lowerCamelCase : Optional[int]=12 , lowerCamelCase : Optional[int]=3072 , lowerCamelCase : List[Any]="gelu" , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : str=0.1 , lowerCamelCase : List[Any]=0.1 , lowerCamelCase : str=0.0 , lowerCamelCase : List[Any]=0.0 , lowerCamelCase : List[str]=0.1 , lowerCamelCase : List[Any]=0.1 , lowerCamelCase : Any=0.02 , lowerCamelCase : Optional[Any]=1E-5 , lowerCamelCase : Union[str, Any]="group" , lowerCamelCase : Tuple="gelu" , lowerCamelCase : Dict=(512, 512, 512, 512, 512, 512, 512) , lowerCamelCase : List[Any]=(5, 2, 2, 2, 2, 2, 2) , lowerCamelCase : str=(10, 3, 3, 3, 3, 2, 2) , lowerCamelCase : int=False , lowerCamelCase : Optional[int]=128 , lowerCamelCase : Optional[int]=16 , lowerCamelCase : str=False , lowerCamelCase : Tuple=True , lowerCamelCase : Dict=0.05 , lowerCamelCase : Union[str, Any]=10 , lowerCamelCase : Any=2 , lowerCamelCase : Union[str, Any]=0.0 , lowerCamelCase : List[str]=10 , lowerCamelCase : str=0 , lowerCamelCase : Tuple=320 , lowerCamelCase : Optional[Any]=2 , lowerCamelCase : List[Any]=0.1 , lowerCamelCase : int=100 , lowerCamelCase : Union[str, Any]=256 , lowerCamelCase : Optional[int]=256 , lowerCamelCase : Optional[int]=0.1 , lowerCamelCase : Any="sum" , lowerCamelCase : str=False , lowerCamelCase : Dict=False , lowerCamelCase : int=256 , lowerCamelCase : Tuple=(512, 512, 512, 512, 1500) , lowerCamelCase : Optional[int]=(5, 3, 3, 1, 1) , lowerCamelCase : int=(1, 2, 3, 1, 1) , lowerCamelCase : int=512 , lowerCamelCase : Optional[Any]=0 , lowerCamelCase : List[str]=1 , lowerCamelCase : Any=2 , lowerCamelCase : Any=False , lowerCamelCase : Any=3 , lowerCamelCase : List[str]=2 , lowerCamelCase : int=3 , lowerCamelCase : str=None , lowerCamelCase : Dict=None , **lowerCamelCase : str , ) -> Optional[int]: super().__init__(**lowerCamelCase , pad_token_id=lowerCamelCase , bos_token_id=lowerCamelCase , eos_token_id=lowerCamelCase ) __snake_case : int = hidden_size __snake_case : Optional[Any] = feat_extract_norm __snake_case : Dict = feat_extract_activation __snake_case : Any = list(lowerCamelCase ) __snake_case : Union[str, Any] = list(lowerCamelCase ) __snake_case : Optional[Any] = list(lowerCamelCase ) __snake_case : str = conv_bias __snake_case : str = num_conv_pos_embeddings __snake_case : Union[str, Any] = num_conv_pos_embedding_groups __snake_case : List[Any] = len(self.conv_dim ) __snake_case : Optional[Any] = num_hidden_layers __snake_case : Tuple = intermediate_size __snake_case : List[Any] = hidden_act __snake_case : int = num_attention_heads __snake_case : List[str] = hidden_dropout __snake_case : Dict = attention_dropout __snake_case : Dict = activation_dropout __snake_case : Tuple = feat_proj_dropout __snake_case : Any = final_dropout __snake_case : Union[str, Any] = layerdrop __snake_case : List[str] = layer_norm_eps __snake_case : Optional[Any] = initializer_range __snake_case : Tuple = vocab_size __snake_case : int = do_stable_layer_norm __snake_case : Any = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( "Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==" " `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =" F' {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,' F' `len(config.conv_kernel) = {len(self.conv_kernel )}`.' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __snake_case : Optional[Any] = apply_spec_augment __snake_case : int = mask_time_prob __snake_case : Optional[int] = mask_time_length __snake_case : Tuple = mask_time_min_masks __snake_case : Tuple = mask_feature_prob __snake_case : Optional[Any] = mask_feature_length __snake_case : Tuple = mask_feature_min_masks # parameters for pretraining with codevector quantized representations __snake_case : Optional[int] = num_codevectors_per_group __snake_case : List[str] = num_codevector_groups __snake_case : List[str] = contrastive_logits_temperature __snake_case : List[Any] = feat_quantizer_dropout __snake_case : List[Any] = num_negatives __snake_case : Tuple = codevector_dim __snake_case : List[str] = proj_codevector_dim __snake_case : int = diversity_loss_weight # ctc loss __snake_case : Optional[Any] = ctc_loss_reduction __snake_case : Any = ctc_zero_infinity # adapter __snake_case : Optional[int] = add_adapter __snake_case : Union[str, Any] = adapter_kernel_size __snake_case : int = adapter_stride __snake_case : Tuple = num_adapter_layers __snake_case : List[Any] = output_hidden_size or hidden_size __snake_case : Dict = adapter_attn_dim # SequenceClassification-specific parameter. Feel free to ignore for other classes. __snake_case : Optional[Any] = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. __snake_case : Optional[int] = list(lowerCamelCase ) __snake_case : List[str] = list(lowerCamelCase ) __snake_case : Optional[int] = list(lowerCamelCase ) __snake_case : List[Any] = xvector_output_dim @property def __snake_case ( self : int ) -> Any: return functools.reduce(operator.mul , self.conv_stride , 1 )
369
def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): if index == r: for j in range(__lowerCamelCase ): print(data[j] , end=" " ) print(" " ) return # When no more elements are there to put in data[] if i >= n: return # current is included, put next at next location __snake_case : Union[str, Any] = arr[i] combination_util(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , index + 1 , __lowerCamelCase , i + 1 ) # current is excluded, replace it with # next (Note that i+1 is passed, but # index is not changed) combination_util(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , i + 1 ) # The main function that prints all combinations # of size r in arr[] of size n. This function # mainly uses combinationUtil() def lowerCAmelCase_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): # A temporary array to store all combination one by one __snake_case : Union[str, Any] = [0] * r # Print all combination using temporary array 'data[]' combination_util(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase , 0 , __lowerCamelCase , 0 ) if __name__ == "__main__": # Driver code to check the function above _snake_case : List[str] = [10, 20, 30, 40, 50] print_combination(arr, len(arr), 3) # This code is contributed by Ambuj sahu
134
0
from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def _snake_case( SCREAMING_SNAKE_CASE__ : Tuple ) -> Tuple: '''simple docstring''' return {key.lstrip('-' ): value for key, value in zip(unknown_args[::2] , unknown_args[1::2] )} def _snake_case( ) -> Dict: '''simple docstring''' A__ = ArgumentParser( 'HuggingFace Datasets CLI tool' , usage='datasets-cli <command> [<args>]' , allow_abbrev=SCREAMING_SNAKE_CASE__ ) A__ = parser.add_subparsers(help='datasets-cli command helpers' ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(SCREAMING_SNAKE_CASE__ ) EnvironmentCommand.register_subcommand(SCREAMING_SNAKE_CASE__ ) TestCommand.register_subcommand(SCREAMING_SNAKE_CASE__ ) RunBeamCommand.register_subcommand(SCREAMING_SNAKE_CASE__ ) DummyDataCommand.register_subcommand(SCREAMING_SNAKE_CASE__ ) # Parse args A__ , A__ = parser.parse_known_args() if not hasattr(SCREAMING_SNAKE_CASE__ , 'func' ): parser.print_help() exit(1 ) A__ = parse_unknown_args(SCREAMING_SNAKE_CASE__ ) # Run A__ = args.func(SCREAMING_SNAKE_CASE__ , **SCREAMING_SNAKE_CASE__ ) service.run() if __name__ == "__main__": main()
7
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, PNDMScheduler, StableDiffusionLDMaDPipeline, UNetaDConditionModel, ) from diffusers.utils import nightly, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS enable_full_determinism() class __lowerCAmelCase ( unittest.TestCase ): snake_case_ : Tuple = StableDiffusionLDMaDPipeline snake_case_ : Optional[int] = TEXT_TO_IMAGE_PARAMS snake_case_ : str = TEXT_TO_IMAGE_BATCH_PARAMS snake_case_ : List[str] = TEXT_TO_IMAGE_IMAGE_PARAMS def UpperCamelCase ( self : Optional[int] ): """simple docstring""" torch.manual_seed(0 ) _UpperCAmelCase = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) _UpperCAmelCase = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=snake_case__ , set_alpha_to_one=snake_case__ , ) torch.manual_seed(0 ) _UpperCAmelCase = AutoencoderKL( block_out_channels=[32, 64] , in_channels=6 , out_channels=6 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) torch.manual_seed(0 ) _UpperCAmelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , ) _UpperCAmelCase = CLIPTextModel(snake_case__ ) _UpperCAmelCase = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) _UpperCAmelCase = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCamelCase ( self : Optional[int] , snake_case__ : Dict , snake_case__ : Optional[int]=0 ): """simple docstring""" if str(snake_case__ ).startswith("mps" ): _UpperCAmelCase = torch.manual_seed(snake_case__ ) else: _UpperCAmelCase = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) _UpperCAmelCase = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def UpperCamelCase ( self : Optional[Any] ): """simple docstring""" _UpperCAmelCase = "cpu" # ensure determinism for the device-dependent torch.Generator _UpperCAmelCase = self.get_dummy_components() _UpperCAmelCase = StableDiffusionLDMaDPipeline(**snake_case__ ) _UpperCAmelCase = ldmad_pipe.to(snake_case__ ) ldmad_pipe.set_progress_bar_config(disable=snake_case__ ) _UpperCAmelCase = self.get_dummy_inputs(snake_case__ ) _UpperCAmelCase = ldmad_pipe(**snake_case__ ) _UpperCAmelCase , _UpperCAmelCase = output.rgb, output.depth _UpperCAmelCase = rgb[0, -3:, -3:, -1] _UpperCAmelCase = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) _UpperCAmelCase = np.array( [0.37_338_176, 0.70_247, 0.74_203_193, 0.51_643_604, 0.58_256_793, 0.60_932_136, 0.4_181_095, 0.48_355_877, 0.46_535_262] ) _UpperCAmelCase = np.array([103.46_727, 85.812_004, 87.849_236] ) assert np.abs(image_slice_rgb.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(image_slice_depth.flatten() - expected_slice_depth ).max() < 1e-2 def UpperCamelCase ( self : List[str] ): """simple docstring""" _UpperCAmelCase = self.get_dummy_components() _UpperCAmelCase = StableDiffusionLDMaDPipeline(**snake_case__ ) _UpperCAmelCase = ldmad_pipe.to(snake_case__ ) ldmad_pipe.set_progress_bar_config(disable=snake_case__ ) _UpperCAmelCase = self.get_dummy_inputs(snake_case__ ) _UpperCAmelCase = 3 * [inputs["prompt"]] # forward _UpperCAmelCase = ldmad_pipe(**snake_case__ ) _UpperCAmelCase , _UpperCAmelCase = output.rgb, output.depth _UpperCAmelCase = rgb_slice_a[0, -3:, -3:, -1] _UpperCAmelCase = depth_slice_a[0, -3:, -1] _UpperCAmelCase = self.get_dummy_inputs(snake_case__ ) _UpperCAmelCase = 3 * [inputs.pop("prompt" )] _UpperCAmelCase = ldmad_pipe.tokenizer( snake_case__ , padding="max_length" , max_length=ldmad_pipe.tokenizer.model_max_length , truncation=snake_case__ , return_tensors="pt" , ) _UpperCAmelCase = text_inputs["input_ids"].to(snake_case__ ) _UpperCAmelCase = ldmad_pipe.text_encoder(snake_case__ )[0] _UpperCAmelCase = prompt_embeds # forward _UpperCAmelCase = ldmad_pipe(**snake_case__ ) _UpperCAmelCase , _UpperCAmelCase = output.rgb, output.depth _UpperCAmelCase = rgb_slice_a[0, -3:, -3:, -1] _UpperCAmelCase = depth_slice_a[0, -3:, -1] assert np.abs(rgb_slice_a.flatten() - rgb_slice_a.flatten() ).max() < 1e-4 assert np.abs(depth_slice_a.flatten() - depth_slice_a.flatten() ).max() < 1e-4 def UpperCamelCase ( self : List[str] ): """simple docstring""" _UpperCAmelCase = "cpu" # ensure determinism for the device-dependent torch.Generator _UpperCAmelCase = self.get_dummy_components() _UpperCAmelCase = PNDMScheduler(skip_prk_steps=snake_case__ ) _UpperCAmelCase = StableDiffusionLDMaDPipeline(**snake_case__ ) _UpperCAmelCase = ldmad_pipe.to(snake_case__ ) ldmad_pipe.set_progress_bar_config(disable=snake_case__ ) _UpperCAmelCase = self.get_dummy_inputs(snake_case__ ) _UpperCAmelCase = "french fries" _UpperCAmelCase = ldmad_pipe(**snake_case__ , negative_prompt=snake_case__ ) _UpperCAmelCase , _UpperCAmelCase = output.rgb, output.depth _UpperCAmelCase = rgb[0, -3:, -3:, -1] _UpperCAmelCase = depth[0, -3:, -1] assert rgb.shape == (1, 64, 64, 3) assert depth.shape == (1, 64, 64) _UpperCAmelCase = np.array( [0.37_044, 0.71_811_503, 0.7_223_251, 0.48_603_675, 0.5_638_391, 0.6_364_948, 0.42_833_704, 0.4_901_315, 0.47_926_217] ) _UpperCAmelCase = np.array([107.84_738, 84.62_802, 89.962_135] ) assert np.abs(rgb_slice.flatten() - expected_slice_rgb ).max() < 1e-2 assert np.abs(depth_slice.flatten() - expected_slice_depth ).max() < 1e-2 @slow @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def UpperCamelCase ( self : Tuple ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self : str , snake_case__ : Optional[int] , snake_case__ : Tuple="cpu" , snake_case__ : Any=torch.floataa , snake_case__ : Dict=0 ): """simple docstring""" _UpperCAmelCase = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) _UpperCAmelCase = np.random.RandomState(snake_case__ ).standard_normal((1, 4, 64, 64) ) _UpperCAmelCase = torch.from_numpy(snake_case__ ).to(device=snake_case__ , dtype=snake_case__ ) _UpperCAmelCase = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCamelCase ( self : Any ): """simple docstring""" _UpperCAmelCase = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d" ) _UpperCAmelCase = ldmad_pipe.to(snake_case__ ) ldmad_pipe.set_progress_bar_config(disable=snake_case__ ) _UpperCAmelCase = self.get_inputs(snake_case__ ) _UpperCAmelCase = ldmad_pipe(**snake_case__ ) _UpperCAmelCase , _UpperCAmelCase = output.rgb, output.depth _UpperCAmelCase = rgb[0, -3:, -3:, -1].flatten() _UpperCAmelCase = rgb[0, -3:, -1].flatten() assert rgb.shape == (1, 512, 512, 3) assert depth.shape == (1, 512, 512) _UpperCAmelCase = np.array( [0.53_805_465, 0.56_707_305, 0.5_486_515, 0.57_012_236, 0.5_814_511, 0.56_253_487, 0.54_843_014, 0.55_092_263, 0.6_459_706] ) _UpperCAmelCase = np.array( [0.9_263_781, 0.6_678_672, 0.5_486_515, 0.92_202_145, 0.67_831_135, 0.56_253_487, 0.9_241_694, 0.7_551_478, 0.6_459_706] ) assert np.abs(rgb_slice - expected_slice_rgb ).max() < 3e-3 assert np.abs(depth_slice - expected_slice_depth ).max() < 3e-3 @nightly @require_torch_gpu class __lowerCAmelCase ( unittest.TestCase ): def UpperCamelCase ( self : Dict ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self : Any , snake_case__ : Optional[Any] , snake_case__ : int="cpu" , snake_case__ : Optional[Any]=torch.floataa , snake_case__ : Optional[Any]=0 ): """simple docstring""" _UpperCAmelCase = torch.Generator(device=snake_case__ ).manual_seed(snake_case__ ) _UpperCAmelCase = np.random.RandomState(snake_case__ ).standard_normal((1, 4, 64, 64) ) _UpperCAmelCase = torch.from_numpy(snake_case__ ).to(device=snake_case__ , dtype=snake_case__ ) _UpperCAmelCase = { "prompt": "a photograph of an astronaut riding a horse", "latents": latents, "generator": generator, "num_inference_steps": 50, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def UpperCamelCase ( self : Union[str, Any] ): """simple docstring""" _UpperCAmelCase = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d" ).to(snake_case__ ) ldmad_pipe.set_progress_bar_config(disable=snake_case__ ) _UpperCAmelCase = self.get_inputs(snake_case__ ) _UpperCAmelCase = ldmad_pipe(**snake_case__ ) _UpperCAmelCase , _UpperCAmelCase = output.rgb, output.depth _UpperCAmelCase = 0.495_586 _UpperCAmelCase = 0.33_795_515 _UpperCAmelCase = 112.48_518 _UpperCAmelCase = 98.489_746 assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3 def UpperCamelCase ( self : Tuple ): """simple docstring""" _UpperCAmelCase = StableDiffusionLDMaDPipeline.from_pretrained("Intel/ldm3d-4c" ).to(snake_case__ ) ldmad_pipe.set_progress_bar_config(disable=snake_case__ ) _UpperCAmelCase = self.get_inputs(snake_case__ ) _UpperCAmelCase = ldmad_pipe(**snake_case__ ) _UpperCAmelCase , _UpperCAmelCase = output.rgb, output.depth _UpperCAmelCase = 0.4_194_127 _UpperCAmelCase = 0.35_375_586 _UpperCAmelCase = 0.5_638_502 _UpperCAmelCase = 0.34_686_103 assert rgb.shape == (1, 512, 512, 3) assert depth.shape == (1, 512, 512, 1) assert np.abs(expected_rgb_mean - rgb.mean() ) < 1e-3 assert np.abs(expected_rgb_std - rgb.std() ) < 1e-3 assert np.abs(expected_depth_mean - depth.mean() ) < 1e-3 assert np.abs(expected_depth_std - depth.std() ) < 1e-3
133
0
'''simple docstring''' import math def UpperCamelCase ( a ) -> list: '''simple docstring''' __magic_name__ = [True] * n __magic_name__ = False __magic_name__ = False __magic_name__ = True for i in range(3 , int(n**0.5 + 1 ) , 2 ): __magic_name__ = i * 2 while index < n: __magic_name__ = False __magic_name__ = index + i __magic_name__ = [2] for i in range(3 , a , 2 ): if is_prime[i]: primes.append(a ) return primes def UpperCamelCase ( a = 9999_6666_3333 ) -> int: '''simple docstring''' __magic_name__ = math.floor(math.sqrt(a ) ) + 100 __magic_name__ = prime_sieve(a ) __magic_name__ = 0 __magic_name__ = 0 __magic_name__ = primes[prime_index] while (last_prime**2) <= limit: __magic_name__ = primes[prime_index + 1] __magic_name__ = last_prime**2 __magic_name__ = next_prime**2 # Get numbers divisible by lps(current) __magic_name__ = lower_bound + last_prime while upper_bound > current <= limit: matches_sum += current current += last_prime # Reset the upper_bound while (upper_bound - next_prime) > limit: upper_bound -= next_prime # Add the numbers divisible by ups(current) __magic_name__ = upper_bound - next_prime while current > lower_bound: matches_sum += current current -= next_prime # Remove the numbers divisible by both ups and lps __magic_name__ = 0 while upper_bound > current <= limit: if current <= lower_bound: # Increment the current number current += last_prime * next_prime continue if current > limit: break # Remove twice since it was added by both ups and lps matches_sum -= current * 2 # Increment the current number current += last_prime * next_prime # Setup for next pair __magic_name__ = next_prime prime_index += 1 return matches_sum if __name__ == "__main__": print(solution())
98
'''simple docstring''' from pathlib import Path import fire from tqdm import tqdm def UpperCamelCase ( a="ro" , a="en" , a="wmt16" , a=None ) -> None: '''simple docstring''' try: import datasets except (ModuleNotFoundError, ImportError): raise ImportError('''run pip install datasets''' ) __magic_name__ = F'''{src_lang}-{tgt_lang}''' print(F'''Converting {dataset}-{pair}''' ) __magic_name__ = datasets.load_dataset(a , a ) if save_dir is None: __magic_name__ = F'''{dataset}-{pair}''' __magic_name__ = Path(a ) save_dir.mkdir(exist_ok=a ) for split in ds.keys(): print(F'''Splitting {split} with {ds[split].num_rows} records''' ) # to save to val.source, val.target like summary datasets __magic_name__ = '''val''' if split == '''validation''' else split __magic_name__ = save_dir.joinpath(F'''{fn}.source''' ) __magic_name__ = save_dir.joinpath(F'''{fn}.target''' ) __magic_name__ = src_path.open('''w+''' ) __magic_name__ = tgt_path.open('''w+''' ) # reader is the bottleneck so writing one record at a time doesn't slow things down for x in tqdm(ds[split] ): __magic_name__ = x['''translation'''] src_fp.write(ex[src_lang] + '''\n''' ) tgt_fp.write(ex[tgt_lang] + '''\n''' ) print(F'''Saved {dataset} dataset to {save_dir}''' ) if __name__ == "__main__": fire.Fire(download_wmt_dataset)
98
1
"""simple docstring""" import numpy as np def A_ ( _lowerCAmelCase : np.array ): """simple docstring""" return 1 / (1 + np.exp(-vector )) def A_ ( _lowerCAmelCase : np.array ): """simple docstring""" return vector * sigmoid(1.7_0_2 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
320
"""simple docstring""" import argparse import torch from transformers import ( SpeechTaConfig, SpeechTaFeatureExtractor, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaProcessor, SpeechTaTokenizer, logging, ) from transformers.tokenization_utils import AddedToken logging.set_verbosity_info() __snake_case = logging.get_logger('''transformers.models.speecht5''') __snake_case = { '''speech_encoder_prenet.layer_norm''': '''speecht5.encoder.prenet.feature_projection.layer_norm''', '''speech_encoder_prenet.post_extract_proj''': '''speecht5.encoder.prenet.feature_projection.projection''', '''speech_encoder_prenet.pos_conv.0''': '''speecht5.encoder.prenet.pos_conv_embed.conv''', '''speech_encoder_prenet.mask_emb''': '''speecht5.encoder.prenet.masked_spec_embed''', } __snake_case = { '''text_encoder_prenet.encoder_prenet.0''': '''speecht5.encoder.prenet.embed_tokens''', '''text_encoder_prenet.encoder_prenet.1.alpha''': '''speecht5.encoder.prenet.encode_positions.alpha''', } __snake_case = { '''speech_decoder_prenet.decoder_prenet.0.0.prenet.0.0''': '''speecht5.decoder.prenet.layers.0''', '''speech_decoder_prenet.decoder_prenet.0.0.prenet.1.0''': '''speecht5.decoder.prenet.layers.1''', '''speech_decoder_prenet.decoder_prenet.0.1''': '''speecht5.decoder.prenet.final_layer''', '''speech_decoder_prenet.decoder_prenet.1.alpha''': '''speecht5.decoder.prenet.encode_positions.alpha''', '''speech_decoder_prenet.spkembs_layer.0''': '''speecht5.decoder.prenet.speaker_embeds_layer''', } __snake_case = { '''speech_decoder_postnet.feat_out''': '''speech_decoder_postnet.feat_out''', '''speech_decoder_postnet.prob_out''': '''speech_decoder_postnet.prob_out''', '''speech_decoder_postnet.postnet.postnet.0.0''': '''speech_decoder_postnet.layers.0.conv''', '''speech_decoder_postnet.postnet.postnet.0.1''': '''speech_decoder_postnet.layers.0.batch_norm''', '''speech_decoder_postnet.postnet.postnet.1.0''': '''speech_decoder_postnet.layers.1.conv''', '''speech_decoder_postnet.postnet.postnet.1.1''': '''speech_decoder_postnet.layers.1.batch_norm''', '''speech_decoder_postnet.postnet.postnet.2.0''': '''speech_decoder_postnet.layers.2.conv''', '''speech_decoder_postnet.postnet.postnet.2.1''': '''speech_decoder_postnet.layers.2.batch_norm''', '''speech_decoder_postnet.postnet.postnet.3.0''': '''speech_decoder_postnet.layers.3.conv''', '''speech_decoder_postnet.postnet.postnet.3.1''': '''speech_decoder_postnet.layers.3.batch_norm''', '''speech_decoder_postnet.postnet.postnet.4.0''': '''speech_decoder_postnet.layers.4.conv''', '''speech_decoder_postnet.postnet.postnet.4.1''': '''speech_decoder_postnet.layers.4.batch_norm''', } __snake_case = { '''text_decoder_prenet.embed_tokens''': '''speecht5.decoder.prenet.embed_tokens''', } __snake_case = { '''text_decoder_postnet.output_projection''': '''text_decoder_postnet.lm_head''', } __snake_case = { '''encoder.layers.*.self_attn.k_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.k_proj''', '''encoder.layers.*.self_attn.v_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.v_proj''', '''encoder.layers.*.self_attn.q_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.q_proj''', '''encoder.layers.*.self_attn.out_proj''': '''speecht5.encoder.wrapped_encoder.layers.*.attention.out_proj''', '''encoder.layers.*.self_attn_layer_norm''': '''speecht5.encoder.wrapped_encoder.layers.*.layer_norm''', '''encoder.layers.*.fc1''': '''speecht5.encoder.wrapped_encoder.layers.*.feed_forward.intermediate_dense''', '''encoder.layers.*.fc2''': '''speecht5.encoder.wrapped_encoder.layers.*.feed_forward.output_dense''', '''encoder.layers.*.final_layer_norm''': '''speecht5.encoder.wrapped_encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''speecht5.encoder.wrapped_encoder.layer_norm''', '''encoder.pos_emb.pe_k''': '''speecht5.encoder.wrapped_encoder.embed_positions.pe_k''', } __snake_case = { '''decoder.layers.*.self_attn.k_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.k_proj''', '''decoder.layers.*.self_attn.v_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.v_proj''', '''decoder.layers.*.self_attn.q_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.q_proj''', '''decoder.layers.*.self_attn.out_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn.out_proj''', '''decoder.layers.*.self_attn_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.self_attn_layer_norm''', '''decoder.layers.*.encoder_attn.k_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.k_proj''', '''decoder.layers.*.encoder_attn.v_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.v_proj''', '''decoder.layers.*.encoder_attn.q_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.q_proj''', '''decoder.layers.*.encoder_attn.out_proj''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn.out_proj''', '''decoder.layers.*.encoder_attn_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.encoder_attn_layer_norm''', '''decoder.layers.*.fc1''': '''speecht5.decoder.wrapped_decoder.layers.*.feed_forward.intermediate_dense''', '''decoder.layers.*.fc2''': '''speecht5.decoder.wrapped_decoder.layers.*.feed_forward.output_dense''', '''decoder.layers.*.final_layer_norm''': '''speecht5.decoder.wrapped_decoder.layers.*.final_layer_norm''', } __snake_case = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_TEXT_DECODER_PRENET, **MAPPING_TEXT_DECODER_POSTNET, } __snake_case = { **MAPPING_TEXT_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } __snake_case = { **MAPPING_SPEECH_ENCODER_PRENET, **MAPPING_ENCODER, **MAPPING_DECODER, **MAPPING_SPEECH_DECODER_PRENET, **MAPPING_SPEECH_DECODER_POSTNET, } __snake_case = [] __snake_case = [ '''encoder.version''', '''encoder.layers.*.norm_k.weight''', '''encoder.layers.*.norm_k.bias''', '''decoder.version''', '''decoder.layers.*.norm_k.weight''', '''decoder.layers.*.norm_k.bias''', '''decoder.pos_emb.pe_k''', '''speech_encoder_prenet.embed_positions._float_tensor''', '''text_decoder_prenet.embed_positions._float_tensor''', ] __snake_case = IGNORE_KEYS + [ '''encoder.proj''', '''text_encoder_prenet.*''', '''speech_decoder_prenet.*''', '''speech_decoder_postnet.*''', ] __snake_case = IGNORE_KEYS + [ '''encoder.proj''', '''speech_encoder_prenet.*''', '''text_decoder_prenet.*''', '''text_decoder_postnet.*''', ] __snake_case = IGNORE_KEYS + [ '''encoder.proj''', '''text_encoder_prenet.*''', '''text_decoder_prenet.*''', '''text_decoder_postnet.*''', ] def A_ ( _lowerCAmelCase : Optional[Any], _lowerCAmelCase : Optional[Any], _lowerCAmelCase : Tuple, _lowerCAmelCase : Dict, _lowerCAmelCase : Optional[int] ): """simple docstring""" for attribute in key.split('''.''' ): _a = getattr(_lowerCAmelCase, _lowerCAmelCase ) if weight_type is not None: _a = getattr(_lowerCAmelCase, _lowerCAmelCase ).shape else: _a = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": _a = value elif weight_type == "weight_g": _a = value elif weight_type == "weight_v": _a = value elif weight_type == "bias": _a = value elif weight_type == "running_mean": _a = value elif weight_type == "running_var": _a = value elif weight_type == "num_batches_tracked": _a = value else: _a = value logger.info(f'{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.' ) def A_ ( _lowerCAmelCase : Optional[Any], _lowerCAmelCase : Tuple ): """simple docstring""" for key in ignore_keys: if key.endswith('''.*''' ): if name.startswith(key[:-1] ): return True elif ".*." in key: _a , _a = key.split('''.*.''' ) if prefix in name and suffix in name: return True elif key in name: return True return False def A_ ( _lowerCAmelCase : Any, _lowerCAmelCase : Union[str, Any], _lowerCAmelCase : int ): """simple docstring""" _a = [] if task == "s2t": _a = hf_model.speechta.encoder.prenet.feature_encoder _a = MAPPING_S2T _a = IGNORE_KEYS_S2T elif task == "t2s": _a = None _a = MAPPING_T2S _a = IGNORE_KEYS_T2S elif task == "s2s": _a = hf_model.speechta.encoder.prenet.feature_encoder _a = MAPPING_S2S _a = IGNORE_KEYS_S2S else: raise ValueError(f'Unsupported task: {task}' ) for name, value in fairseq_dict.items(): if should_ignore(_lowerCAmelCase, _lowerCAmelCase ): logger.info(f'{name} was ignored' ) continue _a = False if "conv_layers" in name: load_conv_layer( _lowerCAmelCase, _lowerCAmelCase, _lowerCAmelCase, _lowerCAmelCase, hf_model.config.feat_extract_norm == '''group''', ) _a = True else: for key, mapped_key in MAPPING.items(): # mapped_key = "speecht5." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if "*" in key: _a , _a = key.split('''.*.''' ) if prefix in name and suffix in name: _a = suffix # if key in name or key.split("w2v_model.")[-1] == name.split(".")[0]: if key in name: _a = True if "*" in mapped_key: _a = name.split(_lowerCAmelCase )[0].split('''.''' )[-2] _a = mapped_key.replace('''*''', _lowerCAmelCase ) if "weight_g" in name: _a = '''weight_g''' elif "weight_v" in name: _a = '''weight_v''' elif "bias" in name: _a = '''bias''' elif "weight" in name: _a = '''weight''' elif "running_mean" in name: _a = '''running_mean''' elif "running_var" in name: _a = '''running_var''' elif "num_batches_tracked" in name: _a = '''num_batches_tracked''' else: _a = None set_recursively(_lowerCAmelCase, _lowerCAmelCase, _lowerCAmelCase, _lowerCAmelCase, _lowerCAmelCase ) continue if not is_used: unused_weights.append(_lowerCAmelCase ) logger.warning(f'Unused weights: {unused_weights}' ) def A_ ( _lowerCAmelCase : Optional[Any], _lowerCAmelCase : Optional[Any], _lowerCAmelCase : Dict, _lowerCAmelCase : List[Any], _lowerCAmelCase : List[Any] ): """simple docstring""" _a = full_name.split('''conv_layers.''' )[-1] _a = name.split('''.''' ) _a = int(items[0] ) _a = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) _a = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) _a = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape} was found.' ) _a = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape} was found.' ) _a = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(_lowerCAmelCase ) @torch.no_grad() def A_ ( _lowerCAmelCase : Union[str, Any], _lowerCAmelCase : Union[str, Any], _lowerCAmelCase : Dict, _lowerCAmelCase : List[Any]=None, _lowerCAmelCase : List[str]=None, _lowerCAmelCase : int=None, ): """simple docstring""" if config_path is not None: _a = SpeechTaConfig.from_pretrained(_lowerCAmelCase ) else: _a = SpeechTaConfig() if task == "s2t": _a = config.max_text_positions _a = SpeechTaForSpeechToText(_lowerCAmelCase ) elif task == "t2s": _a = 18_76 _a = 6_00 _a = config.max_speech_positions _a = SpeechTaForTextToSpeech(_lowerCAmelCase ) elif task == "s2s": _a = 18_76 _a = config.max_speech_positions _a = SpeechTaForSpeechToSpeech(_lowerCAmelCase ) else: raise ValueError(f'Unknown task name: {task}' ) if vocab_path: _a = SpeechTaTokenizer(_lowerCAmelCase, model_max_length=config.max_text_positions ) # Mask token behaves like a normal word, i.e. include the space before it _a = AddedToken('''<mask>''', lstrip=_lowerCAmelCase, rstrip=_lowerCAmelCase ) _a = mask_token tokenizer.add_special_tokens({'''mask_token''': mask_token} ) tokenizer.add_tokens(['''<ctc_blank>'''] ) _a = SpeechTaFeatureExtractor() _a = SpeechTaProcessor(tokenizer=_lowerCAmelCase, feature_extractor=_lowerCAmelCase ) processor.save_pretrained(_lowerCAmelCase ) _a = torch.load(_lowerCAmelCase ) recursively_load_weights(fairseq_checkpoint['''model'''], _lowerCAmelCase, _lowerCAmelCase ) model.save_pretrained(_lowerCAmelCase ) if repo_id: print('''Pushing to the hub...''' ) processor.push_to_hub(_lowerCAmelCase ) model.push_to_hub(_lowerCAmelCase ) if __name__ == "__main__": __snake_case = argparse.ArgumentParser() parser.add_argument( '''--task''', default='''s2t''', type=str, help='''Type of the SpeechT5 model you\'d like to convert. Should be one of \'s2t\', \'t2s\', \'s2s\'.''', ) parser.add_argument('''--checkpoint_path''', required=True, default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--vocab_path''', default=None, type=str, help='''Path to SentencePiece model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--pytorch_dump_folder_path''', required=True, default=None, type=str, help='''Path to the output PyTorch model.''' ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) __snake_case = parser.parse_args() convert_speechta_checkpoint( args.task, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.vocab_path, args.push_to_hub, )
320
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging a_ : List[Any] = logging.get_logger(__name__) class _snake_case ( A__ ): _lowercase : List[str] = ['''pixel_values'''] def __init__( self , a = True , a = None , a = PILImageResampling.BILINEAR , a = True , a = None , a = True , a = 1 / 255 , a = True , a = None , a = None , **a , ) -> None: super().__init__(**a) SCREAMING_SNAKE_CASE = size if size is not None else {'shortest_edge': 256} SCREAMING_SNAKE_CASE = get_size_dict(a , default_to_square=a) SCREAMING_SNAKE_CASE = crop_size if crop_size is not None else {'height': 224, 'width': 224} SCREAMING_SNAKE_CASE = get_size_dict(a) SCREAMING_SNAKE_CASE = do_resize SCREAMING_SNAKE_CASE = size SCREAMING_SNAKE_CASE = resample SCREAMING_SNAKE_CASE = do_center_crop SCREAMING_SNAKE_CASE = crop_size SCREAMING_SNAKE_CASE = do_rescale SCREAMING_SNAKE_CASE = rescale_factor SCREAMING_SNAKE_CASE = do_normalize SCREAMING_SNAKE_CASE = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN SCREAMING_SNAKE_CASE = image_std if image_std is not None else IMAGENET_STANDARD_STD def SCREAMING_SNAKE_CASE__ ( self , a , a , a = PILImageResampling.BICUBIC , a = None , **a , ) -> np.ndarray: SCREAMING_SNAKE_CASE = get_size_dict(a , default_to_square=a) if "shortest_edge" not in size: raise ValueError(f'''The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}''') SCREAMING_SNAKE_CASE = get_resize_output_image_size(a , size=size['shortest_edge'] , default_to_square=a) return resize(a , size=a , resample=a , data_format=a , **a) def SCREAMING_SNAKE_CASE__ ( self , a , a , a = None , **a , ) -> np.ndarray: SCREAMING_SNAKE_CASE = get_size_dict(a) return center_crop(a , size=(size['height'], size['width']) , data_format=a , **a) def SCREAMING_SNAKE_CASE__ ( self , a , a , a = None , **a) -> np.ndarray: return rescale(a , scale=a , data_format=a , **a) def SCREAMING_SNAKE_CASE__ ( self , a , a , a , a = None , **a , ) -> np.ndarray: return normalize(a , mean=a , std=a , data_format=a , **a) def SCREAMING_SNAKE_CASE__ ( self , a , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = ChannelDimension.FIRST , **a , ) -> Optional[Any]: SCREAMING_SNAKE_CASE = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE = size if size is not None else self.size SCREAMING_SNAKE_CASE = get_size_dict(a , default_to_square=a) SCREAMING_SNAKE_CASE = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE = get_size_dict(a) SCREAMING_SNAKE_CASE = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE = make_list_of_images(a) if not valid_images(a): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.') if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.') if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.') if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.') if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.') # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE = [to_numpy_array(a) for image in images] if do_resize: SCREAMING_SNAKE_CASE = [self.resize(image=a , size=a , resample=a) for image in images] if do_center_crop: SCREAMING_SNAKE_CASE = [self.center_crop(image=a , size=a) for image in images] if do_rescale: SCREAMING_SNAKE_CASE = [self.rescale(image=a , scale=a) for image in images] if do_normalize: SCREAMING_SNAKE_CASE = [self.normalize(image=a , mean=a , std=a) for image in images] SCREAMING_SNAKE_CASE = [to_channel_dimension_format(a , a) for image in images] SCREAMING_SNAKE_CASE = {'pixel_values': images} return BatchFeature(data=a , tensor_type=a)
327
from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_herbert import HerbertTokenizer a_ : List[Any] = logging.get_logger(__name__) a_ : Union[str, Any] = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} a_ : str = { 'vocab_file': { 'allegro/herbert-base-cased': 'https://huggingface.co/allegro/herbert-base-cased/resolve/main/vocab.json' }, 'merges_file': { 'allegro/herbert-base-cased': 'https://huggingface.co/allegro/herbert-base-cased/resolve/main/merges.txt' }, } a_ : List[Any] = {'allegro/herbert-base-cased': 5_14} a_ : Dict = {} class _snake_case ( A__ ): _lowercase : Dict = VOCAB_FILES_NAMES _lowercase : int = PRETRAINED_VOCAB_FILES_MAP _lowercase : Any = PRETRAINED_INIT_CONFIGURATION _lowercase : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowercase : Any = HerbertTokenizer def __init__( self , a=None , a=None , a=None , a="<s>" , a="<unk>" , a="<pad>" , a="<mask>" , a="</s>" , **a , ) -> Dict: super().__init__( a , a , tokenizer_file=a , cls_token=a , unk_token=a , pad_token=a , mask_token=a , sep_token=a , **a , ) def SCREAMING_SNAKE_CASE__ ( self , a , a = None) -> List[int]: SCREAMING_SNAKE_CASE = [self.cls_token_id] SCREAMING_SNAKE_CASE = [self.sep_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE__ ( self , a , a = None , a = False) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a) if token_ids_a is None: return [1] + ([0] * len(a)) + [1] return [1] + ([0] * len(a)) + [1] + ([0] * len(a)) + [1] def SCREAMING_SNAKE_CASE__ ( self , a , a = None) -> List[int]: SCREAMING_SNAKE_CASE = [self.sep_token_id] SCREAMING_SNAKE_CASE = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def SCREAMING_SNAKE_CASE__ ( self , a , a = None) -> Tuple[str]: SCREAMING_SNAKE_CASE = self._tokenizer.model.save(a , name=a) return tuple(a)
327
1
"""simple docstring""" import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class lowercase( __a , unittest.TestCase ): '''simple docstring''' lowercase__ = RoCBertTokenizer lowercase__ = None lowercase__ = False lowercase__ = True lowercase__ = filter_non_english def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' super().setUp() _snake_case : List[str] = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """你""", """好""", """是""", """谁""", """a""", """b""", """c""", """d"""] _snake_case : Tuple = {} _snake_case : Any = {} for i, value in enumerate(a_ ): _snake_case : List[str] = i _snake_case : Optional[int] = i _snake_case : List[Any] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""vocab_file"""] ) _snake_case : Dict = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""word_shape_file"""] ) _snake_case : Optional[Any] = os.path.join(self.tmpdirname, VOCAB_FILES_NAMES["""word_pronunciation_file"""] ) with open(self.vocab_file, """w""", encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) with open(self.word_shape_file, """w""", encoding="""utf-8""" ) as word_shape_writer: json.dump(a_, a_, ensure_ascii=a_ ) with open(self.word_pronunciation_file, """w""", encoding="""utf-8""" ) as word_pronunciation_writer: json.dump(a_, a_, ensure_ascii=a_ ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case : Optional[int] = self.tokenizer_class(self.vocab_file, self.word_shape_file, self.word_pronunciation_file ) _snake_case : str = tokenizer.tokenize("""你好[SEP]你是谁""" ) self.assertListEqual(a_, ["""你""", """好""", """[SEP]""", """你""", """是""", """谁"""] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ), [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_shape_ids(a_ ), [5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_pronunciation_ids(a_ ), [5, 6, 2, 5, 7, 8] ) def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : Any = RoCBertBasicTokenizer() self.assertListEqual(tokenizer.tokenize("""ah\u535A\u63A8zz""" ), ["""ah""", """\u535A""", """\u63A8""", """zz"""] ) def UpperCamelCase_ ( self: Any ): '''simple docstring''' _snake_case : Tuple = RoCBertBasicTokenizer(do_lower_case=a_ ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ), ["""hello""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ), ["""hello"""] ) def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' _snake_case : List[str] = RoCBertBasicTokenizer(do_lower_case=a_, strip_accents=a_ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ), ["""hällo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ), ["""h\u00E9llo"""] ) def UpperCamelCase_ ( self: Union[str, Any] ): '''simple docstring''' _snake_case : Any = RoCBertBasicTokenizer(do_lower_case=a_, strip_accents=a_ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ), ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ), ["""hello"""] ) def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : Optional[Any] = RoCBertBasicTokenizer(do_lower_case=a_ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ), ["""hallo""", """!""", """how""", """are""", """you""", """?"""] ) self.assertListEqual(tokenizer.tokenize("""H\u00E9llo""" ), ["""hello"""] ) def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : Optional[int] = RoCBertBasicTokenizer(do_lower_case=a_ ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? """ ), ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def UpperCamelCase_ ( self: List[Any] ): '''simple docstring''' _snake_case : str = RoCBertBasicTokenizer(do_lower_case=a_, strip_accents=a_ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ), ["""HäLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def UpperCamelCase_ ( self: Any ): '''simple docstring''' _snake_case : Union[str, Any] = RoCBertBasicTokenizer(do_lower_case=a_, strip_accents=a_ ) self.assertListEqual( tokenizer.tokenize(""" \tHäLLo!how \n Are yoU? """ ), ["""HaLLo""", """!""", """how""", """Are""", """yoU""", """?"""] ) def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : str = RoCBertBasicTokenizer(do_lower_case=a_, never_split=["""[UNK]"""] ) self.assertListEqual( tokenizer.tokenize(""" \tHeLLo!how \n Are yoU? [UNK]""" ), ["""HeLLo""", """!""", """how""", """Are""", """yoU""", """?""", """[UNK]"""] ) def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Optional[Any] = ["""[UNK]""", """[CLS]""", """[SEP]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing"""] _snake_case : Optional[Any] = {} for i, token in enumerate(a_ ): _snake_case : Dict = i _snake_case : Optional[Any] = RoCBertWordpieceTokenizer(vocab=a_, unk_token="""[UNK]""" ) self.assertListEqual(tokenizer.tokenize("""""" ), [] ) self.assertListEqual(tokenizer.tokenize("""unwanted running""" ), ["""un""", """##want""", """##ed""", """runn""", """##ing"""] ) self.assertListEqual(tokenizer.tokenize("""unwantedX running""" ), ["""[UNK]""", """runn""", """##ing"""] ) def UpperCamelCase_ ( self: List[str] ): '''simple docstring''' self.assertTrue(_is_whitespace(""" """ ) ) self.assertTrue(_is_whitespace("""\t""" ) ) self.assertTrue(_is_whitespace("""\r""" ) ) self.assertTrue(_is_whitespace("""\n""" ) ) self.assertTrue(_is_whitespace("""\u00A0""" ) ) self.assertFalse(_is_whitespace("""A""" ) ) self.assertFalse(_is_whitespace("""-""" ) ) def UpperCamelCase_ ( self: Optional[int] ): '''simple docstring''' self.assertTrue(_is_control("""\u0005""" ) ) self.assertFalse(_is_control("""A""" ) ) self.assertFalse(_is_control(""" """ ) ) self.assertFalse(_is_control("""\t""" ) ) self.assertFalse(_is_control("""\r""" ) ) def UpperCamelCase_ ( self: Optional[Any] ): '''simple docstring''' self.assertTrue(_is_punctuation("""-""" ) ) self.assertTrue(_is_punctuation("""$""" ) ) self.assertTrue(_is_punctuation("""`""" ) ) self.assertTrue(_is_punctuation(""".""" ) ) self.assertFalse(_is_punctuation("""A""" ) ) self.assertFalse(_is_punctuation(""" """ ) ) def UpperCamelCase_ ( self: Dict ): '''simple docstring''' _snake_case : Optional[int] = self.get_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(a_ ) for t in ["""Test""", """\xad""", """test"""]], [["""[UNK]"""], [], ["""[UNK]"""]] ) if self.test_rust_tokenizer: _snake_case : Optional[int] = self.get_rust_tokenizer() self.assertListEqual( [rust_tokenizer.tokenize(a_ ) for t in ["""Test""", """\xad""", """test"""]], [["""[UNK]"""], [], ["""[UNK]"""]] ) def UpperCamelCase_ ( self: Any ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): _snake_case : Optional[int] = self.rust_tokenizer_class.from_pretrained(a_, **a_ ) _snake_case : List[Any] = f"A, naïve {tokenizer_r.mask_token} AllenNLP sentence." _snake_case : List[Any] = tokenizer_r.encode_plus( a_, return_attention_mask=a_, return_token_type_ids=a_, return_offsets_mapping=a_, add_special_tokens=a_, ) _snake_case : Optional[Any] = tokenizer_r.do_lower_case if hasattr(a_, """do_lower_case""" ) else False _snake_case : Optional[Any] = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """A"""), ((1, 2), ""","""), ((3, 5), """na"""), ((5, 6), """##ï"""), ((6, 8), """##ve"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """Allen"""), ((21, 23), """##NL"""), ((23, 24), """##P"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), """a"""), ((1, 2), ""","""), ((3, 8), """naive"""), ((9, 15), tokenizer_r.mask_token), ((16, 21), """allen"""), ((21, 23), """##nl"""), ((23, 24), """##p"""), ((25, 33), """sentence"""), ((33, 34), """."""), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results], tokenizer_r.convert_ids_to_tokens(tokens["""input_ids"""] ) ) self.assertEqual([e[0] for e in expected_results], tokens["""offset_mapping"""] ) def UpperCamelCase_ ( self: int ): '''simple docstring''' _snake_case : Union[str, Any] = ["""的""", """人""", """有"""] _snake_case : Any = """""".join(a_ ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"{tokenizer.__class__.__name__} ({pretrained_name})" ): _snake_case : int = True _snake_case : Tuple = self.tokenizer_class.from_pretrained(a_, **a_ ) _snake_case : List[Any] = self.rust_tokenizer_class.from_pretrained(a_, **a_ ) _snake_case : Optional[Any] = tokenizer_p.encode(a_, add_special_tokens=a_ ) _snake_case : int = tokenizer_r.encode(a_, add_special_tokens=a_ ) _snake_case : Optional[Any] = tokenizer_r.convert_ids_to_tokens(a_ ) _snake_case : Optional[int] = tokenizer_p.convert_ids_to_tokens(a_ ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(a_, a_ ) self.assertListEqual(a_, a_ ) _snake_case : List[str] = False _snake_case : Union[str, Any] = self.rust_tokenizer_class.from_pretrained(a_, **a_ ) _snake_case : str = self.tokenizer_class.from_pretrained(a_, **a_ ) _snake_case : Optional[Any] = tokenizer_r.encode(a_, add_special_tokens=a_ ) _snake_case : Any = tokenizer_p.encode(a_, add_special_tokens=a_ ) _snake_case : Optional[int] = tokenizer_r.convert_ids_to_tokens(a_ ) _snake_case : Tuple = tokenizer_p.convert_ids_to_tokens(a_ ) # it is expected that only the first Chinese character is not preceded by "##". _snake_case : List[Any] = [ f"##{token}" if idx != 0 else token for idx, token in enumerate(a_ ) ] self.assertListEqual(a_, a_ ) self.assertListEqual(a_, a_ ) @slow def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Union[str, Any] = self.tokenizer_class(self.vocab_file, self.word_shape_file, self.word_pronunciation_file ) _snake_case : Tuple = tokenizer.encode("""你好""", add_special_tokens=a_ ) _snake_case : Optional[int] = tokenizer.encode("""你是谁""", add_special_tokens=a_ ) _snake_case : str = tokenizer.build_inputs_with_special_tokens(a_ ) _snake_case : int = tokenizer.build_inputs_with_special_tokens(a_, a_ ) assert encoded_sentence == [1] + text + [2] assert encoded_pair == [1] + text + [2] + text_a + [2] def UpperCamelCase_ ( self: str ): '''simple docstring''' _snake_case : List[Any] = self.get_tokenizers(do_lower_case=a_ ) for tokenizer in tokenizers: with self.subTest(f"{tokenizer.__class__.__name__}" ): _snake_case : int = """你好,你是谁""" _snake_case : List[str] = tokenizer.tokenize(a_ ) _snake_case : List[str] = tokenizer.convert_tokens_to_ids(a_ ) _snake_case : List[Any] = tokenizer.convert_tokens_to_shape_ids(a_ ) _snake_case : Optional[Any] = tokenizer.convert_tokens_to_pronunciation_ids(a_ ) _snake_case : Dict = tokenizer.prepare_for_model( a_, a_, a_, add_special_tokens=a_ ) _snake_case : Dict = tokenizer.encode_plus(a_, add_special_tokens=a_ ) self.assertEqual(a_, a_ )
64
'''simple docstring''' import unittest from transformers import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING, is_vision_available, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class _UpperCamelCase : '''simple docstring''' @staticmethod def UpperCamelCase__ ( *lowerCAmelCase__ : Any , **lowerCAmelCase__ : Any ): """simple docstring""" pass @is_pipeline_test @require_vision @require_torch class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' _A : Optional[int] = MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING def UpperCamelCase__ ( self : str , lowerCAmelCase__ : Dict , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Optional[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = pipeline( """zero-shot-object-detection""" , model="""hf-internal-testing/tiny-random-owlvit-object-detection""" ) __SCREAMING_SNAKE_CASE : Any = [ { """image""": """./tests/fixtures/tests_samples/COCO/000000039769.png""", """candidate_labels""": ["""cat""", """remote""", """couch"""], } ] return object_detector, examples def UpperCamelCase__ ( self : int , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : List[Any] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Optional[int] = object_detector(examples[0] , threshold=0.0 ) __SCREAMING_SNAKE_CASE : Optional[Any] = len(lowerCAmelCase__ ) self.assertGreater(lowerCAmelCase__ , 0 ) self.assertEqual( lowerCAmelCase__ , [ { """score""": ANY(lowerCAmelCase__ ), """label""": ANY(lowerCAmelCase__ ), """box""": {"""xmin""": ANY(lowerCAmelCase__ ), """ymin""": ANY(lowerCAmelCase__ ), """xmax""": ANY(lowerCAmelCase__ ), """ymax""": ANY(lowerCAmelCase__ )}, } for i in range(lowerCAmelCase__ ) ] , ) @require_tf @unittest.skip("""Zero Shot Object Detection not implemented in TF""" ) def UpperCamelCase__ ( self : Optional[int] ): """simple docstring""" pass @require_torch def UpperCamelCase__ ( self : Optional[int] ): """simple docstring""" __SCREAMING_SNAKE_CASE : int = pipeline( """zero-shot-object-detection""" , model="""hf-internal-testing/tiny-random-owlvit-object-detection""" ) __SCREAMING_SNAKE_CASE : int = object_detector( """./tests/fixtures/tests_samples/COCO/000000039769.png""" , candidate_labels=["""cat""", """remote""", """couch"""] , threshold=0.64 , ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"""score""": 0.72_35, """label""": """cat""", """box""": {"""xmin""": 2_0_4, """ymin""": 1_6_7, """xmax""": 2_3_2, """ymax""": 1_9_0}}, {"""score""": 0.72_18, """label""": """remote""", """box""": {"""xmin""": 2_0_4, """ymin""": 1_6_7, """xmax""": 2_3_2, """ymax""": 1_9_0}}, {"""score""": 0.71_84, """label""": """couch""", """box""": {"""xmin""": 2_0_4, """ymin""": 1_6_7, """xmax""": 2_3_2, """ymax""": 1_9_0}}, {"""score""": 0.67_48, """label""": """remote""", """box""": {"""xmin""": 5_7_1, """ymin""": 8_3, """xmax""": 5_9_8, """ymax""": 1_0_3}}, {"""score""": 0.66_56, """label""": """cat""", """box""": {"""xmin""": 5_7_1, """ymin""": 8_3, """xmax""": 5_9_8, """ymax""": 1_0_3}}, {"""score""": 0.66_14, """label""": """couch""", """box""": {"""xmin""": 5_7_1, """ymin""": 8_3, """xmax""": 5_9_8, """ymax""": 1_0_3}}, {"""score""": 0.64_56, """label""": """remote""", """box""": {"""xmin""": 4_9_4, """ymin""": 1_0_5, """xmax""": 5_2_1, """ymax""": 1_2_7}}, {"""score""": 0.6_42, """label""": """remote""", """box""": {"""xmin""": 6_7, """ymin""": 2_7_4, """xmax""": 9_3, """ymax""": 2_9_7}}, {"""score""": 0.64_19, """label""": """cat""", """box""": {"""xmin""": 4_9_4, """ymin""": 1_0_5, """xmax""": 5_2_1, """ymax""": 1_2_7}}, ] , ) __SCREAMING_SNAKE_CASE : List[Any] = object_detector( [ { """image""": """./tests/fixtures/tests_samples/COCO/000000039769.png""", """candidate_labels""": ["""cat""", """remote""", """couch"""], } ] , threshold=0.64 , ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {"""score""": 0.72_35, """label""": """cat""", """box""": {"""xmin""": 2_0_4, """ymin""": 1_6_7, """xmax""": 2_3_2, """ymax""": 1_9_0}}, {"""score""": 0.72_18, """label""": """remote""", """box""": {"""xmin""": 2_0_4, """ymin""": 1_6_7, """xmax""": 2_3_2, """ymax""": 1_9_0}}, {"""score""": 0.71_84, """label""": """couch""", """box""": {"""xmin""": 2_0_4, """ymin""": 1_6_7, """xmax""": 2_3_2, """ymax""": 1_9_0}}, {"""score""": 0.67_48, """label""": """remote""", """box""": {"""xmin""": 5_7_1, """ymin""": 8_3, """xmax""": 5_9_8, """ymax""": 1_0_3}}, {"""score""": 0.66_56, """label""": """cat""", """box""": {"""xmin""": 5_7_1, """ymin""": 8_3, """xmax""": 5_9_8, """ymax""": 1_0_3}}, {"""score""": 0.66_14, """label""": """couch""", """box""": {"""xmin""": 5_7_1, """ymin""": 8_3, """xmax""": 5_9_8, """ymax""": 1_0_3}}, {"""score""": 0.64_56, """label""": """remote""", """box""": {"""xmin""": 4_9_4, """ymin""": 1_0_5, """xmax""": 5_2_1, """ymax""": 1_2_7}}, {"""score""": 0.6_42, """label""": """remote""", """box""": {"""xmin""": 6_7, """ymin""": 2_7_4, """xmax""": 9_3, """ymax""": 2_9_7}}, {"""score""": 0.64_19, """label""": """cat""", """box""": {"""xmin""": 4_9_4, """ymin""": 1_0_5, """xmax""": 5_2_1, """ymax""": 1_2_7}}, ] ] , ) @require_torch @slow def UpperCamelCase__ ( self : str ): """simple docstring""" __SCREAMING_SNAKE_CASE : Any = pipeline("""zero-shot-object-detection""" ) __SCREAMING_SNAKE_CASE : List[str] = object_detector( """http://images.cocodataset.org/val2017/000000039769.jpg""" , candidate_labels=["""cat""", """remote""", """couch"""] , ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"""score""": 0.28_68, """label""": """cat""", """box""": {"""xmin""": 3_2_4, """ymin""": 2_0, """xmax""": 6_4_0, """ymax""": 3_7_3}}, {"""score""": 0.2_77, """label""": """remote""", """box""": {"""xmin""": 4_0, """ymin""": 7_2, """xmax""": 1_7_7, """ymax""": 1_1_5}}, {"""score""": 0.25_37, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 5_5, """xmax""": 3_1_5, """ymax""": 4_7_2}}, {"""score""": 0.14_74, """label""": """remote""", """box""": {"""xmin""": 3_3_5, """ymin""": 7_4, """xmax""": 3_7_1, """ymax""": 1_8_7}}, {"""score""": 0.12_08, """label""": """couch""", """box""": {"""xmin""": 4, """ymin""": 0, """xmax""": 6_4_2, """ymax""": 4_7_6}}, ] , ) __SCREAMING_SNAKE_CASE : Dict = object_detector( [ { """image""": """http://images.cocodataset.org/val2017/000000039769.jpg""", """candidate_labels""": ["""cat""", """remote""", """couch"""], }, { """image""": """http://images.cocodataset.org/val2017/000000039769.jpg""", """candidate_labels""": ["""cat""", """remote""", """couch"""], }, ] , ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ [ {"""score""": 0.28_68, """label""": """cat""", """box""": {"""xmin""": 3_2_4, """ymin""": 2_0, """xmax""": 6_4_0, """ymax""": 3_7_3}}, {"""score""": 0.2_77, """label""": """remote""", """box""": {"""xmin""": 4_0, """ymin""": 7_2, """xmax""": 1_7_7, """ymax""": 1_1_5}}, {"""score""": 0.25_37, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 5_5, """xmax""": 3_1_5, """ymax""": 4_7_2}}, {"""score""": 0.14_74, """label""": """remote""", """box""": {"""xmin""": 3_3_5, """ymin""": 7_4, """xmax""": 3_7_1, """ymax""": 1_8_7}}, {"""score""": 0.12_08, """label""": """couch""", """box""": {"""xmin""": 4, """ymin""": 0, """xmax""": 6_4_2, """ymax""": 4_7_6}}, ], [ {"""score""": 0.28_68, """label""": """cat""", """box""": {"""xmin""": 3_2_4, """ymin""": 2_0, """xmax""": 6_4_0, """ymax""": 3_7_3}}, {"""score""": 0.2_77, """label""": """remote""", """box""": {"""xmin""": 4_0, """ymin""": 7_2, """xmax""": 1_7_7, """ymax""": 1_1_5}}, {"""score""": 0.25_37, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 5_5, """xmax""": 3_1_5, """ymax""": 4_7_2}}, {"""score""": 0.14_74, """label""": """remote""", """box""": {"""xmin""": 3_3_5, """ymin""": 7_4, """xmax""": 3_7_1, """ymax""": 1_8_7}}, {"""score""": 0.12_08, """label""": """couch""", """box""": {"""xmin""": 4, """ymin""": 0, """xmax""": 6_4_2, """ymax""": 4_7_6}}, ], ] , ) @require_tf @unittest.skip("""Zero Shot Object Detection not implemented in TF""" ) def UpperCamelCase__ ( self : Any ): """simple docstring""" pass @require_torch @slow def UpperCamelCase__ ( self : List[str] ): """simple docstring""" __SCREAMING_SNAKE_CASE : Dict = 0.2 __SCREAMING_SNAKE_CASE : Dict = pipeline("""zero-shot-object-detection""" ) __SCREAMING_SNAKE_CASE : Any = object_detector( """http://images.cocodataset.org/val2017/000000039769.jpg""" , candidate_labels=["""cat""", """remote""", """couch"""] , threshold=lowerCAmelCase__ , ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"""score""": 0.28_68, """label""": """cat""", """box""": {"""xmin""": 3_2_4, """ymin""": 2_0, """xmax""": 6_4_0, """ymax""": 3_7_3}}, {"""score""": 0.2_77, """label""": """remote""", """box""": {"""xmin""": 4_0, """ymin""": 7_2, """xmax""": 1_7_7, """ymax""": 1_1_5}}, {"""score""": 0.25_37, """label""": """cat""", """box""": {"""xmin""": 1, """ymin""": 5_5, """xmax""": 3_1_5, """ymax""": 4_7_2}}, ] , ) @require_torch @slow def UpperCamelCase__ ( self : int ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = 2 __SCREAMING_SNAKE_CASE : int = pipeline("""zero-shot-object-detection""" ) __SCREAMING_SNAKE_CASE : int = object_detector( """http://images.cocodataset.org/val2017/000000039769.jpg""" , candidate_labels=["""cat""", """remote""", """couch"""] , top_k=lowerCAmelCase__ , ) self.assertEqual( nested_simplify(lowerCAmelCase__ , decimals=4 ) , [ {"""score""": 0.28_68, """label""": """cat""", """box""": {"""xmin""": 3_2_4, """ymin""": 2_0, """xmax""": 6_4_0, """ymax""": 3_7_3}}, {"""score""": 0.2_77, """label""": """remote""", """box""": {"""xmin""": 4_0, """ymin""": 7_2, """xmax""": 1_7_7, """ymax""": 1_1_5}}, ] , )
112
0
from __future__ import annotations def lowerCAmelCase_ ( snake_case_ ): if len(snake_case_ ) == 0: return array _A , _A : Optional[Any] = min(snake_case_ ), max(snake_case_ ) # Compute the variables _A : str = _max - _min + 1 _A , _A : List[str] = [0] * holes_range, [0] * holes_range # Make the sorting. for i in array: _A : str = i - _min _A : int = i holes_repeat[index] += 1 # Makes the array back by replacing the numbers. _A : List[str] = 0 for i in range(snake_case_ ): while holes_repeat[i] > 0: _A : str = holes[i] index += 1 holes_repeat[i] -= 1 # Returns the sorted array. return array if __name__ == "__main__": import doctest doctest.testmod() _snake_case = input("Enter numbers separated by comma:\n") _snake_case = [int(x) for x in user_input.split(",")] print(pigeon_sort(unsorted))
343
# DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class lowercase : _a = 42 # setable values _a = 42 _a = 42 _a = None @classmethod def a__ ( cls , _a , _a , _a ) -> Tuple: return cls(common=_a , init_noise_sigma=_a , timesteps=_a ) @dataclass class lowercase ( UpperCamelCase__ ): _a = 42 class lowercase ( UpperCamelCase__,UpperCamelCase__ ): _a = [e.name for e in FlaxKarrasDiffusionSchedulers] _a = 42 @property def a__ ( self ) -> Dict: return True @register_to_config def __init__( self , _a = 1000 , _a = 0.0001 , _a = 0.02 , _a = "linear" , _a = None , _a = "fixed_small" , _a = True , _a = "epsilon" , _a = jnp.floataa , ) -> Tuple: _A : Tuple = dtype def a__ ( self , _a = None ) -> DDPMSchedulerState: if common is None: _A : Dict = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution _A : Union[str, Any] = jnp.array(1.0 , dtype=self.dtype ) _A : Tuple = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=_a , init_noise_sigma=_a , timesteps=_a , ) def a__ ( self , _a , _a , _a = None ) -> jnp.ndarray: return sample def a__ ( self , _a , _a , _a = () ) -> DDPMSchedulerState: _A : Any = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 _A : Dict = (jnp.arange(0 , _a ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=_a , timesteps=_a , ) def a__ ( self , _a , _a , _a=None , _a=None ) -> Optional[int]: _A : Optional[Any] = state.common.alphas_cumprod[t] _A : int = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample _A : List[str] = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: _A : Optional[Any] = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": _A : Optional[Any] = jnp.clip(_a , a_min=1e-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": _A : Any = jnp.log(jnp.clip(_a , a_min=1e-20 ) ) elif variance_type == "fixed_large": _A : Optional[Any] = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log _A : Tuple = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": _A : str = variance _A : Union[str, Any] = state.common.betas[t] _A : Tuple = (predicted_variance + 1) / 2 _A : List[str] = frac * max_log + (1 - frac) * min_log return variance def a__ ( self , _a , _a , _a , _a , _a = None , _a = True , ) -> Union[FlaxDDPMSchedulerOutput, Tuple]: _A : Dict = timestep if key is None: _A : int = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: _A , _A : List[str] = jnp.split(_a , sample.shape[1] , axis=1 ) else: _A : int = None # 1. compute alphas, betas _A : int = state.common.alphas_cumprod[t] _A : List[str] = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) _A : Union[str, Any] = 1 - alpha_prod_t _A : Optional[int] = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": _A : Dict = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": _A : Optional[int] = model_output elif self.config.prediction_type == "v_prediction": _A : Any = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( F'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` ''' """ for the FlaxDDPMScheduler.""" ) # 3. Clip "predicted x_0" if self.config.clip_sample: _A : Union[str, Any] = jnp.clip(_a , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _A : List[Any] = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t _A : Dict = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _A : int = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): _A : Tuple = jax.random.split(_a , num=1 ) _A : Dict = jax.random.normal(_a , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(_a , _a , predicted_variance=_a ) ** 0.5) * noise _A : int = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) _A : Union[str, Any] = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=_a , state=_a ) def a__ ( self , _a , _a , _a , _a , ) -> jnp.ndarray: return add_noise_common(state.common , _a , _a , _a ) def a__ ( self , _a , _a , _a , _a , ) -> jnp.ndarray: return get_velocity_common(state.common , _a , _a , _a ) def __len__( self ) -> List[Any]: return self.config.num_train_timesteps
343
1
"""simple docstring""" from queue import Queue from typing import TYPE_CHECKING, Optional if TYPE_CHECKING: from ..models.auto import AutoTokenizer class _A : def A__ ( self , __lowerCAmelCase ): """simple docstring""" raise NotImplementedError() def A__ ( self ): """simple docstring""" raise NotImplementedError() class _A ( lowerCAmelCase ): def __init__( self , __lowerCAmelCase , __lowerCAmelCase = False , **__lowerCAmelCase ): """simple docstring""" lowercase = tokenizer lowercase = skip_prompt lowercase = decode_kwargs # variables used in the streaming process lowercase = [] lowercase = 0 lowercase = True def A__ ( self , __lowerCAmelCase ): """simple docstring""" if len(value.shape ) > 1 and value.shape[0] > 1: raise ValueError("""TextStreamer only supports batch size 1""" ) elif len(value.shape ) > 1: lowercase = value[0] if self.skip_prompt and self.next_tokens_are_prompt: lowercase = False return # Add the new token to the cache and decodes the entire thing. self.token_cache.extend(value.tolist() ) lowercase = self.tokenizer.decode(self.token_cache , **self.decode_kwargs ) # After the symbol for a new line, we flush the cache. if text.endswith("""\n""" ): lowercase = text[self.print_len :] lowercase = [] lowercase = 0 # If the last token is a CJK character, we print the characters. elif len(__lowerCAmelCase ) > 0 and self._is_chinese_char(ord(text[-1] ) ): lowercase = text[self.print_len :] self.print_len += len(__lowerCAmelCase ) # Otherwise, prints until the last space char (simple heuristic to avoid printing incomplete words, # which may change with the subsequent token -- there are probably smarter ways to do this!) else: lowercase = text[self.print_len : text.rfind(""" """ ) + 1] self.print_len += len(__lowerCAmelCase ) self.on_finalized_text(__lowerCAmelCase ) def A__ ( self ): """simple docstring""" if len(self.token_cache ) > 0: lowercase = self.tokenizer.decode(self.token_cache , **self.decode_kwargs ) lowercase = text[self.print_len :] lowercase = [] lowercase = 0 else: lowercase = """""" lowercase = True self.on_finalized_text(__lowerCAmelCase , stream_end=__lowerCAmelCase ) def A__ ( self , __lowerCAmelCase , __lowerCAmelCase = False ): """simple docstring""" print(__lowerCAmelCase , flush=__lowerCAmelCase , end="""""" if not stream_end else None ) def A__ ( self , __lowerCAmelCase ): """simple docstring""" if ( (cp >= 0X4_e00 and cp <= 0X9_fff) or (cp >= 0X3_400 and cp <= 0X4_dbf) # or (cp >= 0X20_000 and cp <= 0X2a_6df) # or (cp >= 0X2a_700 and cp <= 0X2b_73f) # or (cp >= 0X2b_740 and cp <= 0X2b_81f) # or (cp >= 0X2b_820 and cp <= 0X2c_eaf) # or (cp >= 0Xf_900 and cp <= 0Xf_aff) or (cp >= 0X2f_800 and cp <= 0X2f_a1f) # ): # return True return False class _A ( lowerCAmelCase ): def __init__( self , __lowerCAmelCase , __lowerCAmelCase = False , __lowerCAmelCase = None , **__lowerCAmelCase ): """simple docstring""" super().__init__(__lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ) lowercase = Queue() lowercase = None lowercase = timeout def A__ ( self , __lowerCAmelCase , __lowerCAmelCase = False ): """simple docstring""" self.text_queue.put(__lowerCAmelCase , timeout=self.timeout ) if stream_end: self.text_queue.put(self.stop_signal , timeout=self.timeout ) def __iter__( self ): """simple docstring""" return self def A__ ( self ): """simple docstring""" lowercase = self.text_queue.get(timeout=self.timeout ) if value == self.stop_signal: raise StopIteration() else: return value
197
"""simple docstring""" def UpperCAmelCase__ ( lowerCAmelCase__ :str , lowerCAmelCase__ :Optional[Any] ) -> int: '''simple docstring''' lowercase = [0 for i in range(r + 1 )] # nc0 = 1 lowercase = 1 for i in range(1 , n + 1 ): # to compute current row from previous row. lowercase = min(lowerCAmelCase__ , lowerCAmelCase__ ) while j > 0: c[j] += c[j - 1] j -= 1 return c[r] print(binomial_coefficient(n=1_0, r=5))
197
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) SCREAMING_SNAKE_CASE__ = { """configuration_blip""": [ """BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BlipConfig""", """BlipTextConfig""", """BlipVisionConfig""", ], """processing_blip""": ["""BlipProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = ["""BlipImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = [ """BLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """BlipModel""", """BlipPreTrainedModel""", """BlipForConditionalGeneration""", """BlipForQuestionAnswering""", """BlipVisionModel""", """BlipTextModel""", """BlipForImageTextRetrieval""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = [ """TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFBlipModel""", """TFBlipPreTrainedModel""", """TFBlipForConditionalGeneration""", """TFBlipForQuestionAnswering""", """TFBlipVisionModel""", """TFBlipTextModel""", """TFBlipForImageTextRetrieval""", ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys SCREAMING_SNAKE_CASE__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
352
from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class __lowerCamelCase ( snake_case_ ): """simple docstring""" def A__ ( self , UpperCAmelCase ) -> float: '''simple docstring''' return 0.0 def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: np.ndarray , __lowerCamelCase: int ): '''simple docstring''' lowercase_ = min([-20, np.min(fft_results[1 : samplerate // 2 - 1] )] ) lowercase_ = max([20, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: FilterType , __lowerCamelCase: int ): '''simple docstring''' lowercase_ = 512 lowercase_ = [1] + [0] * (size - 1) lowercase_ = [filter_type.process(__lowerCamelCase ) for item in inputs] lowercase_ = [0] * (samplerate - size) # zero-padding outputs += filler lowercase_ = np.abs(np.fft.fft(__lowerCamelCase ) ) lowercase_ = 20 * np.logaa(__lowerCamelCase ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel("Frequency (Hz)" ) plt.xscale("log" ) # Display within reasonable bounds lowercase_ = get_bounds(__lowerCamelCase , __lowerCamelCase ) plt.ylim(max([-80, bounds[0]] ) , min([80, bounds[1]] ) ) plt.ylabel("Gain (dB)" ) plt.plot(__lowerCamelCase ) plt.show() def SCREAMING_SNAKE_CASE_ ( __lowerCamelCase: FilterType , __lowerCamelCase: int ): '''simple docstring''' lowercase_ = 512 lowercase_ = [1] + [0] * (size - 1) lowercase_ = [filter_type.process(__lowerCamelCase ) for item in inputs] lowercase_ = [0] * (samplerate - size) # zero-padding outputs += filler lowercase_ = np.angle(np.fft.fft(__lowerCamelCase ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel("Frequency (Hz)" ) plt.xscale("log" ) plt.ylim(-2 * pi , 2 * pi ) plt.ylabel("Phase shift (Radians)" ) plt.plot(np.unwrap(__lowerCamelCase , -2 * pi ) ) plt.show()
297
0
"""simple docstring""" import argparse import json import re from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileNetVaConfig, MobileNetVaForImageClassification, MobileNetVaImageProcessor, load_tf_weights_in_mobilenet_va, ) from transformers.utils import logging logging.set_verbosity_info() __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) def _a ( _SCREAMING_SNAKE_CASE ) -> Optional[int]: snake_case_ = MobileNetVaConfig(layer_norm_eps=0.001 ) if "_quant" in model_name: raise ValueError("""Quantized models are not supported.""" ) snake_case_ = re.match(r"""^mobilenet_v1_([^_]*)_([^_]*)$""" , _SCREAMING_SNAKE_CASE ) if matches: snake_case_ = float(matches[1] ) snake_case_ = int(matches[2] ) # The TensorFlow version of MobileNetV1 predicts 1001 classes instead of # the usual 1000. The first class (index 0) is "background". snake_case_ = 1_001 snake_case_ = """imagenet-1k-id2label.json""" snake_case_ = """huggingface/label-files""" snake_case_ = json.load(open(hf_hub_download(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , repo_type="""dataset""" ) , """r""" ) ) snake_case_ = {int(_SCREAMING_SNAKE_CASE ) + 1: v for k, v in idalabel.items()} snake_case_ = """background""" snake_case_ = idalabel snake_case_ = {v: k for k, v in idalabel.items()} return config def _a ( ) -> Dict: snake_case_ = """http://images.cocodataset.org/val2017/000000039769.jpg""" snake_case_ = Image.open(requests.get(_SCREAMING_SNAKE_CASE , stream=_SCREAMING_SNAKE_CASE ).raw ) return im @torch.no_grad() def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=False ) -> Optional[Any]: snake_case_ = get_mobilenet_va_config(_SCREAMING_SNAKE_CASE ) # Load 🤗 model snake_case_ = MobileNetVaForImageClassification(_SCREAMING_SNAKE_CASE ).eval() # Load weights from TensorFlow checkpoint load_tf_weights_in_mobilenet_va(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # Check outputs on an image, prepared by MobileNetV1ImageProcessor snake_case_ = MobileNetVaImageProcessor( crop_size={"""width""": config.image_size, """height""": config.image_size} , size={"""shortest_edge""": config.image_size + 32} , ) snake_case_ = image_processor(images=prepare_img() , return_tensors="""pt""" ) snake_case_ = model(**_SCREAMING_SNAKE_CASE ) snake_case_ = outputs.logits assert logits.shape == (1, 1_001) if model_name == "mobilenet_v1_1.0_224": snake_case_ = torch.tensor([-4.1739, -1.1233, 3.1205] ) elif model_name == "mobilenet_v1_0.75_192": snake_case_ = torch.tensor([-3.9440, -2.3141, -0.3333] ) else: snake_case_ = None if expected_logits is not None: assert torch.allclose(logits[0, :3] , _SCREAMING_SNAKE_CASE , atol=1E-4 ) Path(_SCREAMING_SNAKE_CASE ).mkdir(exist_ok=_SCREAMING_SNAKE_CASE ) print(f"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(_SCREAMING_SNAKE_CASE ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(_SCREAMING_SNAKE_CASE ) if push_to_hub: print("""Pushing to the hub...""" ) snake_case_ = """google/""" + model_name image_processor.push_to_hub(_SCREAMING_SNAKE_CASE ) model.push_to_hub(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __SCREAMING_SNAKE_CASE : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='mobilenet_v1_1.0_224', type=str, help='Name of the MobileNetV1 model you\'d like to convert. Should in the form \'mobilenet_v1_<depth>_<size>\'.', ) parser.add_argument( '--checkpoint_path', required=True, type=str, help='Path to the original TensorFlow checkpoint (.ckpt file).' ) parser.add_argument( '--pytorch_dump_folder_path', required=True, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) __SCREAMING_SNAKE_CASE : Dict = parser.parse_args() convert_movilevit_checkpoint( args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
347
"""simple docstring""" import datasets __SCREAMING_SNAKE_CASE : Tuple = '\\n@InProceedings{conneau2018xnli,\n author = "Conneau, Alexis\n and Rinott, Ruty\n and Lample, Guillaume\n and Williams, Adina\n and Bowman, Samuel R.\n and Schwenk, Holger\n and Stoyanov, Veselin",\n title = "XNLI: Evaluating Cross-lingual Sentence Representations",\n booktitle = "Proceedings of the 2018 Conference on Empirical Methods\n in Natural Language Processing",\n year = "2018",\n publisher = "Association for Computational Linguistics",\n location = "Brussels, Belgium",\n}\n' __SCREAMING_SNAKE_CASE : Dict = '\\nXNLI is a subset of a few thousand examples from MNLI which has been translated\ninto a 14 different languages (some low-ish resource). As with MNLI, the goal is\nto predict textual entailment (does sentence A imply/contradict/neither sentence\nB) and is a classification task (given two sentences, predict one of three\nlabels).\n' __SCREAMING_SNAKE_CASE : List[str] = '\nComputes XNLI score which is just simple accuracy.\nArgs:\n predictions: Predicted labels.\n references: Ground truth labels.\nReturns:\n \'accuracy\': accuracy\nExamples:\n\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> xnli_metric = datasets.load_metric("xnli")\n >>> results = xnli_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n' def _a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) -> List[Any]: return (preds == labels).mean() @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class __A (datasets.Metric): '''simple docstring''' def lowerCAmelCase ( self : str ) ->Any: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""int64""" if self.config_name != """sts-b""" else """float32""" ), """references""": datasets.Value("""int64""" if self.config_name != """sts-b""" else """float32""" ), } ) , codebase_urls=[] , reference_urls=[] , format="""numpy""" , ) def lowerCAmelCase ( self : Dict , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Any ) ->int: """simple docstring""" return {"accuracy": simple_accuracy(UpperCAmelCase_ , UpperCAmelCase_ )}
347
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE__ = { 'configuration_nllb_moe': [ 'NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP', 'NllbMoeConfig', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE__ = [ 'NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST', 'NllbMoeForConditionalGeneration', 'NllbMoeModel', 'NllbMoePreTrainedModel', 'NllbMoeTop2Router', 'NllbMoeSparseMLP', ] if TYPE_CHECKING: from .configuration_nllb_moe import ( NLLB_MOE_PRETRAINED_CONFIG_ARCHIVE_MAP, NllbMoeConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nllb_moe import ( NLLB_MOE_PRETRAINED_MODEL_ARCHIVE_LIST, NllbMoeForConditionalGeneration, NllbMoeModel, NllbMoePreTrainedModel, NllbMoeSparseMLP, NllbMoeTopaRouter, ) else: import sys SCREAMING_SNAKE_CASE__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
183
'''simple docstring''' import importlib.util import json import os import warnings from dataclasses import dataclass, field import torch from ..training_args import TrainingArguments from ..utils import cached_property, is_sagemaker_dp_enabled, logging SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) def lowercase__ ( )-> Tuple: # Get the sagemaker specific mp parameters from smp_options variable. UpperCamelCase = os.getenv("""SM_HP_MP_PARAMETERS""" , """{}""" ) try: # Parse it and check the field "partitions" is included, it is required for model parallel. UpperCamelCase = json.loads(__UpperCamelCase ) if "partitions" not in smp_options: return False except json.JSONDecodeError: return False # Get the sagemaker specific framework parameters from mpi_options variable. UpperCamelCase = os.getenv("""SM_FRAMEWORK_PARAMS""" , """{}""" ) try: # Parse it and check the field "sagemaker_distributed_dataparallel_enabled". UpperCamelCase = json.loads(__UpperCamelCase ) if not mpi_options.get("""sagemaker_mpi_enabled""" , __UpperCamelCase ): return False except json.JSONDecodeError: return False # Lastly, check if the `smdistributed` module is present. return importlib.util.find_spec("""smdistributed""" ) is not None if is_sagemaker_model_parallel_available(): import smdistributed.modelparallel.torch as smp smp.init() @dataclass class a_ ( lowerCamelCase ): lowercase = field( default="""""" , metadata={"""help""": """Used by the SageMaker launcher to send mp-specific args. Ignored in SageMakerTrainer"""} , ) def A__ ( self ) -> Tuple: """simple docstring""" super().__post_init__() warnings.warn( """`SageMakerTrainingArguments` is deprecated and will be removed in v5 of Transformers. You can use """ """`TrainingArguments` instead.""" , _SCREAMING_SNAKE_CASE , ) @cached_property def A__ ( self ) -> "torch.device": """simple docstring""" logger.info("""PyTorch: setting up devices""" ) if torch.distributed.is_available() and torch.distributed.is_initialized() and self.local_rank == -1: logger.warning( """torch.distributed process group is initialized, but local_rank == -1. """ """In order to use Torch DDP, launch your script with `python -m torch.distributed.launch""" ) if self.no_cuda: UpperCamelCase = torch.device("""cpu""" ) UpperCamelCase = 0 elif is_sagemaker_model_parallel_available(): UpperCamelCase = smp.local_rank() UpperCamelCase = torch.device("""cuda""" , _SCREAMING_SNAKE_CASE ) UpperCamelCase = 1 elif is_sagemaker_dp_enabled(): import smdistributed.dataparallel.torch.torch_smddp # noqa: F401 torch.distributed.init_process_group(backend="""smddp""" , timeout=self.ddp_timeout_delta ) UpperCamelCase = int(os.getenv("""SMDATAPARALLEL_LOCAL_RANK""" ) ) UpperCamelCase = torch.device("""cuda""" , self.local_rank ) UpperCamelCase = 1 elif self.local_rank == -1: # if n_gpu is > 1 we'll use nn.DataParallel. # If you only want to use a specific subset of GPUs use `CUDA_VISIBLE_DEVICES=0` # Explicitly set CUDA to the first (index 0) CUDA device, otherwise `set_device` will # trigger an error that a device index is missing. Index 0 takes into account the # GPUs available in the environment, so `CUDA_VISIBLE_DEVICES=1,2` with `cuda:0` # will use the first GPU in that env, i.e. GPU#1 UpperCamelCase = torch.device("""cuda:0""" if torch.cuda.is_available() else """cpu""" ) # Sometimes the line in the postinit has not been run before we end up here, so just checking we're not at # the default value. UpperCamelCase = torch.cuda.device_count() else: # Here, we'll use torch.distributed. # Initializes the distributed backend which will take care of synchronizing nodes/GPUs if not torch.distributed.is_initialized(): torch.distributed.init_process_group(backend="""nccl""" , timeout=self.ddp_timeout_delta ) UpperCamelCase = torch.device("""cuda""" , self.local_rank ) UpperCamelCase = 1 if device.type == "cuda": torch.cuda.set_device(_SCREAMING_SNAKE_CASE ) return device @property def A__ ( self ) -> Tuple: """simple docstring""" if is_sagemaker_model_parallel_available(): return smp.dp_size() return super().world_size @property def A__ ( self ) -> Optional[Any]: """simple docstring""" return not is_sagemaker_model_parallel_available() @property def A__ ( self ) -> str: """simple docstring""" return False
183
1
from ...configuration_utils import PretrainedConfig from ...utils import logging a__: Dict = logging.get_logger(__name__) a__: Union[str, Any] = { 'naver-clova-ix/donut-base': 'https://huggingface.co/naver-clova-ix/donut-base/resolve/main/config.json', # See all Donut models at https://huggingface.co/models?filter=donut-swin } class SCREAMING_SNAKE_CASE__ ( UpperCamelCase__ ): __SCREAMING_SNAKE_CASE = '''donut-swin''' __SCREAMING_SNAKE_CASE = { '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self,__lowerCamelCase=224,__lowerCamelCase=4,__lowerCamelCase=3,__lowerCamelCase=96,__lowerCamelCase=[2, 2, 6, 2],__lowerCamelCase=[3, 6, 12, 24],__lowerCamelCase=7,__lowerCamelCase=4.0,__lowerCamelCase=True,__lowerCamelCase=0.0,__lowerCamelCase=0.0,__lowerCamelCase=0.1,__lowerCamelCase="gelu",__lowerCamelCase=False,__lowerCamelCase=0.02,__lowerCamelCase=1E-5,**__lowerCamelCase,): super().__init__(**__lowerCamelCase ) A__ = image_size A__ = patch_size A__ = num_channels A__ = embed_dim A__ = depths A__ = len(__lowerCamelCase ) A__ = num_heads A__ = window_size A__ = mlp_ratio A__ = qkv_bias A__ = hidden_dropout_prob A__ = attention_probs_dropout_prob A__ = drop_path_rate A__ = hidden_act A__ = use_absolute_embeddings A__ = layer_norm_eps A__ = initializer_range # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model A__ = int(embed_dim * 2 ** (len(__lowerCamelCase ) - 1) )
193
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device a__: List[str] = False class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): pass @nightly @require_torch_gpu class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def UpperCamelCase ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase ( self ): A__ = VersatileDiffusionTextToImagePipeline.from_pretrained('''shi-labs/versatile-diffusion''' ) # remove text_unet pipe.remove_unused_weights() pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) A__ = '''A painting of a squirrel eating a burger ''' A__ = torch.manual_seed(0 ) A__ = pipe( prompt=__lowerCamelCase,generator=__lowerCamelCase,guidance_scale=7.5,num_inference_steps=2,output_type='''numpy''' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(__lowerCamelCase ) A__ = VersatileDiffusionTextToImagePipeline.from_pretrained(__lowerCamelCase ) pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) A__ = generator.manual_seed(0 ) A__ = pipe( prompt=__lowerCamelCase,generator=__lowerCamelCase,guidance_scale=7.5,num_inference_steps=2,output_type='''numpy''' ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def UpperCamelCase ( self ): A__ = VersatileDiffusionTextToImagePipeline.from_pretrained( '''shi-labs/versatile-diffusion''',torch_dtype=torch.floataa ) pipe.to(__lowerCamelCase ) pipe.set_progress_bar_config(disable=__lowerCamelCase ) A__ = '''A painting of a squirrel eating a burger ''' A__ = torch.manual_seed(0 ) A__ = pipe( prompt=__lowerCamelCase,generator=__lowerCamelCase,guidance_scale=7.5,num_inference_steps=50,output_type='''numpy''' ).images A__ = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) A__ = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
193
1
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging if TYPE_CHECKING: from ... import FeatureExtractionMixin, TensorType UpperCAmelCase =logging.get_logger(__name__) UpperCAmelCase ={ 'openai/imagegpt-small': '', 'openai/imagegpt-medium': '', 'openai/imagegpt-large': '', } class lowerCamelCase__ ( lowercase__ ): '''simple docstring''' _lowerCamelCase = 'imagegpt' _lowerCamelCase = ['past_key_values'] _lowerCamelCase = { 'hidden_size': 'n_embd', 'max_position_embeddings': 'n_positions', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self ,lowerCamelCase_=5_1_2 + 1 ,lowerCamelCase_=3_2 * 3_2 ,lowerCamelCase_=5_1_2 ,lowerCamelCase_=2_4 ,lowerCamelCase_=8 ,lowerCamelCase_=None ,lowerCamelCase_="quick_gelu" ,lowerCamelCase_=0.1 ,lowerCamelCase_=0.1 ,lowerCamelCase_=0.1 ,lowerCamelCase_=1E-5 ,lowerCamelCase_=0.02 ,lowerCamelCase_=True ,lowerCamelCase_=True ,lowerCamelCase_=False ,lowerCamelCase_=False ,lowerCamelCase_=False ,**lowerCamelCase_ ,) -> List[str]: A = vocab_size A = n_positions A = n_embd A = n_layer A = n_head A = n_inner A = activation_function A = resid_pdrop A = embd_pdrop A = attn_pdrop A = layer_norm_epsilon A = initializer_range A = scale_attn_weights A = use_cache A = scale_attn_by_inverse_layer_idx A = reorder_and_upcast_attn A = tie_word_embeddings super().__init__(tie_word_embeddings=_UpperCamelCase ,**_UpperCamelCase ) class lowerCamelCase__ ( lowercase__ ): '''simple docstring''' @property def UpperCamelCase__ ( self ) -> List[Any]: return OrderedDict( [ ("""input_ids""", {0: """batch""", 1: """sequence"""}), ] ) def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_ = 1 ,lowerCamelCase_ = -1 ,lowerCamelCase_ = False ,lowerCamelCase_ = None ,lowerCamelCase_ = 3 ,lowerCamelCase_ = 3_2 ,lowerCamelCase_ = 3_2 ,) -> List[Any]: A = self._generate_dummy_images(_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ,_UpperCamelCase ) A = dict(preprocessor(images=_UpperCamelCase ,return_tensors=_UpperCamelCase ) ) return inputs
358
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_VISION_2_SEQ_MAPPING if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_VISION_2_SEQ_MAPPING UpperCAmelCase =logging.get_logger(__name__) @add_end_docstrings(SCREAMING_SNAKE_CASE ) class lowerCamelCase__ ( SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self ,*lowerCamelCase_ ,**lowerCamelCase_ ) -> Any: super().__init__(*lowerCamelCase_ ,**lowerCamelCase_ ) requires_backends(self ,"""vision""" ) self.check_model_type( TF_MODEL_FOR_VISION_2_SEQ_MAPPING if self.framework == """tf""" else MODEL_FOR_VISION_2_SEQ_MAPPING ) def UpperCamelCase__ ( self ,lowerCamelCase_=None ,lowerCamelCase_=None ,lowerCamelCase_=None ) -> int: A = {} A = {} if prompt is not None: A = prompt if generate_kwargs is not None: A = generate_kwargs if max_new_tokens is not None: if "generate_kwargs" not in forward_kwargs: A = {} if "max_new_tokens" in forward_kwargs["generate_kwargs"]: raise ValueError( """'max_new_tokens' is defined twice, once in 'generate_kwargs' and once as a direct parameter,""" """ please use only one""" ) A = max_new_tokens return preprocess_params, forward_kwargs, {} def __call__( self ,lowerCamelCase_ ,**lowerCamelCase_ ) -> Any: return super().__call__(lowerCamelCase_ ,**lowerCamelCase_ ) def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_=None ) -> Optional[Any]: A = load_image(lowerCamelCase_ ) if prompt is not None: if not isinstance(lowerCamelCase_ ,lowerCamelCase_ ): raise ValueError( f'Received an invalid text input, got - {type(lowerCamelCase_ )} - but expected a single string. ' """Note also that one single text can be provided for conditional image to text generation.""" ) A = self.model.config.model_type if model_type == "git": A = self.image_processor(images=lowerCamelCase_ ,return_tensors=self.framework ) A = self.tokenizer(text=lowerCamelCase_ ,add_special_tokens=lowerCamelCase_ ).input_ids A = [self.tokenizer.cls_token_id] + input_ids A = torch.tensor(lowerCamelCase_ ).unsqueeze(0 ) model_inputs.update({"""input_ids""": input_ids} ) elif model_type == "pix2struct": A = self.image_processor(images=lowerCamelCase_ ,header_text=lowerCamelCase_ ,return_tensors=self.framework ) elif model_type != "vision-encoder-decoder": # vision-encoder-decoder does not support conditional generation A = self.image_processor(images=lowerCamelCase_ ,return_tensors=self.framework ) A = self.tokenizer(lowerCamelCase_ ,return_tensors=self.framework ) model_inputs.update(lowerCamelCase_ ) else: raise ValueError(f'Model type {model_type} does not support conditional text generation' ) else: A = self.image_processor(images=lowerCamelCase_ ,return_tensors=self.framework ) if self.model.config.model_type == "git" and prompt is None: A = None return model_inputs def UpperCamelCase__ ( self ,lowerCamelCase_ ,lowerCamelCase_=None ) -> Optional[int]: # Git model sets `model_inputs["input_ids"] = None` in `preprocess` (when `prompt=None`). In batch model, the # pipeline will group them into a list of `None`, which fail `_forward`. Avoid this by checking it first. if ( "input_ids" in model_inputs and isinstance(model_inputs["""input_ids"""] ,lowerCamelCase_ ) and all(x is None for x in model_inputs["""input_ids"""] ) ): A = None if generate_kwargs is None: A = {} # FIXME: We need to pop here due to a difference in how `generation.py` and `generation.tf_utils.py` # parse inputs. In the Tensorflow version, `generate` raises an error if we don't use `input_ids` whereas # the PyTorch version matches it with `self.model.main_input_name` or `self.model.encoder.main_input_name` # in the `_prepare_model_inputs` method. A = model_inputs.pop(self.model.main_input_name ) A = self.model.generate(lowerCamelCase_ ,**lowerCamelCase_ ,**lowerCamelCase_ ) return model_outputs def UpperCamelCase__ ( self ,lowerCamelCase_ ) -> Optional[Any]: A = [] for output_ids in model_outputs: A = { """generated_text""": self.tokenizer.decode( lowerCamelCase_ ,skip_special_tokens=lowerCamelCase_ ,) } records.append(lowerCamelCase_ ) return records
77
0
import gc import random import unittest import numpy as np import torch from transformers import CLIPImageProcessor, CLIPVisionConfig, CLIPVisionModel from diffusers import HeunDiscreteScheduler, PriorTransformer, ShapEImgaImgPipeline from diffusers.pipelines.shap_e import ShapERenderer from diffusers.utils import floats_tensor, load_image, load_numpy, slow from diffusers.utils.testing_utils import require_torch_gpu, torch_device from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference class A__ ( __UpperCAmelCase , unittest.TestCase ): """simple docstring""" __A : Optional[int] = ShapEImgaImgPipeline __A : Tuple = ['''image'''] __A : Any = ['''image'''] __A : Optional[Any] = [ '''num_images_per_prompt''', '''num_inference_steps''', '''generator''', '''latents''', '''guidance_scale''', '''frame_size''', '''output_type''', '''return_dict''', ] __A : Dict = False @property def __lowercase ( self) -> Any: '''simple docstring''' return 32 @property def __lowercase ( self) -> Optional[int]: '''simple docstring''' return 32 @property def __lowercase ( self) -> Optional[Any]: '''simple docstring''' return self.time_input_dim * 4 @property def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' return 8 @property def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' torch.manual_seed(0) a__ : Any = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , image_size=64 , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , num_attention_heads=4 , num_channels=3 , num_hidden_layers=5 , patch_size=1 , ) a__ : Dict = CLIPVisionModel(lowercase) return model @property def __lowercase ( self) -> List[Any]: '''simple docstring''' a__ : str = CLIPImageProcessor( crop_size=224 , do_center_crop=lowercase , do_normalize=lowercase , do_resize=lowercase , image_mean=[0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73] , image_std=[0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11] , resample=3 , size=224 , ) return image_processor @property def __lowercase ( self) -> str: '''simple docstring''' torch.manual_seed(0) a__ : str = { 'num_attention_heads': 2, 'attention_head_dim': 16, 'embedding_dim': self.time_input_dim, 'num_embeddings': 32, 'embedding_proj_dim': self.text_embedder_hidden_size, 'time_embed_dim': self.time_embed_dim, 'num_layers': 1, 'clip_embed_dim': self.time_input_dim * 2, 'additional_embeddings': 0, 'time_embed_act_fn': 'gelu', 'norm_in_type': 'layer', 'embedding_proj_norm_type': 'layer', 'encoder_hid_proj_type': None, 'added_emb_type': None, } a__ : Any = PriorTransformer(**lowercase) return model @property def __lowercase ( self) -> Any: '''simple docstring''' torch.manual_seed(0) a__ : List[Any] = { 'param_shapes': ( (self.renderer_dim, 93), (self.renderer_dim, 8), (self.renderer_dim, 8), (self.renderer_dim, 8), ), 'd_latent': self.time_input_dim, 'd_hidden': self.renderer_dim, 'n_output': 12, 'background': ( 0.1, 0.1, 0.1, ), } a__ : List[str] = ShapERenderer(**lowercase) return model def __lowercase ( self) -> str: '''simple docstring''' a__ : Dict = self.dummy_prior a__ : List[str] = self.dummy_image_encoder a__ : int = self.dummy_image_processor a__ : str = self.dummy_renderer a__ : Optional[int] = HeunDiscreteScheduler( beta_schedule='exp' , num_train_timesteps=1024 , prediction_type='sample' , use_karras_sigmas=lowercase , clip_sample=lowercase , clip_sample_range=1.0 , ) a__ : List[Any] = { 'prior': prior, 'image_encoder': image_encoder, 'image_processor': image_processor, 'renderer': renderer, 'scheduler': scheduler, } return components def __lowercase ( self , lowercase , lowercase=0) -> List[str]: '''simple docstring''' a__ : List[str] = floats_tensor((1, 3, 64, 64) , rng=random.Random(lowercase)).to(lowercase) if str(lowercase).startswith('mps'): a__ : List[str] = torch.manual_seed(lowercase) else: a__ : str = torch.Generator(device=lowercase).manual_seed(lowercase) a__ : Tuple = { 'image': input_image, 'generator': generator, 'num_inference_steps': 1, 'frame_size': 32, 'output_type': 'np', } return inputs def __lowercase ( self) -> Any: '''simple docstring''' a__ : int = 'cpu' a__ : List[str] = self.get_dummy_components() a__ : Dict = self.pipeline_class(**lowercase) a__ : Optional[int] = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__ : Tuple = pipe(**self.get_dummy_inputs(lowercase)) a__ : Any = output.images[0] a__ : Any = image[0, -3:, -3:, -1] assert image.shape == (20, 32, 32, 3) a__ : List[str] = np.array( [ 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, 0.00_03_92_16, ]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 def __lowercase ( self) -> Any: '''simple docstring''' self._test_inference_batch_consistent(batch_sizes=[1, 2]) def __lowercase ( self) -> Union[str, Any]: '''simple docstring''' a__ : str = torch_device == 'cpu' a__ : Tuple = True self._test_inference_batch_single_identical( batch_size=2 , test_max_difference=lowercase , relax_max_difference=lowercase , ) def __lowercase ( self) -> Tuple: '''simple docstring''' a__ : List[str] = self.get_dummy_components() a__ : str = self.pipeline_class(**lowercase) a__ : List[str] = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__ : Optional[Any] = 1 a__ : List[str] = 2 a__ : Optional[Any] = self.get_dummy_inputs(lowercase) for key in inputs.keys(): if key in self.batch_params: a__ : Any = batch_size * [inputs[key]] a__ : int = pipe(**lowercase , num_images_per_prompt=lowercase)[0] assert images.shape[0] == batch_size * num_images_per_prompt @slow @require_torch_gpu class A__ ( unittest.TestCase ): """simple docstring""" def __lowercase ( self) -> Dict: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowercase ( self) -> Dict: '''simple docstring''' a__ : str = load_image( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/corgi.png') a__ : str = load_numpy( 'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main' '/shap_e/test_shap_e_img2img_out.npy') a__ : List[str] = ShapEImgaImgPipeline.from_pretrained('openai/shap-e-img2img') a__ : Tuple = pipe.to(lowercase) pipe.set_progress_bar_config(disable=lowercase) a__ : List[Any] = torch.Generator(device=lowercase).manual_seed(0) a__ : Optional[int] = pipe( lowercase , generator=lowercase , guidance_scale=3.0 , num_inference_steps=64 , frame_size=64 , output_type='np' , ).images[0] assert images.shape == (20, 64, 64, 3) assert_mean_pixel_difference(lowercase , lowercase)
99
"""simple docstring""" import logging from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import arg_to_scheduler from transformers import TrainingArguments A : Dict = logging.getLogger(__name__) @dataclass class _UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' __UpperCAmelCase : Optional[float] =field( default=0.0 ,metadata={"""help""": """The label smoothing epsilon to apply (if not zero)."""} ) __UpperCAmelCase : bool =field(default=lowerCAmelCase__ ,metadata={"""help""": """Whether to SortishSamler or not."""} ) __UpperCAmelCase : bool =field( default=lowerCAmelCase__ ,metadata={"""help""": """Whether to use generate to calculate generative metrics (ROUGE, BLEU)."""} ) __UpperCAmelCase : bool =field(default=lowerCAmelCase__ ,metadata={"""help""": """whether to use adafactor"""} ) __UpperCAmelCase : Optional[float] =field( default=lowerCAmelCase__ ,metadata={"""help""": """Encoder layer dropout probability. Goes into model.config."""} ) __UpperCAmelCase : Optional[float] =field( default=lowerCAmelCase__ ,metadata={"""help""": """Decoder layer dropout probability. Goes into model.config."""} ) __UpperCAmelCase : Optional[float] =field(default=lowerCAmelCase__ ,metadata={"""help""": """Dropout probability. Goes into model.config."""} ) __UpperCAmelCase : Optional[float] =field( default=lowerCAmelCase__ ,metadata={"""help""": """Attention dropout probability. Goes into model.config."""} ) __UpperCAmelCase : Optional[str] =field( default="""linear""" ,metadata={"""help""": F'''Which lr scheduler to use. Selected in {sorted(arg_to_scheduler.keys() )}'''} ,)
57
0
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import SegformerImageProcessor, SwinConfig, UperNetConfig, UperNetForSemanticSegmentation def lowerCAmelCase( __lowerCamelCase ): __a = 384 __a = 7 if "tiny" in model_name: __a = 96 __a = (2, 2, 6, 2) __a = (3, 6, 12, 24) elif "small" in model_name: __a = 96 __a = (2, 2, 18, 2) __a = (3, 6, 12, 24) elif "base" in model_name: __a = 128 __a = (2, 2, 18, 2) __a = (4, 8, 16, 32) __a = 12 __a = 512 elif "large" in model_name: __a = 192 __a = (2, 2, 18, 2) __a = (6, 12, 24, 48) __a = 12 __a = 768 # set label information __a = 150 __a = """huggingface/label-files""" __a = """ade20k-id2label.json""" __a = json.load(open(hf_hub_download(_a , _a , repo_type='dataset' ) , 'r' ) ) __a = {int(_a ): v for k, v in idalabel.items()} __a = {v: k for k, v in idalabel.items()} __a = SwinConfig( embed_dim=_a , depths=_a , num_heads=_a , window_size=_a , out_features=['stage1', 'stage2', 'stage3', 'stage4'] , ) __a = UperNetConfig( backbone_config=_a , auxiliary_in_channels=_a , num_labels=_a , idalabel=_a , labelaid=_a , ) return config def lowerCAmelCase( __lowerCamelCase ): __a = [] # fmt: off # stem rename_keys.append(('backbone.patch_embed.projection.weight', 'backbone.embeddings.patch_embeddings.projection.weight') ) rename_keys.append(('backbone.patch_embed.projection.bias', 'backbone.embeddings.patch_embeddings.projection.bias') ) rename_keys.append(('backbone.patch_embed.norm.weight', 'backbone.embeddings.norm.weight') ) rename_keys.append(('backbone.patch_embed.norm.bias', 'backbone.embeddings.norm.bias') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.norm1.weight''', f'''backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.weight''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.norm1.bias''', f'''backbone.encoder.layers.{i}.blocks.{j}.layernorm_before.bias''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_bias_table''', f'''backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_bias_table''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.attn.w_msa.relative_position_index''', f'''backbone.encoder.layers.{i}.blocks.{j}.attention.self.relative_position_index''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.weight''', f'''backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.weight''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.attn.w_msa.proj.bias''', f'''backbone.encoder.layers.{i}.blocks.{j}.attention.output.dense.bias''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.norm2.weight''', f'''backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.weight''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.norm2.bias''', f'''backbone.encoder.layers.{i}.blocks.{j}.layernorm_after.bias''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.weight''', f'''backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.weight''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.ffn.layers.0.0.bias''', f'''backbone.encoder.layers.{i}.blocks.{j}.intermediate.dense.bias''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.ffn.layers.1.weight''', f'''backbone.encoder.layers.{i}.blocks.{j}.output.dense.weight''') ) rename_keys.append((f'''backbone.stages.{i}.blocks.{j}.ffn.layers.1.bias''', f'''backbone.encoder.layers.{i}.blocks.{j}.output.dense.bias''') ) if i < 3: rename_keys.append((f'''backbone.stages.{i}.downsample.reduction.weight''', f'''backbone.encoder.layers.{i}.downsample.reduction.weight''') ) rename_keys.append((f'''backbone.stages.{i}.downsample.norm.weight''', f'''backbone.encoder.layers.{i}.downsample.norm.weight''') ) rename_keys.append((f'''backbone.stages.{i}.downsample.norm.bias''', f'''backbone.encoder.layers.{i}.downsample.norm.bias''') ) rename_keys.append((f'''backbone.norm{i}.weight''', f'''backbone.hidden_states_norms.stage{i+1}.weight''') ) rename_keys.append((f'''backbone.norm{i}.bias''', f'''backbone.hidden_states_norms.stage{i+1}.bias''') ) # decode head rename_keys.extend( [ ('decode_head.conv_seg.weight', 'decode_head.classifier.weight'), ('decode_head.conv_seg.bias', 'decode_head.classifier.bias'), ('auxiliary_head.conv_seg.weight', 'auxiliary_head.classifier.weight'), ('auxiliary_head.conv_seg.bias', 'auxiliary_head.classifier.bias'), ] ) # fmt: on return rename_keys def lowerCAmelCase( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __a = dct.pop(_a ) __a = val def lowerCAmelCase( __lowerCamelCase , __lowerCamelCase ): __a = [int(backbone_config.embed_dim * 2**i ) for i in range(len(backbone_config.depths ) )] for i in range(len(backbone_config.depths ) ): __a = num_features[i] for j in range(backbone_config.depths[i] ): # fmt: off # read in weights + bias of input projection layer (in original implementation, this is a single matrix + bias) __a = state_dict.pop(f'''backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.weight''' ) __a = state_dict.pop(f'''backbone.stages.{i}.blocks.{j}.attn.w_msa.qkv.bias''' ) # next, add query, keys and values (in that order) to the state dict __a = in_proj_weight[:dim, :] __a = in_proj_bias[: dim] __a = in_proj_weight[ dim : dim * 2, : ] __a = in_proj_bias[ dim : dim * 2 ] __a = in_proj_weight[ -dim :, : ] __a = in_proj_bias[-dim :] # fmt: on def lowerCAmelCase( __lowerCamelCase ): __a = x.shape __a = x.reshape(_a , 4 , in_channel // 4 ) __a = x[:, [0, 2, 1, 3], :].transpose(1 , 2 ).reshape(_a , _a ) return x def lowerCAmelCase( __lowerCamelCase ): __a = x.shape __a = x.reshape(_a , in_channel // 4 , 4 ) __a = x[:, :, [0, 2, 1, 3]].transpose(1 , 2 ).reshape(_a , _a ) return x def lowerCAmelCase( __lowerCamelCase ): __a = x.shape[0] __a = x.reshape(4 , in_channel // 4 ) __a = x[[0, 2, 1, 3], :].transpose(0 , 1 ).reshape(_a ) return x def lowerCAmelCase( __lowerCamelCase ): __a = x.shape[0] __a = x.reshape(in_channel // 4 , 4 ) __a = x[:, [0, 2, 1, 3]].transpose(0 , 1 ).reshape(_a ) return x def lowerCAmelCase( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase ): __a = { """upernet-swin-tiny""": """https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_tiny_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210531_112542-e380ad3e.pth""", """upernet-swin-small""": """https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K/upernet_swin_small_patch4_window7_512x512_160k_ade20k_pretrain_224x224_1K_20210526_192015-ee2fff1c.pth""", """upernet-swin-base""": """https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K/upernet_swin_base_patch4_window12_512x512_160k_ade20k_pretrain_384x384_22K_20210531_125459-429057bf.pth""", """upernet-swin-large""": """https://download.openmmlab.com/mmsegmentation/v0.5/swin/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k/upernet_swin_large_patch4_window12_512x512_pretrain_384x384_22K_160k_ade20k_20220318_091743-9ba68901.pth""", } __a = model_name_to_url[model_name] __a = torch.hub.load_state_dict_from_url(_a , map_location='cpu' , file_name=_a )[ """state_dict""" ] for name, param in state_dict.items(): print(_a , param.shape ) __a = get_upernet_config(_a ) __a = UperNetForSemanticSegmentation(_a ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): __a = state_dict.pop(_a ) if "bn" in key: __a = key.replace('bn' , 'batch_norm' ) __a = val # rename keys __a = create_rename_keys(_a ) for src, dest in rename_keys: rename_key(_a , _a , _a ) read_in_q_k_v(_a , config.backbone_config ) # fix downsample parameters for key, value in state_dict.items(): if "downsample" in key: if "reduction" in key: __a = reverse_correct_unfold_reduction_order(_a ) if "norm" in key: __a = reverse_correct_unfold_norm_order(_a ) model.load_state_dict(_a ) # verify on image __a = """https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg""" __a = Image.open(requests.get(_a , stream=_a ).raw ).convert('RGB' ) __a = SegformerImageProcessor() __a = processor(_a , return_tensors='pt' ).pixel_values with torch.no_grad(): __a = model(_a ) __a = outputs.logits print(logits.shape ) print('First values of logits:' , logits[0, 0, :3, :3] ) # assert values if model_name == "upernet-swin-tiny": __a = torch.tensor( [[-7.59_58, -7.59_58, -7.43_02], [-7.59_58, -7.59_58, -7.43_02], [-7.47_97, -7.47_97, -7.30_68]] ) elif model_name == "upernet-swin-small": __a = torch.tensor( [[-7.19_21, -7.19_21, -6.95_32], [-7.19_21, -7.19_21, -6.95_32], [-7.09_08, -7.09_08, -6.85_34]] ) elif model_name == "upernet-swin-base": __a = torch.tensor( [[-6.58_51, -6.58_51, -6.43_30], [-6.58_51, -6.58_51, -6.43_30], [-6.47_63, -6.47_63, -6.32_54]] ) elif model_name == "upernet-swin-large": __a = torch.tensor( [[-7.52_97, -7.52_97, -7.38_02], [-7.52_97, -7.52_97, -7.38_02], [-7.40_44, -7.40_44, -7.25_86]] ) print('Logits:' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , _a , atol=1e-4 ) print('Looks ok!' ) if pytorch_dump_folder_path is not None: print(f'''Saving model {model_name} to {pytorch_dump_folder_path}''' ) model.save_pretrained(_a ) print(f'''Saving processor to {pytorch_dump_folder_path}''' ) processor.save_pretrained(_a ) if push_to_hub: print(f'''Pushing model and processor for {model_name} to hub''' ) model.push_to_hub(f'''openmmlab/{model_name}''' ) processor.push_to_hub(f'''openmmlab/{model_name}''' ) if __name__ == "__main__": lowerCamelCase_ : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""upernet-swin-tiny""", type=str, choices=[F'''upernet-swin-{size}''' for size in ["""tiny""", """small""", """base""", """large"""]], help="""Name of the Swin + UperNet model you\'d like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) lowerCamelCase_ : int = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
365
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCamelCase_ : Dict = {"""configuration_swin""": ["""SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP""", """SwinConfig""", """SwinOnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : int = [ """SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """SwinForImageClassification""", """SwinForMaskedImageModeling""", """SwinModel""", """SwinPreTrainedModel""", """SwinBackbone""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : Any = [ """TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFSwinForImageClassification""", """TFSwinForMaskedImageModeling""", """TFSwinModel""", """TFSwinPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swin import SWIN_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinConfig, SwinOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swin import ( SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, SwinBackbone, SwinForImageClassification, SwinForMaskedImageModeling, SwinModel, SwinPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_swin import ( TF_SWIN_PRETRAINED_MODEL_ARCHIVE_LIST, TFSwinForImageClassification, TFSwinForMaskedImageModeling, TFSwinModel, TFSwinPreTrainedModel, ) else: import sys lowerCamelCase_ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
197
0
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __lowercase (lowerCAmelCase_ , lowerCAmelCase_ , unittest.TestCase ): _UpperCamelCase = CycleDiffusionPipeline _UpperCamelCase = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { "negative_prompt", "height", "width", "negative_prompt_embeds", } _UpperCamelCase = PipelineTesterMixin.required_optional_params - {"latents"} _UpperCamelCase = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""source_prompt"""} ) _UpperCamelCase = IMAGE_TO_IMAGE_IMAGE_PARAMS _UpperCamelCase = IMAGE_TO_IMAGE_IMAGE_PARAMS def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' torch.manual_seed(0 ) __lowerCAmelCase : Optional[int] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=32 , ) __lowerCAmelCase : str = DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , num_train_timesteps=1000 , clip_sample=lowerCamelCase__ , set_alpha_to_one=lowerCamelCase__ , ) torch.manual_seed(0 ) __lowerCAmelCase : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) torch.manual_seed(0 ) __lowerCAmelCase : Union[str, Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) __lowerCAmelCase : Tuple = CLIPTextModel(lowerCamelCase__ ) __lowerCAmelCase : Dict = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) __lowerCAmelCase : int = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def UpperCamelCase__ ( self , A_ , A_=0 ) ->Union[str, Any]: '''simple docstring''' __lowerCAmelCase : Optional[Any] = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) __lowerCAmelCase : Tuple = image / 2 + 0.5 if str(lowerCamelCase__ ).startswith('''mps''' ): __lowerCAmelCase : Optional[Any] = torch.manual_seed(lowerCamelCase__ ) else: __lowerCAmelCase : Any = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) __lowerCAmelCase : str = { '''prompt''': '''An astronaut riding an elephant''', '''source_prompt''': '''An astronaut riding a horse''', '''image''': image, '''generator''': generator, '''num_inference_steps''': 2, '''eta''': 0.1, '''strength''': 0.8, '''guidance_scale''': 3, '''source_guidance_scale''': 1, '''output_type''': '''numpy''', } return inputs def UpperCamelCase__ ( self ) ->str: '''simple docstring''' __lowerCAmelCase : int = '''cpu''' # ensure determinism for the device-dependent torch.Generator __lowerCAmelCase : Optional[Any] = self.get_dummy_components() __lowerCAmelCase : Any = CycleDiffusionPipeline(**lowerCamelCase__ ) __lowerCAmelCase : Optional[int] = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __lowerCAmelCase : Union[str, Any] = self.get_dummy_inputs(lowerCamelCase__ ) __lowerCAmelCase : Any = pipe(**lowerCamelCase__ ) __lowerCAmelCase : Tuple = output.images __lowerCAmelCase : Union[str, Any] = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) __lowerCAmelCase : Optional[Any] = np.array([0.4_459, 0.4_943, 0.4_544, 0.6_643, 0.5_474, 0.4_327, 0.5_701, 0.5_959, 0.5_179] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @unittest.skipIf(torch_device != '''cuda''' , '''This test requires a GPU''' ) def UpperCamelCase__ ( self ) ->Tuple: '''simple docstring''' __lowerCAmelCase : List[str] = self.get_dummy_components() for name, module in components.items(): if hasattr(lowerCamelCase__ , '''half''' ): __lowerCAmelCase : Any = module.half() __lowerCAmelCase : Tuple = CycleDiffusionPipeline(**lowerCamelCase__ ) __lowerCAmelCase : List[Any] = pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __lowerCAmelCase : Tuple = self.get_dummy_inputs(lowerCamelCase__ ) __lowerCAmelCase : Optional[Any] = pipe(**lowerCamelCase__ ) __lowerCAmelCase : str = output.images __lowerCAmelCase : Optional[int] = images[0, -3:, -3:, -1] assert images.shape == (1, 32, 32, 3) __lowerCAmelCase : List[Any] = np.array([0.3_506, 0.4_543, 0.446, 0.4_575, 0.5_195, 0.4_155, 0.5_273, 0.518, 0.4_116] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 @skip_mps def UpperCamelCase__ ( self ) ->Optional[int]: '''simple docstring''' return super().test_save_load_local() @unittest.skip('''non-deterministic pipeline''' ) def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' return super().test_inference_batch_single_identical() @skip_mps def UpperCamelCase__ ( self ) ->List[Any]: '''simple docstring''' return super().test_dict_tuple_outputs_equivalent() @skip_mps def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return super().test_save_load_optional_components() @skip_mps def UpperCamelCase__ ( self ) ->Optional[Any]: '''simple docstring''' return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class __lowercase (unittest.TestCase ): def UpperCamelCase__ ( self ) ->Union[str, Any]: '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self ) ->int: '''simple docstring''' __lowerCAmelCase : Tuple = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/cycle-diffusion/black_colored_car.png''' ) __lowerCAmelCase : Union[str, Any] = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy''' ) __lowerCAmelCase : Union[str, Any] = init_image.resize((512, 512) ) __lowerCAmelCase : str = '''CompVis/stable-diffusion-v1-4''' __lowerCAmelCase : str = DDIMScheduler.from_pretrained(lowerCamelCase__ , subfolder='''scheduler''' ) __lowerCAmelCase : Dict = CycleDiffusionPipeline.from_pretrained( lowerCamelCase__ , scheduler=lowerCamelCase__ , safety_checker=lowerCamelCase__ , torch_dtype=torch.floataa , revision='''fp16''' ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing() __lowerCAmelCase : Optional[Any] = '''A black colored car''' __lowerCAmelCase : Tuple = '''A blue colored car''' __lowerCAmelCase : Optional[Any] = torch.manual_seed(0 ) __lowerCAmelCase : List[str] = pipe( prompt=lowerCamelCase__ , source_prompt=lowerCamelCase__ , image=lowerCamelCase__ , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=lowerCamelCase__ , output_type='''np''' , ) __lowerCAmelCase : str = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5e-1 def UpperCamelCase__ ( self ) ->List[str]: '''simple docstring''' __lowerCAmelCase : List[str] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/cycle-diffusion/black_colored_car.png''' ) __lowerCAmelCase : Tuple = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy''' ) __lowerCAmelCase : List[Any] = init_image.resize((512, 512) ) __lowerCAmelCase : Union[str, Any] = '''CompVis/stable-diffusion-v1-4''' __lowerCAmelCase : str = DDIMScheduler.from_pretrained(lowerCamelCase__ , subfolder='''scheduler''' ) __lowerCAmelCase : str = CycleDiffusionPipeline.from_pretrained(lowerCamelCase__ , scheduler=lowerCamelCase__ , safety_checker=lowerCamelCase__ ) pipe.to(lowerCamelCase__ ) pipe.set_progress_bar_config(disable=lowerCamelCase__ ) pipe.enable_attention_slicing() __lowerCAmelCase : Tuple = '''A black colored car''' __lowerCAmelCase : Union[str, Any] = '''A blue colored car''' __lowerCAmelCase : Dict = torch.manual_seed(0 ) __lowerCAmelCase : Any = pipe( prompt=lowerCamelCase__ , source_prompt=lowerCamelCase__ , image=lowerCamelCase__ , num_inference_steps=100 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=lowerCamelCase__ , output_type='''np''' , ) __lowerCAmelCase : Tuple = output.images assert np.abs(image - expected_image ).max() < 2e-2
275
from typing import List from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = { """snap-research/efficientformer-l1-300""": ( """https://huggingface.co/snap-research/efficientformer-l1-300/resolve/main/config.json""" ), } class UpperCamelCase__ ( lowerCAmelCase_ ): '''simple docstring''' __snake_case : int = "efficientformer" def __init__( self : Optional[int] ,lowerCamelCase__ : List[int] = [3, 2, 6, 4] ,lowerCamelCase__ : List[int] = [48, 96, 224, 448] ,lowerCamelCase__ : List[bool] = [True, True, True, True] ,lowerCamelCase__ : int = 448 ,lowerCamelCase__ : int = 32 ,lowerCamelCase__ : int = 4 ,lowerCamelCase__ : int = 7 ,lowerCamelCase__ : int = 5 ,lowerCamelCase__ : int = 8 ,lowerCamelCase__ : int = 4 ,lowerCamelCase__ : float = 0.0 ,lowerCamelCase__ : int = 16 ,lowerCamelCase__ : int = 3 ,lowerCamelCase__ : int = 3 ,lowerCamelCase__ : int = 3 ,lowerCamelCase__ : int = 2 ,lowerCamelCase__ : int = 1 ,lowerCamelCase__ : float = 0.0 ,lowerCamelCase__ : int = 1 ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : bool = True ,lowerCamelCase__ : float = 1e-5 ,lowerCamelCase__ : str = "gelu" ,lowerCamelCase__ : float = 0.02 ,lowerCamelCase__ : float = 1e-1_2 ,lowerCamelCase__ : int = 224 ,lowerCamelCase__ : float = 1e-0_5 ,**lowerCamelCase__ : str ,) -> None: '''simple docstring''' super().__init__(**lowerCamelCase__ ) SCREAMING_SNAKE_CASE = hidden_act SCREAMING_SNAKE_CASE = hidden_dropout_prob SCREAMING_SNAKE_CASE = hidden_sizes SCREAMING_SNAKE_CASE = num_hidden_layers SCREAMING_SNAKE_CASE = num_attention_heads SCREAMING_SNAKE_CASE = initializer_range SCREAMING_SNAKE_CASE = layer_norm_eps SCREAMING_SNAKE_CASE = patch_size SCREAMING_SNAKE_CASE = num_channels SCREAMING_SNAKE_CASE = depths SCREAMING_SNAKE_CASE = mlp_expansion_ratio SCREAMING_SNAKE_CASE = downsamples SCREAMING_SNAKE_CASE = dim SCREAMING_SNAKE_CASE = key_dim SCREAMING_SNAKE_CASE = attention_ratio SCREAMING_SNAKE_CASE = resolution SCREAMING_SNAKE_CASE = pool_size SCREAMING_SNAKE_CASE = downsample_patch_size SCREAMING_SNAKE_CASE = downsample_stride SCREAMING_SNAKE_CASE = downsample_pad SCREAMING_SNAKE_CASE = drop_path_rate SCREAMING_SNAKE_CASE = num_metaad_blocks SCREAMING_SNAKE_CASE = distillation SCREAMING_SNAKE_CASE = use_layer_scale SCREAMING_SNAKE_CASE = layer_scale_init_value SCREAMING_SNAKE_CASE = image_size SCREAMING_SNAKE_CASE = batch_norm_eps
296
0
import contextlib import importlib import io import unittest import transformers # Try to import everything from transformers to ensure every object can be loaded. from transformers import * # noqa F406 from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, require_flax, require_tf, require_torch from transformers.utils import ContextManagers, find_labels, is_flax_available, is_tf_available, is_torch_available if is_torch_available(): from transformers import BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification if is_tf_available(): from transformers import TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification if is_flax_available(): from transformers import FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification __lowerCamelCase : Tuple = DUMMY_UNKNOWN_IDENTIFIER # An actual model hosted on huggingface.co __lowerCamelCase : Optional[Any] = '''main''' # Default branch name __lowerCamelCase : Optional[int] = '''f2c752cfc5c0ab6f4bdec59acea69eefbee381c2''' # One particular commit (not the top of `main`) __lowerCamelCase : Any = '''aaaaaaa''' # This commit does not exist, so we should 404. __lowerCamelCase : List[str] = '''d9e9f15bc825e4b2c9249e9578f884bbcb5e3684''' # Sha-1 of config.json on the top of `main`, for checking purposes __lowerCamelCase : List[Any] = '''4b243c475af8d0a7754e87d7d096c92e5199ec2fe168a2ee7998e3b8e9bcb1d3''' @contextlib.contextmanager def __SCREAMING_SNAKE_CASE ( ) -> str: """simple docstring""" print("""Welcome!""" ) yield print("""Bye!""" ) @contextlib.contextmanager def __SCREAMING_SNAKE_CASE ( ) -> Any: """simple docstring""" print("""Bonjour!""" ) yield print("""Au revoir!""" ) class __snake_case ( unittest.TestCase ): def __a ( self : List[Any] ): """simple docstring""" assert transformers.__spec__ is not None assert importlib.util.find_spec("""transformers""" ) is not None class __snake_case ( unittest.TestCase ): @unittest.mock.patch("""sys.stdout""" , new_callable=io.StringIO ) def __a ( self : List[str] , _lowercase : str ): """simple docstring""" with ContextManagers([] ): print("""Transformers are awesome!""" ) # The print statement adds a new line at the end of the output self.assertEqual(mock_stdout.getvalue() , """Transformers are awesome!\n""" ) @unittest.mock.patch("""sys.stdout""" , new_callable=io.StringIO ) def __a ( self : Optional[Any] , _lowercase : str ): """simple docstring""" with ContextManagers([context_en()] ): print("""Transformers are awesome!""" ) # The output should be wrapped with an English welcome and goodbye self.assertEqual(mock_stdout.getvalue() , """Welcome!\nTransformers are awesome!\nBye!\n""" ) @unittest.mock.patch("""sys.stdout""" , new_callable=io.StringIO ) def __a ( self : Tuple , _lowercase : Dict ): """simple docstring""" with ContextManagers([context_fr(), context_en()] ): print("""Transformers are awesome!""" ) # The output should be wrapped with an English and French welcome and goodbye self.assertEqual(mock_stdout.getvalue() , """Bonjour!\nWelcome!\nTransformers are awesome!\nBye!\nAu revoir!\n""" ) @require_torch def __a ( self : Union[str, Any] ): """simple docstring""" self.assertEqual(find_labels(_lowercase ) , ["""labels"""] ) self.assertEqual(find_labels(_lowercase ) , ["""labels""", """next_sentence_label"""] ) self.assertEqual(find_labels(_lowercase ) , ["""start_positions""", """end_positions"""] ) class __snake_case ( lowerCamelCase_ ): pass self.assertEqual(find_labels(_lowercase ) , ["""labels"""] ) @require_tf def __a ( self : Any ): """simple docstring""" self.assertEqual(find_labels(_lowercase ) , ["""labels"""] ) self.assertEqual(find_labels(_lowercase ) , ["""labels""", """next_sentence_label"""] ) self.assertEqual(find_labels(_lowercase ) , ["""start_positions""", """end_positions"""] ) class __snake_case ( lowerCamelCase_ ): pass self.assertEqual(find_labels(_lowercase ) , ["""labels"""] ) @require_flax def __a ( self : Union[str, Any] ): """simple docstring""" self.assertEqual(find_labels(_lowercase ) , [] ) self.assertEqual(find_labels(_lowercase ) , [] ) self.assertEqual(find_labels(_lowercase ) , [] ) class __snake_case ( lowerCamelCase_ ): pass self.assertEqual(find_labels(_lowercase ) , [] )
204
import contextlib import importlib import io import unittest import transformers # Try to import everything from transformers to ensure every object can be loaded. from transformers import * # noqa F406 from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, require_flax, require_tf, require_torch from transformers.utils import ContextManagers, find_labels, is_flax_available, is_tf_available, is_torch_available if is_torch_available(): from transformers import BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification if is_tf_available(): from transformers import TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification if is_flax_available(): from transformers import FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification __lowerCamelCase : Tuple = DUMMY_UNKNOWN_IDENTIFIER # An actual model hosted on huggingface.co __lowerCamelCase : Optional[Any] = '''main''' # Default branch name __lowerCamelCase : Optional[int] = '''f2c752cfc5c0ab6f4bdec59acea69eefbee381c2''' # One particular commit (not the top of `main`) __lowerCamelCase : Any = '''aaaaaaa''' # This commit does not exist, so we should 404. __lowerCamelCase : List[str] = '''d9e9f15bc825e4b2c9249e9578f884bbcb5e3684''' # Sha-1 of config.json on the top of `main`, for checking purposes __lowerCamelCase : List[Any] = '''4b243c475af8d0a7754e87d7d096c92e5199ec2fe168a2ee7998e3b8e9bcb1d3''' @contextlib.contextmanager def __SCREAMING_SNAKE_CASE ( ) -> str: """simple docstring""" print("""Welcome!""" ) yield print("""Bye!""" ) @contextlib.contextmanager def __SCREAMING_SNAKE_CASE ( ) -> Any: """simple docstring""" print("""Bonjour!""" ) yield print("""Au revoir!""" ) class __snake_case ( unittest.TestCase ): def __a ( self : List[Any] ): """simple docstring""" assert transformers.__spec__ is not None assert importlib.util.find_spec("""transformers""" ) is not None class __snake_case ( unittest.TestCase ): @unittest.mock.patch("""sys.stdout""" , new_callable=io.StringIO ) def __a ( self : List[str] , _lowercase : str ): """simple docstring""" with ContextManagers([] ): print("""Transformers are awesome!""" ) # The print statement adds a new line at the end of the output self.assertEqual(mock_stdout.getvalue() , """Transformers are awesome!\n""" ) @unittest.mock.patch("""sys.stdout""" , new_callable=io.StringIO ) def __a ( self : Optional[Any] , _lowercase : str ): """simple docstring""" with ContextManagers([context_en()] ): print("""Transformers are awesome!""" ) # The output should be wrapped with an English welcome and goodbye self.assertEqual(mock_stdout.getvalue() , """Welcome!\nTransformers are awesome!\nBye!\n""" ) @unittest.mock.patch("""sys.stdout""" , new_callable=io.StringIO ) def __a ( self : Tuple , _lowercase : Dict ): """simple docstring""" with ContextManagers([context_fr(), context_en()] ): print("""Transformers are awesome!""" ) # The output should be wrapped with an English and French welcome and goodbye self.assertEqual(mock_stdout.getvalue() , """Bonjour!\nWelcome!\nTransformers are awesome!\nBye!\nAu revoir!\n""" ) @require_torch def __a ( self : Union[str, Any] ): """simple docstring""" self.assertEqual(find_labels(_lowercase ) , ["""labels"""] ) self.assertEqual(find_labels(_lowercase ) , ["""labels""", """next_sentence_label"""] ) self.assertEqual(find_labels(_lowercase ) , ["""start_positions""", """end_positions"""] ) class __snake_case ( lowerCamelCase_ ): pass self.assertEqual(find_labels(_lowercase ) , ["""labels"""] ) @require_tf def __a ( self : Any ): """simple docstring""" self.assertEqual(find_labels(_lowercase ) , ["""labels"""] ) self.assertEqual(find_labels(_lowercase ) , ["""labels""", """next_sentence_label"""] ) self.assertEqual(find_labels(_lowercase ) , ["""start_positions""", """end_positions"""] ) class __snake_case ( lowerCamelCase_ ): pass self.assertEqual(find_labels(_lowercase ) , ["""labels"""] ) @require_flax def __a ( self : Union[str, Any] ): """simple docstring""" self.assertEqual(find_labels(_lowercase ) , [] ) self.assertEqual(find_labels(_lowercase ) , [] ) self.assertEqual(find_labels(_lowercase ) , [] ) class __snake_case ( lowerCamelCase_ ): pass self.assertEqual(find_labels(_lowercase ) , [] )
204
1
import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __lowerCAmelCase : Dict =logging.get_logger(__name__) __lowerCAmelCase : Dict ={'vocab_file': 'vocab.txt'} __lowerCAmelCase : List[Any] ={ 'vocab_file': { 'openbmb/cpm-ant-10b': 'https://huggingface.co/openbmb/cpm-ant-10b/blob/main/vocab.txt', }, } __lowerCAmelCase : Any ={ 'openbmb/cpm-ant-10b': 1_0_2_4, } def _UpperCamelCase ( lowercase__ ): __SCREAMING_SNAKE_CASE : Optional[Any] = collections.OrderedDict() with open(lowercase__ , '''r''' , encoding='''utf-8''' ) as reader: __SCREAMING_SNAKE_CASE : Optional[int] = reader.readlines() for index, token in enumerate(lowercase__ ): __SCREAMING_SNAKE_CASE : int = token.rstrip('''\n''' ) __SCREAMING_SNAKE_CASE : Tuple = index return vocab class _lowercase ( A__ ): '''simple docstring''' def __init__( self :List[Any] , lowerCAmelCase__ :List[str] , lowerCAmelCase__ :List[Any]="<unk>" , lowerCAmelCase__ :str=200 ) -> Optional[int]: __SCREAMING_SNAKE_CASE : Optional[Any] = vocab __SCREAMING_SNAKE_CASE : Union[str, Any] = unk_token __SCREAMING_SNAKE_CASE : Optional[Any] = max_input_chars_per_word def __magic_name__( self :str , lowerCAmelCase__ :Tuple ) -> Dict: __SCREAMING_SNAKE_CASE : Union[str, Any] = list(lowerCAmelCase__ ) if len(lowerCAmelCase__ ) > self.max_input_chars_per_word: return [self.unk_token] __SCREAMING_SNAKE_CASE : Union[str, Any] = 0 __SCREAMING_SNAKE_CASE : int = [] while start < len(lowerCAmelCase__ ): __SCREAMING_SNAKE_CASE : Optional[int] = len(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : int = None while start < end: __SCREAMING_SNAKE_CASE : List[str] = ''''''.join(chars[start:end] ) if substr in self.vocab: __SCREAMING_SNAKE_CASE : Optional[int] = substr break end -= 1 if cur_substr is None: sub_tokens.append(self.unk_token ) start += 1 else: sub_tokens.append(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = end return sub_tokens class _lowercase ( A__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : Any = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : List[Any] = ['''input_ids''', '''attention_mask'''] SCREAMING_SNAKE_CASE__ : Any = False def __init__( self :Tuple , lowerCAmelCase__ :int , lowerCAmelCase__ :int="<d>" , lowerCAmelCase__ :int="</d>" , lowerCAmelCase__ :Optional[Any]="<s>" , lowerCAmelCase__ :Optional[Any]="</s>" , lowerCAmelCase__ :Optional[int]="<pad>" , lowerCAmelCase__ :Dict="<unk>" , lowerCAmelCase__ :Any="</n>" , lowerCAmelCase__ :List[str]="</_>" , lowerCAmelCase__ :Optional[Any]="left" , **lowerCAmelCase__ :Tuple , ) -> Optional[Any]: requires_backends(self , ['''jieba'''] ) super().__init__( bod_token=lowerCAmelCase__ , eod_token=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , line_token=lowerCAmelCase__ , space_token=lowerCAmelCase__ , padding_side=lowerCAmelCase__ , **lowerCAmelCase__ , ) __SCREAMING_SNAKE_CASE : Optional[int] = bod_token __SCREAMING_SNAKE_CASE : List[str] = eod_token __SCREAMING_SNAKE_CASE : Optional[int] = load_vocab(lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : int = self.encoder[space_token] __SCREAMING_SNAKE_CASE : str = self.encoder[line_token] del self.encoder[space_token] del self.encoder[line_token] __SCREAMING_SNAKE_CASE : Dict = collections.OrderedDict(sorted(self.encoder.items() , key=lambda lowerCAmelCase__ : x[1] ) ) __SCREAMING_SNAKE_CASE : Tuple = {v: k for k, v in self.encoder.items()} __SCREAMING_SNAKE_CASE : Any = WordpieceTokenizer(vocab=self.encoder , unk_token=self.unk_token ) @property def __magic_name__( self :List[Any] ) -> int: return self.encoder[self.bod_token] @property def __magic_name__( self :int ) -> List[Any]: return self.encoder[self.eod_token] @property def __magic_name__( self :List[Any] ) -> Union[str, Any]: return self.encoder["\n"] @property def __magic_name__( self :Optional[int] ) -> int: return len(self.encoder ) def __magic_name__( self :int ) -> Optional[int]: return dict(self.encoder , **self.added_tokens_encoder ) def __magic_name__( self :str , lowerCAmelCase__ :List[str] ) -> Any: __SCREAMING_SNAKE_CASE : int = [] for x in jieba.cut(lowerCAmelCase__ , cut_all=lowerCAmelCase__ ): output_tokens.extend(self.wordpiece_tokenizer.tokenize(lowerCAmelCase__ ) ) return output_tokens def __magic_name__( self :str , lowerCAmelCase__ :Tuple , **lowerCAmelCase__ :Optional[int] ) -> List[str]: __SCREAMING_SNAKE_CASE : Any = [i for i in token_ids if i >= 0] __SCREAMING_SNAKE_CASE : int = [ x for x in token_ids if x != self.pad_token_id and x != self.eos_token_id and x != self.bos_token_id ] return super()._decode(lowerCAmelCase__ , **lowerCAmelCase__ ) def __magic_name__( self :int , lowerCAmelCase__ :Tuple ) -> List[str]: return token in self.encoder def __magic_name__( self :Tuple , lowerCAmelCase__ :List[str] ) -> str: return "".join(lowerCAmelCase__ ) def __magic_name__( self :List[str] , lowerCAmelCase__ :List[Any] ) -> List[Any]: return self.encoder.get(lowerCAmelCase__ , self.encoder.get(self.unk_token ) ) def __magic_name__( self :Any , lowerCAmelCase__ :Any ) -> Optional[Any]: return self.decoder.get(lowerCAmelCase__ , self.unk_token ) def __magic_name__( self :List[str] , lowerCAmelCase__ :str , lowerCAmelCase__ :Optional[str] = None ) -> Tuple[str]: if os.path.isdir(lowerCAmelCase__ ): __SCREAMING_SNAKE_CASE : Any = os.path.join( lowerCAmelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) else: __SCREAMING_SNAKE_CASE : Union[str, Any] = (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory __SCREAMING_SNAKE_CASE : Optional[int] = 0 if " " in self.encoder: __SCREAMING_SNAKE_CASE : List[str] = self.encoder[''' '''] del self.encoder[" "] if "\n" in self.encoder: __SCREAMING_SNAKE_CASE : Tuple = self.encoder['''\n'''] del self.encoder["\n"] __SCREAMING_SNAKE_CASE : Optional[int] = collections.OrderedDict(sorted(self.encoder.items() , key=lambda lowerCAmelCase__ : x[1] ) ) with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' ) as writer: for token, token_index in self.encoder.items(): if index != token_index: logger.warning( f'''Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.''' ''' Please check that the vocabulary is not corrupted!''' ) __SCREAMING_SNAKE_CASE : Any = token_index writer.write(token + '''\n''' ) index += 1 return (vocab_file,) def __magic_name__( self :Optional[Any] , lowerCAmelCase__ :List[int] , lowerCAmelCase__ :List[int] = None ) -> List[int]: if token_ids_a is None: return [self.bos_token_id] + token_ids_a return [self.bos_token_id] + token_ids_a + [self.bos_token_id] + token_ids_a def __magic_name__( self :str , lowerCAmelCase__ :List[int] , lowerCAmelCase__ :Optional[List[int]] = None , lowerCAmelCase__ :bool = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCAmelCase__ , token_ids_a=lowerCAmelCase__ , already_has_special_tokens=lowerCAmelCase__ ) if token_ids_a is not None: return [1] + ([0] * len(lowerCAmelCase__ )) + [1] + ([0] * len(lowerCAmelCase__ )) return [1] + ([0] * len(lowerCAmelCase__ ))
9
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __lowerCAmelCase : Optional[int] =logging.get_logger(__name__) __lowerCAmelCase : Optional[Any] ={'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} __lowerCAmelCase : List[str] ={ 'tokenizer_file': { 'EleutherAI/gpt-neox-20b': 'https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/tokenizer.json', }, } __lowerCAmelCase : Optional[int] ={ 'gpt-neox-20b': 2_0_4_8, } class _lowercase ( A__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : Dict = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : Optional[Any] = ['''input_ids''', '''attention_mask'''] def __init__( self :int , lowerCAmelCase__ :Any=None , lowerCAmelCase__ :Optional[Any]=None , lowerCAmelCase__ :List[Any]=None , lowerCAmelCase__ :str="<|endoftext|>" , lowerCAmelCase__ :str="<|endoftext|>" , lowerCAmelCase__ :Dict="<|endoftext|>" , lowerCAmelCase__ :Union[str, Any]=False , **lowerCAmelCase__ :List[str] , ) -> Any: super().__init__( lowerCAmelCase__ , lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , bos_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , add_prefix_space=lowerCAmelCase__ , **lowerCAmelCase__ , ) __SCREAMING_SNAKE_CASE : List[Any] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' , lowerCAmelCase__ ) != add_prefix_space: __SCREAMING_SNAKE_CASE : List[str] = getattr(lowerCAmelCase__ , pre_tok_state.pop('''type''' ) ) __SCREAMING_SNAKE_CASE : str = add_prefix_space __SCREAMING_SNAKE_CASE : Any = pre_tok_class(**lowerCAmelCase__ ) __SCREAMING_SNAKE_CASE : Optional[int] = add_prefix_space def __magic_name__( self :Union[str, Any] , lowerCAmelCase__ :str , lowerCAmelCase__ :Optional[str] = None ) -> Tuple[str]: __SCREAMING_SNAKE_CASE : List[str] = self._tokenizer.model.save(lowerCAmelCase__ , name=lowerCAmelCase__ ) return tuple(lowerCAmelCase__ ) def __magic_name__( self :Optional[Any] , lowerCAmelCase__ :"Conversation" ) -> List[int]: __SCREAMING_SNAKE_CASE : Optional[Any] = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) + [self.eos_token_id] ) if len(lowerCAmelCase__ ) > self.model_max_length: __SCREAMING_SNAKE_CASE : List[str] = input_ids[-self.model_max_length :] return input_ids
9
1
from __future__ import annotations def __lowerCamelCase (UpperCAmelCase__ : list , UpperCAmelCase__ : int , UpperCAmelCase__ : int , UpperCAmelCase__ : int ): SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = input_list[low:mid], input_list[mid : high + 1] while left and right: result.append((left if left[0] <= right[0] else right).pop(0 ) ) SCREAMING_SNAKE_CASE = result + left + right return input_list def __lowerCamelCase (UpperCAmelCase__ : list ): if len(UpperCAmelCase__ ) <= 1: return input_list SCREAMING_SNAKE_CASE = list(UpperCAmelCase__ ) # iteration for two-way merging SCREAMING_SNAKE_CASE = 2 while p <= len(UpperCAmelCase__ ): # getting low, high and middle value for merge-sort of single list for i in range(0 , len(UpperCAmelCase__ ) , UpperCAmelCase__ ): SCREAMING_SNAKE_CASE = i SCREAMING_SNAKE_CASE = i + p - 1 SCREAMING_SNAKE_CASE = (low + high + 1) // 2 SCREAMING_SNAKE_CASE = merge(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) # final merge of last two parts if p * 2 >= len(UpperCAmelCase__ ): SCREAMING_SNAKE_CASE = i SCREAMING_SNAKE_CASE = merge(UpperCAmelCase__ , 0 , UpperCAmelCase__ , len(UpperCAmelCase__ ) - 1 ) break p *= 2 return input_list if __name__ == "__main__": _lowerCamelCase : str = input('''Enter numbers separated by a comma:\n''').strip() if user_input == "": _lowerCamelCase : Optional[Any] = [] else: _lowerCamelCase : Any = [int(item.strip()) for item in user_input.split(''',''')] print(iter_merge_sort(unsorted))
351
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL _lowerCamelCase : List[Any] = logging.get_logger(__name__) class lowercase ( a ): lowercase__ : List[str] = ["""pixel_values"""] def __init__( self : List[str] , _UpperCamelCase : bool = True , _UpperCamelCase : Dict[str, int] = None , _UpperCamelCase : float = None , _UpperCamelCase : PILImageResampling = PILImageResampling.BILINEAR , _UpperCamelCase : bool = True , _UpperCamelCase : Union[int, float] = 1 / 255 , _UpperCamelCase : bool = True , _UpperCamelCase : Optional[Union[float, List[float]]] = None , _UpperCamelCase : Optional[Union[float, List[float]]] = None , **_UpperCamelCase : Dict , ) -> None: '''simple docstring''' super().__init__(**_UpperCamelCase ) SCREAMING_SNAKE_CASE = size if size is not None else {"shortest_edge": 384} SCREAMING_SNAKE_CASE = get_size_dict(_UpperCamelCase , default_to_square=_UpperCamelCase ) SCREAMING_SNAKE_CASE = do_resize SCREAMING_SNAKE_CASE = size # Default value set here for backwards compatibility where the value in config is None SCREAMING_SNAKE_CASE = crop_pct if crop_pct is not None else 224 / 256 SCREAMING_SNAKE_CASE = resample SCREAMING_SNAKE_CASE = do_rescale SCREAMING_SNAKE_CASE = rescale_factor SCREAMING_SNAKE_CASE = do_normalize SCREAMING_SNAKE_CASE = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN SCREAMING_SNAKE_CASE = image_std if image_std is not None else IMAGENET_STANDARD_STD def __snake_case( self : Optional[Any] , _UpperCamelCase : np.ndarray , _UpperCamelCase : Dict[str, int] , _UpperCamelCase : float , _UpperCamelCase : PILImageResampling = PILImageResampling.BICUBIC , _UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **_UpperCamelCase : Optional[int] , ) -> np.ndarray: '''simple docstring''' SCREAMING_SNAKE_CASE = get_size_dict(_UpperCamelCase , default_to_square=_UpperCamelCase ) if "shortest_edge" not in size: raise ValueError(F"Size dictionary must contain 'shortest_edge' key. Got {size.keys()}" ) SCREAMING_SNAKE_CASE = size["shortest_edge"] if shortest_edge < 384: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct SCREAMING_SNAKE_CASE = int(shortest_edge / crop_pct ) SCREAMING_SNAKE_CASE = get_resize_output_image_size(_UpperCamelCase , size=_UpperCamelCase , default_to_square=_UpperCamelCase ) SCREAMING_SNAKE_CASE = resize(image=_UpperCamelCase , size=_UpperCamelCase , resample=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=_UpperCamelCase , size=(shortest_edge, shortest_edge) , data_format=_UpperCamelCase , **_UpperCamelCase ) else: # warping (no cropping) when evaluated at 384 or larger return resize( _UpperCamelCase , size=(shortest_edge, shortest_edge) , resample=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def __snake_case( self : int , _UpperCamelCase : np.ndarray , _UpperCamelCase : Union[int, float] , _UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **_UpperCamelCase : Optional[Any] , ) -> List[Any]: '''simple docstring''' return rescale(_UpperCamelCase , scale=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def __snake_case( self : Tuple , _UpperCamelCase : np.ndarray , _UpperCamelCase : Union[float, List[float]] , _UpperCamelCase : Union[float, List[float]] , _UpperCamelCase : Optional[Union[str, ChannelDimension]] = None , **_UpperCamelCase : List[Any] , ) -> np.ndarray: '''simple docstring''' return normalize(_UpperCamelCase , mean=_UpperCamelCase , std=_UpperCamelCase , data_format=_UpperCamelCase , **_UpperCamelCase ) def __snake_case( self : List[str] , _UpperCamelCase : ImageInput , _UpperCamelCase : bool = None , _UpperCamelCase : Dict[str, int] = None , _UpperCamelCase : float = None , _UpperCamelCase : PILImageResampling = None , _UpperCamelCase : bool = None , _UpperCamelCase : float = None , _UpperCamelCase : bool = None , _UpperCamelCase : Optional[Union[float, List[float]]] = None , _UpperCamelCase : Optional[Union[float, List[float]]] = None , _UpperCamelCase : Optional[Union[str, TensorType]] = None , _UpperCamelCase : ChannelDimension = ChannelDimension.FIRST , **_UpperCamelCase : Any , ) -> PIL.Image.Image: '''simple docstring''' SCREAMING_SNAKE_CASE = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE = crop_pct if crop_pct is not None else self.crop_pct SCREAMING_SNAKE_CASE = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE = size if size is not None else self.size SCREAMING_SNAKE_CASE = get_size_dict(_UpperCamelCase , default_to_square=_UpperCamelCase ) SCREAMING_SNAKE_CASE = make_list_of_images(_UpperCamelCase ) if not valid_images(_UpperCamelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_resize and size["shortest_edge"] < 384 and crop_pct is None: raise ValueError("crop_pct must be specified if size < 384." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE = [to_numpy_array(_UpperCamelCase ) for image in images] if do_resize: SCREAMING_SNAKE_CASE = [self.resize(image=_UpperCamelCase , size=_UpperCamelCase , crop_pct=_UpperCamelCase , resample=_UpperCamelCase ) for image in images] if do_rescale: SCREAMING_SNAKE_CASE = [self.rescale(image=_UpperCamelCase , scale=_UpperCamelCase ) for image in images] if do_normalize: SCREAMING_SNAKE_CASE = [self.normalize(image=_UpperCamelCase , mean=_UpperCamelCase , std=_UpperCamelCase ) for image in images] SCREAMING_SNAKE_CASE = [to_channel_dimension_format(_UpperCamelCase , _UpperCamelCase ) for image in images] SCREAMING_SNAKE_CASE = {"pixel_values": images} return BatchFeature(data=_UpperCamelCase , tensor_type=_UpperCamelCase )
206
0
'''simple docstring''' from ..utils import DummyObject, requires_backends class _A ( metaclass=__SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : List[Any] = ["onnx"] def __init__( self , *__UpperCAmelCase , **__UpperCAmelCase ) -> List[str]: '''simple docstring''' requires_backends(self , ["""onnx"""] ) @classmethod def __A ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' requires_backends(cls , ["""onnx"""] ) @classmethod def __A ( cls , *__UpperCAmelCase , **__UpperCAmelCase ) -> Tuple: '''simple docstring''' requires_backends(cls , ["""onnx"""] )
254
'''simple docstring''' # Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version _UpperCamelCase = get_logger(__name__) class _A : _SCREAMING_SNAKE_CASE : Dict = "dummy_data" _SCREAMING_SNAKE_CASE : int = "datasets" _SCREAMING_SNAKE_CASE : Union[str, Any] = False def __init__( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase = None , __UpperCAmelCase = False , __UpperCAmelCase = True , __UpperCAmelCase = None , ) -> List[Any]: '''simple docstring''' __UpperCAmelCase : int = 0 __UpperCAmelCase : Tuple = dataset_name __UpperCAmelCase : List[Any] = cache_dir __UpperCAmelCase : List[Any] = use_local_dummy_data __UpperCAmelCase : str = config # download_callbacks take a single url as input __UpperCAmelCase : List[Callable] = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root __UpperCAmelCase : Tuple = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general __UpperCAmelCase : Dict = str(__UpperCAmelCase ) # to be downloaded __UpperCAmelCase : Any = None __UpperCAmelCase : List[str] = None @property def __A ( self ) -> Dict: '''simple docstring''' if self._dummy_file is None: __UpperCAmelCase : Dict = self.download_dummy_data() return self._dummy_file @property def __A ( self ) -> Any: '''simple docstring''' if self.config is not None: # structure is dummy / config_name / version_name return os.path.join("""dummy""" , self.config.name , self.version_name ) # structure is dummy / version_name return os.path.join("""dummy""" , self.version_name ) @property def __A ( self ) -> Optional[Any]: '''simple docstring''' return os.path.join(self.dummy_data_folder , """dummy_data.zip""" ) def __A ( self ) -> List[Any]: '''simple docstring''' __UpperCAmelCase : Tuple = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) __UpperCAmelCase : Tuple = cached_path( __UpperCAmelCase , cache_dir=self.cache_dir , extract_compressed_file=__UpperCAmelCase , force_extract=__UpperCAmelCase ) return os.path.join(__UpperCAmelCase , self.dummy_file_name ) @property def __A ( self ) -> Any: '''simple docstring''' return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file ) @property def __A ( self ) -> List[Any]: '''simple docstring''' if self._bucket_url is None: __UpperCAmelCase : Optional[Any] = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , """/""" ) ) return self._bucket_url @property def __A ( self ) -> Dict: '''simple docstring''' # return full path if its a dir if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , """/""" ).split("""/""" )[:-1] ) def __A ( self , __UpperCAmelCase , *__UpperCAmelCase ) -> Optional[int]: '''simple docstring''' if self.load_existing_dummy_data: # dummy data is downloaded and tested __UpperCAmelCase : Dict = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned __UpperCAmelCase : Optional[int] = self.dummy_file_name # special case when data_url is a dict if isinstance(__UpperCAmelCase , __UpperCAmelCase ): return self.create_dummy_data_dict(__UpperCAmelCase , __UpperCAmelCase ) elif isinstance(__UpperCAmelCase , (list, tuple) ): return self.create_dummy_data_list(__UpperCAmelCase , __UpperCAmelCase ) else: return self.create_dummy_data_single(__UpperCAmelCase , __UpperCAmelCase ) def __A ( self , __UpperCAmelCase , *__UpperCAmelCase ) -> Dict: '''simple docstring''' return self.download_and_extract(__UpperCAmelCase ) def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> Any: '''simple docstring''' return self.download_and_extract(__UpperCAmelCase ) def __A ( self , __UpperCAmelCase , *__UpperCAmelCase , **__UpperCAmelCase ) -> Union[str, Any]: '''simple docstring''' return path def __A ( self ) -> List[Any]: '''simple docstring''' return {} def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> Tuple: '''simple docstring''' __UpperCAmelCase : str = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(__UpperCAmelCase , __UpperCAmelCase ): for single_url in single_urls: download_callback(__UpperCAmelCase ) else: __UpperCAmelCase : Union[str, Any] = single_urls download_callback(__UpperCAmelCase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(__UpperCAmelCase , __UpperCAmelCase ): __UpperCAmelCase : str = [os.path.join(__UpperCAmelCase , urllib.parse.quote_plus(Path(__UpperCAmelCase ).name ) ) for x in single_urls] else: __UpperCAmelCase : List[str] = single_urls __UpperCAmelCase : Any = os.path.join(__UpperCAmelCase , urllib.parse.quote_plus(Path(__UpperCAmelCase ).name ) ) __UpperCAmelCase : Union[str, Any] = value # make sure that values are unique if all(isinstance(__UpperCAmelCase , __UpperCAmelCase ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique __UpperCAmelCase : Tuple = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[Any]: '''simple docstring''' __UpperCAmelCase : Optional[Any] = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one __UpperCAmelCase : Tuple = all(bool(re.findall("""[0-9]{3,}-of-[0-9]{3,}""" , __UpperCAmelCase ) ) for url in data_url ) __UpperCAmelCase : str = all( url.startswith("""https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed""" ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): __UpperCAmelCase : int = [data_url[0]] * len(__UpperCAmelCase ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(__UpperCAmelCase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus __UpperCAmelCase : List[str] = os.path.join(__UpperCAmelCase , urllib.parse.quote_plus(single_url.split("""/""" )[-1] ) ) dummy_data_list.append(__UpperCAmelCase ) return dummy_data_list def __A ( self , __UpperCAmelCase , __UpperCAmelCase ) -> Optional[int]: '''simple docstring''' for download_callback in self.download_callbacks: download_callback(__UpperCAmelCase ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus __UpperCAmelCase : List[str] = os.path.join(__UpperCAmelCase , urllib.parse.quote_plus(data_url.split("""/""" )[-1] ) ) if os.path.exists(__UpperCAmelCase ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def __A ( self ) -> Tuple: '''simple docstring''' pass def __A ( self ) -> int: '''simple docstring''' pass def __A ( self , __UpperCAmelCase ) -> Any: '''simple docstring''' def _iter_archive_members(__UpperCAmelCase ): # this preserves the order of the members inside the ZIP archive __UpperCAmelCase : Dict = Path(self.dummy_file ).parent __UpperCAmelCase : Dict = path.relative_to(__UpperCAmelCase ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: __UpperCAmelCase : List[str] = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(__UpperCAmelCase ) __UpperCAmelCase : Any = Path(__UpperCAmelCase ) __UpperCAmelCase : int = _iter_archive_members(__UpperCAmelCase ) if self.use_local_dummy_data else path.rglob("""*""" ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith((""".""", """__""") ): yield file_path.relative_to(__UpperCAmelCase ).as_posix(), file_path.open("""rb""" ) def __A ( self , __UpperCAmelCase ) -> Dict: '''simple docstring''' if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): __UpperCAmelCase : List[Any] = [paths] for path in paths: if os.path.isfile(__UpperCAmelCase ): if os.path.basename(__UpperCAmelCase ).startswith((""".""", """__""") ): return yield path else: for dirpath, dirnames, filenames in os.walk(__UpperCAmelCase ): if os.path.basename(__UpperCAmelCase ).startswith((""".""", """__""") ): continue dirnames.sort() for filename in sorted(__UpperCAmelCase ): if filename.startswith((""".""", """__""") ): continue yield os.path.join(__UpperCAmelCase , __UpperCAmelCase )
254
1
import json import os import unittest from transformers.models.blenderbot_small.tokenization_blenderbot_small import ( VOCAB_FILES_NAMES, BlenderbotSmallTokenizer, ) from ...test_tokenization_common import TokenizerTesterMixin class lowercase ( a , unittest.TestCase ): lowercase__ : Optional[Any] = BlenderbotSmallTokenizer lowercase__ : Dict = False def __snake_case( self : Optional[int] ) -> List[str]: '''simple docstring''' super().setUp() SCREAMING_SNAKE_CASE = ["__start__", "adapt", "act", "ap@@", "te", "__end__", "__unk__"] SCREAMING_SNAKE_CASE = dict(zip(_UpperCamelCase , range(len(_UpperCamelCase ) ) ) ) SCREAMING_SNAKE_CASE = ["#version: 0.2", "a p", "t e</w>", "ap t</w>", "a d", "ad apt</w>", "a c", "ac t</w>", ""] SCREAMING_SNAKE_CASE = {"unk_token": "__unk__", "bos_token": "__start__", "eos_token": "__end__"} SCREAMING_SNAKE_CASE = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) SCREAMING_SNAKE_CASE = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(_UpperCamelCase ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(_UpperCamelCase ) ) def __snake_case( self : int , **_UpperCamelCase : List[Any] ) -> List[Any]: '''simple docstring''' kwargs.update(self.special_tokens_map ) return BlenderbotSmallTokenizer.from_pretrained(self.tmpdirname , **_UpperCamelCase ) def __snake_case( self : List[str] , _UpperCamelCase : str ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = "adapt act apte" SCREAMING_SNAKE_CASE = "adapt act apte" return input_text, output_text def __snake_case( self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = BlenderbotSmallTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) SCREAMING_SNAKE_CASE = "adapt act apte" SCREAMING_SNAKE_CASE = ["adapt", "act", "ap@@", "te"] SCREAMING_SNAKE_CASE = tokenizer.tokenize(_UpperCamelCase ) self.assertListEqual(_UpperCamelCase , _UpperCamelCase ) SCREAMING_SNAKE_CASE = [tokenizer.bos_token] + tokens + [tokenizer.eos_token] SCREAMING_SNAKE_CASE = [0, 1, 2, 3, 4, 5] self.assertListEqual(tokenizer.convert_tokens_to_ids(_UpperCamelCase ) , _UpperCamelCase ) def __snake_case( self : Tuple ) -> Any: '''simple docstring''' SCREAMING_SNAKE_CASE = BlenderbotSmallTokenizer.from_pretrained("facebook/blenderbot-90M" ) assert tok("sam" ).input_ids == [1_384] SCREAMING_SNAKE_CASE = "I am a small frog." SCREAMING_SNAKE_CASE = tok([src_text] , padding=_UpperCamelCase , truncation=_UpperCamelCase )["input_ids"] SCREAMING_SNAKE_CASE = tok.batch_decode(_UpperCamelCase , skip_special_tokens=_UpperCamelCase , clean_up_tokenization_spaces=_UpperCamelCase )[0] assert src_text != decoded # I wish it did! assert decoded == "i am a small frog ." def __snake_case( self : Dict ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = BlenderbotSmallTokenizer.from_pretrained("facebook/blenderbot-90M" ) SCREAMING_SNAKE_CASE = "I am a small frog ." SCREAMING_SNAKE_CASE = "." SCREAMING_SNAKE_CASE = tok(_UpperCamelCase )["input_ids"] SCREAMING_SNAKE_CASE = tok(_UpperCamelCase )["input_ids"] assert encoded[-1] == encoded_dot[0]
354
import argparse import collections import json import os import re import string import sys import numpy as np _lowerCamelCase : Dict = re.compile(r'''\b(a|an|the)\b''', re.UNICODE) _lowerCamelCase : Optional[int] = None def __lowerCamelCase (): SCREAMING_SNAKE_CASE = argparse.ArgumentParser("Official evaluation script for SQuAD version 2.0." ) parser.add_argument("data_file" , metavar="data.json" , help="Input data JSON file." ) parser.add_argument("pred_file" , metavar="pred.json" , help="Model predictions." ) parser.add_argument( "--out-file" , "-o" , metavar="eval.json" , help="Write accuracy metrics to file (default is stdout)." ) parser.add_argument( "--na-prob-file" , "-n" , metavar="na_prob.json" , help="Model estimates of probability of no answer." ) parser.add_argument( "--na-prob-thresh" , "-t" , type=UpperCAmelCase__ , default=1.0 , help="Predict \"\" if no-answer probability exceeds this (default = 1.0)." , ) parser.add_argument( "--out-image-dir" , "-p" , metavar="out_images" , default=UpperCAmelCase__ , help="Save precision-recall curves to directory." ) parser.add_argument("--verbose" , "-v" , action="store_true" ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def __lowerCamelCase (UpperCAmelCase__ : Optional[int] ): SCREAMING_SNAKE_CASE = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: SCREAMING_SNAKE_CASE = bool(qa["answers"]["text"] ) return qid_to_has_ans def __lowerCamelCase (UpperCAmelCase__ : Union[str, Any] ): def remove_articles(UpperCAmelCase__ : List[str] ): return ARTICLES_REGEX.sub(" " , UpperCAmelCase__ ) def white_space_fix(UpperCAmelCase__ : Dict ): return " ".join(text.split() ) def remove_punc(UpperCAmelCase__ : str ): SCREAMING_SNAKE_CASE = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(UpperCAmelCase__ : Dict ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(UpperCAmelCase__ ) ) ) ) def __lowerCamelCase (UpperCAmelCase__ : List[str] ): if not s: return [] return normalize_answer(UpperCAmelCase__ ).split() def __lowerCamelCase (UpperCAmelCase__ : Dict , UpperCAmelCase__ : List[str] ): return int(normalize_answer(UpperCAmelCase__ ) == normalize_answer(UpperCAmelCase__ ) ) def __lowerCamelCase (UpperCAmelCase__ : Any , UpperCAmelCase__ : str ): SCREAMING_SNAKE_CASE = get_tokens(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = get_tokens(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = collections.Counter(UpperCAmelCase__ ) & collections.Counter(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = sum(common.values() ) if len(UpperCAmelCase__ ) == 0 or len(UpperCAmelCase__ ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 SCREAMING_SNAKE_CASE = 1.0 * num_same / len(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = 1.0 * num_same / len(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = (2 * precision * recall) / (precision + recall) return fa def __lowerCamelCase (UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : Tuple ): SCREAMING_SNAKE_CASE = {} SCREAMING_SNAKE_CASE = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: SCREAMING_SNAKE_CASE = qa["id"] SCREAMING_SNAKE_CASE = [t for t in qa["answers"]["text"] if normalize_answer(UpperCAmelCase__ )] if not gold_answers: # For unanswerable questions, only correct answer is empty string SCREAMING_SNAKE_CASE = [""] if qid not in preds: print(F"Missing prediction for {qid}" ) continue SCREAMING_SNAKE_CASE = preds[qid] # Take max over all gold answers SCREAMING_SNAKE_CASE = max(compute_exact(UpperCAmelCase__ , UpperCAmelCase__ ) for a in gold_answers ) SCREAMING_SNAKE_CASE = max(compute_fa(UpperCAmelCase__ , UpperCAmelCase__ ) for a in gold_answers ) return exact_scores, fa_scores def __lowerCamelCase (UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Dict , UpperCAmelCase__ : str ): SCREAMING_SNAKE_CASE = {} for qid, s in scores.items(): SCREAMING_SNAKE_CASE = na_probs[qid] > na_prob_thresh if pred_na: SCREAMING_SNAKE_CASE = float(not qid_to_has_ans[qid] ) else: SCREAMING_SNAKE_CASE = s return new_scores def __lowerCamelCase (UpperCAmelCase__ : List[str] , UpperCAmelCase__ : str , UpperCAmelCase__ : Dict=None ): if not qid_list: SCREAMING_SNAKE_CASE = len(UpperCAmelCase__ ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores.values() ) / total), ("f1", 100.0 * sum(fa_scores.values() ) / total), ("total", total), ] ) else: SCREAMING_SNAKE_CASE = len(UpperCAmelCase__ ) return collections.OrderedDict( [ ("exact", 100.0 * sum(exact_scores[k] for k in qid_list ) / total), ("f1", 100.0 * sum(fa_scores[k] for k in qid_list ) / total), ("total", total), ] ) def __lowerCamelCase (UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[str] ): for k in new_eval: SCREAMING_SNAKE_CASE = new_eval[k] def __lowerCamelCase (UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : int , UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : List[str] ): plt.step(UpperCAmelCase__ , UpperCAmelCase__ , color="b" , alpha=0.2 , where="post" ) plt.fill_between(UpperCAmelCase__ , UpperCAmelCase__ , step="post" , alpha=0.2 , color="b" ) plt.xlabel("Recall" ) plt.ylabel("Precision" ) plt.xlim([0.0, 1.05] ) plt.ylim([0.0, 1.05] ) plt.title(UpperCAmelCase__ ) plt.savefig(UpperCAmelCase__ ) plt.clf() def __lowerCamelCase (UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Tuple , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : int=None , UpperCAmelCase__ : str=None ): SCREAMING_SNAKE_CASE = sorted(UpperCAmelCase__ , key=lambda UpperCAmelCase__ : na_probs[k] ) SCREAMING_SNAKE_CASE = 0.0 SCREAMING_SNAKE_CASE = 1.0 SCREAMING_SNAKE_CASE = 0.0 SCREAMING_SNAKE_CASE = [1.0] SCREAMING_SNAKE_CASE = [0.0] SCREAMING_SNAKE_CASE = 0.0 for i, qid in enumerate(UpperCAmelCase__ ): if qid_to_has_ans[qid]: true_pos += scores[qid] SCREAMING_SNAKE_CASE = true_pos / float(i + 1 ) SCREAMING_SNAKE_CASE = true_pos / float(UpperCAmelCase__ ) if i == len(UpperCAmelCase__ ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(UpperCAmelCase__ ) recalls.append(UpperCAmelCase__ ) if out_image: plot_pr_curve(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) return {"ap": 100.0 * avg_prec} def __lowerCamelCase (UpperCAmelCase__ : Dict , UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Any , UpperCAmelCase__ : List[str] ): if out_image_dir and not os.path.exists(UpperCAmelCase__ ): os.makedirs(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return SCREAMING_SNAKE_CASE = make_precision_recall_eval( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , out_image=os.path.join(UpperCAmelCase__ , "pr_exact.png" ) , title="Precision-Recall curve for Exact Match score" , ) SCREAMING_SNAKE_CASE = make_precision_recall_eval( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , out_image=os.path.join(UpperCAmelCase__ , "pr_f1.png" ) , title="Precision-Recall curve for F1 score" , ) SCREAMING_SNAKE_CASE = {k: float(UpperCAmelCase__ ) for k, v in qid_to_has_ans.items()} SCREAMING_SNAKE_CASE = make_precision_recall_eval( UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , out_image=os.path.join(UpperCAmelCase__ , "pr_oracle.png" ) , title="Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)" , ) merge_eval(UpperCAmelCase__ , UpperCAmelCase__ , "pr_exact" ) merge_eval(UpperCAmelCase__ , UpperCAmelCase__ , "pr_f1" ) merge_eval(UpperCAmelCase__ , UpperCAmelCase__ , "pr_oracle" ) def __lowerCamelCase (UpperCAmelCase__ : List[Any] , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : int ): if not qid_list: return SCREAMING_SNAKE_CASE = [na_probs[k] for k in qid_list] SCREAMING_SNAKE_CASE = np.ones_like(UpperCAmelCase__ ) / float(len(UpperCAmelCase__ ) ) plt.hist(UpperCAmelCase__ , weights=UpperCAmelCase__ , bins=2_0 , range=(0.0, 1.0) ) plt.xlabel("Model probability of no-answer" ) plt.ylabel("Proportion of dataset" ) plt.title(F"Histogram of no-answer probability: {name}" ) plt.savefig(os.path.join(UpperCAmelCase__ , F"na_prob_hist_{name}.png" ) ) plt.clf() def __lowerCamelCase (UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[Any] , UpperCAmelCase__ : List[str] , UpperCAmelCase__ : Dict ): SCREAMING_SNAKE_CASE = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) SCREAMING_SNAKE_CASE = num_no_ans SCREAMING_SNAKE_CASE = cur_score SCREAMING_SNAKE_CASE = 0.0 SCREAMING_SNAKE_CASE = sorted(UpperCAmelCase__ , key=lambda UpperCAmelCase__ : na_probs[k] ) for i, qid in enumerate(UpperCAmelCase__ ): if qid not in scores: continue if qid_to_has_ans[qid]: SCREAMING_SNAKE_CASE = scores[qid] else: if preds[qid]: SCREAMING_SNAKE_CASE = -1 else: SCREAMING_SNAKE_CASE = 0 cur_score += diff if cur_score > best_score: SCREAMING_SNAKE_CASE = cur_score SCREAMING_SNAKE_CASE = na_probs[qid] return 100.0 * best_score / len(UpperCAmelCase__ ), best_thresh def __lowerCamelCase (UpperCAmelCase__ : Union[str, Any] , UpperCAmelCase__ : str , UpperCAmelCase__ : Dict , UpperCAmelCase__ : Optional[int] , UpperCAmelCase__ : int , UpperCAmelCase__ : Optional[Any] ): SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = find_best_thresh(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = find_best_thresh(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = best_exact SCREAMING_SNAKE_CASE = exact_thresh SCREAMING_SNAKE_CASE = best_fa SCREAMING_SNAKE_CASE = fa_thresh def __lowerCamelCase (): with open(OPTS.data_file ) as f: SCREAMING_SNAKE_CASE = json.load(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = dataset_json["data"] with open(OPTS.pred_file ) as f: SCREAMING_SNAKE_CASE = json.load(UpperCAmelCase__ ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: SCREAMING_SNAKE_CASE = json.load(UpperCAmelCase__ ) else: SCREAMING_SNAKE_CASE = {k: 0.0 for k in preds} SCREAMING_SNAKE_CASE = make_qid_to_has_ans(UpperCAmelCase__ ) # maps qid to True/False SCREAMING_SNAKE_CASE = [k for k, v in qid_to_has_ans.items() if v] SCREAMING_SNAKE_CASE = [k for k, v in qid_to_has_ans.items() if not v] SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = get_raw_scores(UpperCAmelCase__ , UpperCAmelCase__ ) SCREAMING_SNAKE_CASE = apply_no_ans_threshold(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , OPTS.na_prob_thresh ) SCREAMING_SNAKE_CASE = apply_no_ans_threshold(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , OPTS.na_prob_thresh ) SCREAMING_SNAKE_CASE = make_eval_dict(UpperCAmelCase__ , UpperCAmelCase__ ) if has_ans_qids: SCREAMING_SNAKE_CASE = make_eval_dict(UpperCAmelCase__ , UpperCAmelCase__ , qid_list=UpperCAmelCase__ ) merge_eval(UpperCAmelCase__ , UpperCAmelCase__ , "HasAns" ) if no_ans_qids: SCREAMING_SNAKE_CASE = make_eval_dict(UpperCAmelCase__ , UpperCAmelCase__ , qid_list=UpperCAmelCase__ ) merge_eval(UpperCAmelCase__ , UpperCAmelCase__ , "NoAns" ) if OPTS.na_prob_file: find_all_best_thresh(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , OPTS.out_image_dir ) histogram_na_prob(UpperCAmelCase__ , UpperCAmelCase__ , OPTS.out_image_dir , "hasAns" ) histogram_na_prob(UpperCAmelCase__ , UpperCAmelCase__ , OPTS.out_image_dir , "noAns" ) if OPTS.out_file: with open(OPTS.out_file , "w" ) as f: json.dump(UpperCAmelCase__ , UpperCAmelCase__ ) else: print(json.dumps(UpperCAmelCase__ , indent=2 ) ) if __name__ == "__main__": _lowerCamelCase : Optional[Any] = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use('''Agg''') import matplotlib.pyplot as plt main()
206
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __snake_case = { '''configuration_groupvit''': [ '''GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GroupViTConfig''', '''GroupViTOnnxConfig''', '''GroupViTTextConfig''', '''GroupViTVisionConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = [ '''GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GroupViTModel''', '''GroupViTPreTrainedModel''', '''GroupViTTextModel''', '''GroupViTVisionModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = [ '''TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFGroupViTModel''', '''TFGroupViTPreTrainedModel''', '''TFGroupViTTextModel''', '''TFGroupViTVisionModel''', ] if TYPE_CHECKING: from .configuration_groupvit import ( GROUPVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GroupViTConfig, GroupViTOnnxConfig, GroupViTTextConfig, GroupViTVisionConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_groupvit import ( GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, GroupViTModel, GroupViTPreTrainedModel, GroupViTTextModel, GroupViTVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_groupvit import ( TF_GROUPVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFGroupViTModel, TFGroupViTPreTrainedModel, TFGroupViTTextModel, TFGroupViTVisionModel, ) else: import sys __snake_case = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
310
import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def _A ( _lowercase ) -> Dict: """simple docstring""" if is_torch_version('<' , '2.0.0' ) or not hasattr(_lowercase , '_dynamo' ): return False return isinstance(_lowercase , torch._dynamo.eval_frame.OptimizedModule ) def _A ( _lowercase , _lowercase = True ) -> Optional[int]: """simple docstring""" __UpperCamelCase = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) __UpperCamelCase = is_compiled_module(_lowercase ) if is_compiled: __UpperCamelCase = model __UpperCamelCase = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(_lowercase , _lowercase ): __UpperCamelCase = model.module if not keep_fpaa_wrapper: __UpperCamelCase = getattr(_lowercase , 'forward' ) __UpperCamelCase = model.__dict__.pop('_original_forward' , _lowercase ) if original_forward is not None: while hasattr(_lowercase , '__wrapped__' ): __UpperCamelCase = forward.__wrapped__ if forward == original_forward: break __UpperCamelCase = forward if getattr(_lowercase , '_converted_to_transformer_engine' , _lowercase ): convert_model(_lowercase , to_transformer_engine=_lowercase ) if is_compiled: __UpperCamelCase = model __UpperCamelCase = compiled_model return model def _A ( ) -> Any: """simple docstring""" PartialState().wait_for_everyone() def _A ( _lowercase , _lowercase ) -> Any: """simple docstring""" if PartialState().distributed_type == DistributedType.TPU: xm.save(_lowercase , _lowercase ) elif PartialState().local_process_index == 0: torch.save(_lowercase , _lowercase ) @contextmanager def _A ( **_lowercase ) -> Union[str, Any]: """simple docstring""" for key, value in kwargs.items(): __UpperCamelCase = str(_lowercase ) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def _A ( _lowercase ) -> Tuple: """simple docstring""" if not hasattr(_lowercase , '__qualname__' ) and not hasattr(_lowercase , '__name__' ): __UpperCamelCase = getattr(_lowercase , '__class__' , _lowercase ) if hasattr(_lowercase , '__qualname__' ): return obj.__qualname__ if hasattr(_lowercase , '__name__' ): return obj.__name__ return str(_lowercase ) def _A ( _lowercase , _lowercase ) -> Any: """simple docstring""" for key, value in source.items(): if isinstance(_lowercase , _lowercase ): __UpperCamelCase = destination.setdefault(_lowercase , {} ) merge_dicts(_lowercase , _lowercase ) else: __UpperCamelCase = value return destination def _A ( _lowercase = None ) -> bool: """simple docstring""" if port is None: __UpperCamelCase = 2_95_00 with socket.socket(socket.AF_INET , socket.SOCK_STREAM ) as s: return s.connect_ex(('localhost', port) ) == 0
310
1
_UpperCAmelCase : str = { "A": ".-", "B": "-...", "C": "-.-.", "D": "-..", "E": ".", "F": "..-.", "G": "--.", "H": "....", "I": "..", "J": ".---", "K": "-.-", "L": ".-..", "M": "--", "N": "-.", "O": "---", "P": ".--.", "Q": "--.-", "R": ".-.", "S": "...", "T": "-", "U": "..-", "V": "...-", "W": ".--", "X": "-..-", "Y": "-.--", "Z": "--..", "1": ".----", "2": "..---", "3": "...--", "4": "....-", "5": ".....", "6": "-....", "7": "--...", "8": "---..", "9": "----.", "0": "-----", "&": ".-...", "@": ".--.-.", ":": "---...", ",": "--..--", ".": ".-.-.-", "'": ".----.", "\"": ".-..-.", "?": "..--..", "/": "-..-.", "=": "-...-", "+": ".-.-.", "-": "-....-", "(": "-.--.", ")": "-.--.-", "!": "-.-.--", " ": "/" } # Exclamation mark is not in ITU-R recommendation # fmt: on _UpperCAmelCase : List[Any] = {value: key for key, value in MORSE_CODE_DICT.items()} def UpperCAmelCase__ ( lowerCamelCase ): return " ".join(MORSE_CODE_DICT[char] for char in message.upper() ) def UpperCAmelCase__ ( lowerCamelCase ): return "".join(REVERSE_DICT[char] for char in message.split() ) def UpperCAmelCase__ ( ): lowercase :str = "Morse code here!" print(lowerCamelCase ) lowercase :Optional[int] = encrypt(lowerCamelCase ) print(lowerCamelCase ) lowercase :Optional[Any] = decrypt(lowerCamelCase ) print(lowerCamelCase ) if __name__ == "__main__": main()
158
def UpperCAmelCase__ ( lowerCamelCase ): if not isinstance(lowerCamelCase, lowerCamelCase ): raise TypeError("Input value must be an 'int' type" ) lowercase :int = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
158
1
from math import asin, atan, cos, radians, sin, sqrt, tan __snake_case = 6_378_137.0 __snake_case = 6_356_752.314_245 __snake_case = 6_37_81_37 def _lowercase ( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ) -> float: '''simple docstring''' SCREAMING_SNAKE_CASE__ = (AXIS_A - AXIS_B) / AXIS_A SCREAMING_SNAKE_CASE__ = atan((1 - flattening) * tan(radians(UpperCamelCase_ ) ) ) SCREAMING_SNAKE_CASE__ = atan((1 - flattening) * tan(radians(UpperCamelCase_ ) ) ) SCREAMING_SNAKE_CASE__ = radians(UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = radians(UpperCamelCase_ ) # Equation SCREAMING_SNAKE_CASE__ = sin((phi_a - phi_a) / 2 ) SCREAMING_SNAKE_CASE__ = sin((lambda_a - lambda_a) / 2 ) # Square both values sin_sq_phi *= sin_sq_phi sin_sq_lambda *= sin_sq_lambda SCREAMING_SNAKE_CASE__ = sqrt(sin_sq_phi + (cos(UpperCamelCase_ ) * cos(UpperCamelCase_ ) * sin_sq_lambda) ) return 2 * RADIUS * asin(UpperCamelCase_ ) if __name__ == "__main__": import doctest doctest.testmod()
176
import json import os import unittest from transformers.models.gptsan_japanese.tokenization_gptsan_japanese import ( VOCAB_FILES_NAMES, GPTSanJapaneseTokenizer, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowercase__ ( _UpperCAmelCase , unittest.TestCase ): A__ : Any =GPTSanJapaneseTokenizer A__ : str =False A__ : int ={"""do_clean_text""": False, """add_prefix_space""": False} def A_ ( self : Any ): super().setUp() # fmt: off SCREAMING_SNAKE_CASE__ = ['こん', 'こんに', 'にちは', 'ばんは', '世界,㔺界', '、', '。', '<BR>', '<SP>', '<TAB>', '<URL>', '<EMAIL>', '<TEL>', '<DATE>', '<PRICE>', '<BLOCK>', '<KIGOU>', '<U2000U2BFF>', '<|emoji1|>', '<unk>', '<|bagoftoken|>', '<|endoftext|>'] # fmt: on SCREAMING_SNAKE_CASE__ = {'emoji': {'\ud83d\ude00': '<|emoji1|>'}, 'emoji_inv': {'<|emoji1|>': '\ud83d\ude00'}} # 😀 SCREAMING_SNAKE_CASE__ = {'unk_token': '<unk>'} SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) SCREAMING_SNAKE_CASE__ = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['emoji_file'] ) with open(self.vocab_file , 'w' , encoding='utf-8' ) as vocab_writer: vocab_writer.write(''.join([x + '\n' for x in vocab_tokens] ) ) with open(self.emoji_file , 'w' ) as emoji_writer: emoji_writer.write(json.dumps(UpperCAmelCase_ ) ) def A_ ( self : str , **UpperCAmelCase_ : Optional[Any] ): kwargs.update(self.special_tokens_map ) return GPTSanJapaneseTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase_ ) def A_ ( self : int , UpperCAmelCase_ : List[str] ): SCREAMING_SNAKE_CASE__ = 'こんにちは、世界。 \nこんばんは、㔺界。😀' SCREAMING_SNAKE_CASE__ = 'こんにちは、世界。 \nこんばんは、世界。😀' return input_text, output_text def A_ ( self : Any , UpperCAmelCase_ : str ): SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ = self.get_input_output_texts(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.encode(UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.decode(UpperCAmelCase_ , clean_up_tokenization_spaces=UpperCAmelCase_ ) return text, ids def A_ ( self : str ): pass # TODO add if relevant def A_ ( self : Tuple ): pass # TODO add if relevant def A_ ( self : int ): pass # TODO add if relevant def A_ ( self : Optional[int] ): SCREAMING_SNAKE_CASE__ = self.get_tokenizer() # Testing tokenization SCREAMING_SNAKE_CASE__ = 'こんにちは、世界。 こんばんは、㔺界。' SCREAMING_SNAKE_CASE__ = ['こん', 'にちは', '、', '世界', '。', '<SP>', 'こん', 'ばんは', '、', '㔺界', '。'] SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) # Testing conversion to ids without special tokens SCREAMING_SNAKE_CASE__ = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6] SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) # Testing conversion to ids with special tokens SCREAMING_SNAKE_CASE__ = tokens + [tokenizer.unk_token] SCREAMING_SNAKE_CASE__ = [0, 2, 5, 4, 6, 8, 0, 3, 5, 4, 6, 19] SCREAMING_SNAKE_CASE__ = tokenizer.convert_tokens_to_ids(UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def A_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE__ = self.get_tokenizer() # Testing tokenization SCREAMING_SNAKE_CASE__ = 'こんにちは、<|bagoftoken|>世界。こんばんは、<|bagoftoken|>㔺界。' SCREAMING_SNAKE_CASE__ = 'こんにちは、、、、世界。こんばんは、、、、世界。' SCREAMING_SNAKE_CASE__ = tokenizer.encode(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.decode(UpperCAmelCase_ ) self.assertEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @slow def A_ ( self : str ): SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization SCREAMING_SNAKE_CASE__ = 'こんにちは、世界。' SCREAMING_SNAKE_CASE__ = 'こんばんは、㔺界。😀' SCREAMING_SNAKE_CASE__ = 'こんにちは、世界。こんばんは、世界。😀' SCREAMING_SNAKE_CASE__ = tokenizer.encode(prefix_text + input_text ) SCREAMING_SNAKE_CASE__ = tokenizer.encode('' , prefix_text=prefix_text + input_text ) SCREAMING_SNAKE_CASE__ = tokenizer.encode(UpperCAmelCase_ , prefix_text=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.decode(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.decode(UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.decode(UpperCAmelCase_ ) self.assertEqual(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertEqual(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @slow def A_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) # Testing tokenization SCREAMING_SNAKE_CASE__ = 'こんにちは、世界。' SCREAMING_SNAKE_CASE__ = 'こんばんは、㔺界。😀' SCREAMING_SNAKE_CASE__ = len(tokenizer.encode(UpperCAmelCase_ ) ) - 2 SCREAMING_SNAKE_CASE__ = len(tokenizer.encode(UpperCAmelCase_ ) ) - 2 SCREAMING_SNAKE_CASE__ = [1] + [0] * (len_prefix + len_text + 1) SCREAMING_SNAKE_CASE__ = [1] * (len_prefix + len_text + 1) + [0] SCREAMING_SNAKE_CASE__ = [1] + [1] * (len_prefix) + [0] * (len_text + 1) SCREAMING_SNAKE_CASE__ = tokenizer(prefix_text + input_text ).token_type_ids SCREAMING_SNAKE_CASE__ = tokenizer('' , prefix_text=prefix_text + input_text ).token_type_ids SCREAMING_SNAKE_CASE__ = tokenizer(UpperCAmelCase_ , prefix_text=UpperCAmelCase_ ).token_type_ids self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @slow def A_ ( self : Tuple ): SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) SCREAMING_SNAKE_CASE__ = tokenizer.encode('あンいワ' ) SCREAMING_SNAKE_CASE__ = tokenizer.encode('' , prefix_text='あンいワ' ) SCREAMING_SNAKE_CASE__ = tokenizer.encode('いワ' , prefix_text='あン' ) self.assertEqual(tokenizer.decode(UpperCAmelCase_ ) , tokenizer.decode(UpperCAmelCase_ ) ) self.assertEqual(tokenizer.decode(UpperCAmelCase_ ) , tokenizer.decode(UpperCAmelCase_ ) ) self.assertNotEqual(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertNotEqual(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertEqual(x_token_a[1] , x_token_a[-1] ) # SEG token self.assertEqual(x_token_a[1] , x_token_a[3] ) # SEG token @slow def A_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE__ = self.tokenizer_class.from_pretrained('Tanrei/GPTSAN-japanese' ) SCREAMING_SNAKE_CASE__ = [['武田信玄', 'は、'], ['織田信長', 'の配下の、']] SCREAMING_SNAKE_CASE__ = tokenizer(UpperCAmelCase_ , padding=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = tokenizer.batch_encode_plus(UpperCAmelCase_ , padding=UpperCAmelCase_ ) # fmt: off SCREAMING_SNAKE_CASE__ = [[35993, 8640, 25948, 35998, 30647, 35675, 35999, 35999], [35993, 10382, 9868, 35998, 30646, 9459, 30646, 35675]] SCREAMING_SNAKE_CASE__ = [[1, 1, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0]] SCREAMING_SNAKE_CASE__ = [[1, 1, 1, 1, 1, 1, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1]] # fmt: on self.assertListEqual(x_token.input_ids , UpperCAmelCase_ ) self.assertListEqual(x_token.token_type_ids , UpperCAmelCase_ ) self.assertListEqual(x_token.attention_mask , UpperCAmelCase_ ) self.assertListEqual(x_token_a.input_ids , UpperCAmelCase_ ) self.assertListEqual(x_token_a.token_type_ids , UpperCAmelCase_ ) self.assertListEqual(x_token_a.attention_mask , UpperCAmelCase_ ) def A_ ( self : Tuple ): # Intentionally convert some words to accommodate character fluctuations unique to Japanese pass def A_ ( self : List[str] ): # tokenizer has no padding token pass
176
1
"""simple docstring""" __UpperCamelCase : str = [ (1_0_0_0, '''M'''), (9_0_0, '''CM'''), (5_0_0, '''D'''), (4_0_0, '''CD'''), (1_0_0, '''C'''), (9_0, '''XC'''), (5_0, '''L'''), (4_0, '''XL'''), (1_0, '''X'''), (9, '''IX'''), (5, '''V'''), (4, '''IV'''), (1, '''I'''), ] def __SCREAMING_SNAKE_CASE ( A_ ): lowerCAmelCase__ : List[str] = {'''I''': 1, '''V''': 5, '''X''': 10, '''L''': 50, '''C''': 1_00, '''D''': 5_00, '''M''': 10_00} lowerCAmelCase__ : Tuple = 0 lowerCAmelCase__ : Union[str, Any] = 0 while place < len(__lowerCAmelCase ): if (place + 1 < len(__lowerCAmelCase )) and (vals[roman[place]] < vals[roman[place + 1]]): total += vals[roman[place + 1]] - vals[roman[place]] place += 2 else: total += vals[roman[place]] place += 1 return total def __SCREAMING_SNAKE_CASE ( A_ ): lowerCAmelCase__ : List[str] = [] for arabic, roman in ROMAN: ((lowerCAmelCase__) ,(lowerCAmelCase__)) : str = divmod(__lowerCAmelCase , __lowerCAmelCase ) result.append(roman * factor ) if number == 0: break return "".join(__lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
370
"""simple docstring""" import os from collections import namedtuple import pytest from datasets import ClassLabel, Features, Sequence, Value from datasets.commands.test import TestCommand from datasets.info import DatasetInfo, DatasetInfosDict __UpperCamelCase : Union[str, Any] = namedtuple( '''_TestCommandArgs''', [ '''dataset''', '''name''', '''cache_dir''', '''data_dir''', '''all_configs''', '''save_infos''', '''ignore_verifications''', '''force_redownload''', '''clear_cache''', ], defaults=[None, None, None, False, False, False, False, False], ) def __SCREAMING_SNAKE_CASE ( A_ , A_ ): return (abs(source - target ) / target) < 0.01 @pytest.mark.integration def __SCREAMING_SNAKE_CASE ( A_ ): lowerCAmelCase__ : Dict = _TestCommandArgs(dataset=A_ , all_configs=A_ , save_infos=A_ ) lowerCAmelCase__ : Optional[int] = TestCommand(*A_ ) test_command.run() lowerCAmelCase__ : int = os.path.join(A_ , '''README.md''' ) assert os.path.exists(A_ ) lowerCAmelCase__ : List[Any] = DatasetInfosDict.from_directory(A_ ) lowerCAmelCase__ : List[str] = DatasetInfosDict( { '''default''': DatasetInfo( features=Features( { '''tokens''': Sequence(Value('''string''' ) ), '''ner_tags''': Sequence( ClassLabel(names=['''O''', '''B-PER''', '''I-PER''', '''B-ORG''', '''I-ORG''', '''B-LOC''', '''I-LOC'''] ) ), '''langs''': Sequence(Value('''string''' ) ), '''spans''': Sequence(Value('''string''' ) ), } ) , splits=[ { '''name''': '''train''', '''num_bytes''': 2_35_15_63, '''num_examples''': 1_00_00, }, { '''name''': '''validation''', '''num_bytes''': 23_84_18, '''num_examples''': 10_00, }, ] , download_size=3_94_06_80 , dataset_size=2_58_99_81 , ) } ) assert dataset_infos.keys() == expected_dataset_infos.keys() for key in DatasetInfo._INCLUDED_INFO_IN_YAML: lowerCAmelCase__ ,lowerCAmelCase__ : List[Any] = getattr(dataset_infos['''default'''] , A_ ), getattr(expected_dataset_infos['''default'''] , A_ ) if key == "num_bytes": assert is_apercent_close(A_ , A_ ) elif key == "splits": assert list(A_ ) == list(A_ ) for split in result: assert result[split].name == expected[split].name assert result[split].num_examples == expected[split].num_examples assert is_apercent_close(result[split].num_bytes , expected[split].num_bytes ) else: result == expected
74
0
"""simple docstring""" def _A (__a , __a ) -> int: """simple docstring""" return int(input_a == input_a == 0 ) def _A () -> None: """simple docstring""" print('''Truth Table of NOR Gate:''' ) print('''| Input 1 | Input 2 | Output |''' ) print(f'| 0 | 0 | {nor_gate(0 , 0 )} |' ) print(f'| 0 | 1 | {nor_gate(0 , 1 )} |' ) print(f'| 1 | 0 | {nor_gate(1 , 0 )} |' ) print(f'| 1 | 1 | {nor_gate(1 , 1 )} |' ) if __name__ == "__main__": import doctest doctest.testmod() main()
91
"""simple docstring""" import unittest from diffusers.pipelines.pipeline_utils import is_safetensors_compatible class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def _SCREAMING_SNAKE_CASE ( self : str): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Optional[Any] = [ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertTrue(is_safetensors_compatible(lowercase_)) def _SCREAMING_SNAKE_CASE ( self : Optional[int]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Optional[Any] = [ '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertTrue(is_safetensors_compatible(lowercase_)) def _SCREAMING_SNAKE_CASE ( self : List[str]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Optional[int] = [ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', '''unet/diffusion_pytorch_model.bin''', # Removed: 'unet/diffusion_pytorch_model.safetensors', ] self.assertFalse(is_safetensors_compatible(lowercase_)) def _SCREAMING_SNAKE_CASE ( self : int): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[Any] = [ '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', ] self.assertTrue(is_safetensors_compatible(lowercase_)) def _SCREAMING_SNAKE_CASE ( self : Optional[int]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Optional[int] = [ '''safety_checker/pytorch_model.bin''', '''safety_checker/model.safetensors''', '''vae/diffusion_pytorch_model.bin''', '''vae/diffusion_pytorch_model.safetensors''', '''text_encoder/pytorch_model.bin''', # Removed: 'text_encoder/model.safetensors', '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] self.assertFalse(is_safetensors_compatible(lowercase_)) def _SCREAMING_SNAKE_CASE ( self : List[Any]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : str = [ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] SCREAMING_SNAKE_CASE_ : Any = '''fp16''' self.assertTrue(is_safetensors_compatible(lowercase_ , variant=lowercase_)) def _SCREAMING_SNAKE_CASE ( self : str): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Optional[int] = [ '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] SCREAMING_SNAKE_CASE_ : Dict = '''fp16''' self.assertTrue(is_safetensors_compatible(lowercase_ , variant=lowercase_)) def _SCREAMING_SNAKE_CASE ( self : int): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[str] = [ '''unet/diffusion_pytorch_model.bin''', '''unet/diffusion_pytorch_model.safetensors''', ] SCREAMING_SNAKE_CASE_ : Any = '''fp16''' self.assertTrue(is_safetensors_compatible(lowercase_ , variant=lowercase_)) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Tuple = [ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', '''unet/diffusion_pytorch_model.fp16.bin''', # Removed: 'unet/diffusion_pytorch_model.fp16.safetensors', ] SCREAMING_SNAKE_CASE_ : List[Any] = '''fp16''' self.assertFalse(is_safetensors_compatible(lowercase_ , variant=lowercase_)) def _SCREAMING_SNAKE_CASE ( self : List[Any]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : List[Any] = [ '''text_encoder/pytorch_model.fp16.bin''', '''text_encoder/model.fp16.safetensors''', ] SCREAMING_SNAKE_CASE_ : Any = '''fp16''' self.assertTrue(is_safetensors_compatible(lowercase_ , variant=lowercase_)) def _SCREAMING_SNAKE_CASE ( self : List[str]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Optional[Any] = [ '''text_encoder/pytorch_model.bin''', '''text_encoder/model.safetensors''', ] SCREAMING_SNAKE_CASE_ : List[Any] = '''fp16''' self.assertTrue(is_safetensors_compatible(lowercase_ , variant=lowercase_)) def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): '''simple docstring''' SCREAMING_SNAKE_CASE_ : Optional[int] = [ '''safety_checker/pytorch_model.fp16.bin''', '''safety_checker/model.fp16.safetensors''', '''vae/diffusion_pytorch_model.fp16.bin''', '''vae/diffusion_pytorch_model.fp16.safetensors''', '''text_encoder/pytorch_model.fp16.bin''', # 'text_encoder/model.fp16.safetensors', '''unet/diffusion_pytorch_model.fp16.bin''', '''unet/diffusion_pytorch_model.fp16.safetensors''', ] SCREAMING_SNAKE_CASE_ : str = '''fp16''' self.assertFalse(is_safetensors_compatible(lowercase_ , variant=lowercase_))
91
1
'''simple docstring''' from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, ) @flax.struct.dataclass class A__ ( _snake_case ): lowercase = 42 lowercase = 42 class A__ ( nn.Module ): lowercase = 42 lowercase = (16, 32, 96, 256) lowercase = jnp.floataa def snake_case_ ( self ) -> Any: '''simple docstring''' A_ = nn.Conv( self.block_out_channels[0] , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) A_ = [] for i in range(len(self.block_out_channels ) - 1 ): A_ = self.block_out_channels[i] A_ = self.block_out_channels[i + 1] A_ = nn.Conv( UpperCamelCase__ , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) blocks.append(UpperCamelCase__ ) A_ = nn.Conv( UpperCamelCase__ , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) blocks.append(UpperCamelCase__ ) A_ = blocks A_ = nn.Conv( self.conditioning_embedding_channels , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) def __call__( self , UpperCamelCase__ ) -> int: '''simple docstring''' A_ = self.conv_in(UpperCamelCase__ ) A_ = nn.silu(UpperCamelCase__ ) for block in self.blocks: A_ = block(UpperCamelCase__ ) A_ = nn.silu(UpperCamelCase__ ) A_ = self.conv_out(UpperCamelCase__ ) return embedding @flax_register_to_config class A__ ( nn.Module , _snake_case , _snake_case ): lowercase = 32 lowercase = 4 lowercase = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) lowercase = False lowercase = (320, 640, 1_280, 1_280) lowercase = 2 lowercase = 8 lowercase = None lowercase = 1_280 lowercase = 0.0 lowercase = False lowercase = jnp.floataa lowercase = True lowercase = 0 lowercase = "rgb" lowercase = (16, 32, 96, 256) def snake_case_ ( self , UpperCamelCase__ ) -> FrozenDict: '''simple docstring''' # init input tensors A_ = (1, self.in_channels, self.sample_size, self.sample_size) A_ = jnp.zeros(UpperCamelCase__ , dtype=jnp.floataa ) A_ = jnp.ones((1,) , dtype=jnp.intaa ) A_ = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) A_ = (1, 3, self.sample_size * 8, self.sample_size * 8) A_ = jnp.zeros(UpperCamelCase__ , dtype=jnp.floataa ) A_ , A_ = jax.random.split(UpperCamelCase__ ) A_ = {"""params""": params_rng, """dropout""": dropout_rng} return self.init(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ )["params"] def snake_case_ ( self ) -> List[str]: '''simple docstring''' A_ = self.block_out_channels A_ = block_out_channels[0] * 4 # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. A_ = self.num_attention_heads or self.attention_head_dim # input A_ = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time A_ = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) A_ = FlaxTimestepEmbedding(UpperCamelCase__ , dtype=self.dtype ) A_ = FlaxControlNetConditioningEmbedding( conditioning_embedding_channels=block_out_channels[0] , block_out_channels=self.conditioning_embedding_out_channels , ) A_ = self.only_cross_attention if isinstance(UpperCamelCase__ , UpperCamelCase__ ): A_ = (only_cross_attention,) * len(self.down_block_types ) if isinstance(UpperCamelCase__ , UpperCamelCase__ ): A_ = (num_attention_heads,) * len(self.down_block_types ) # down A_ = [] A_ = [] A_ = block_out_channels[0] A_ = nn.Conv( UpperCamelCase__ , kernel_size=(1, 1) , padding="""VALID""" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(UpperCamelCase__ ) for i, down_block_type in enumerate(self.down_block_types ): A_ = output_channel A_ = block_out_channels[i] A_ = i == len(UpperCamelCase__ ) - 1 if down_block_type == "CrossAttnDownBlock2D": A_ = FlaxCrossAttnDownBlockaD( in_channels=UpperCamelCase__ , out_channels=UpperCamelCase__ , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , dtype=self.dtype , ) else: A_ = FlaxDownBlockaD( in_channels=UpperCamelCase__ , out_channels=UpperCamelCase__ , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(UpperCamelCase__ ) for _ in range(self.layers_per_block ): A_ = nn.Conv( UpperCamelCase__ , kernel_size=(1, 1) , padding="""VALID""" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(UpperCamelCase__ ) if not is_final_block: A_ = nn.Conv( UpperCamelCase__ , kernel_size=(1, 1) , padding="""VALID""" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(UpperCamelCase__ ) A_ = down_blocks A_ = controlnet_down_blocks # mid A_ = block_out_channels[-1] A_ = FlaxUNetMidBlockaDCrossAttn( in_channels=UpperCamelCase__ , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , dtype=self.dtype , ) A_ = nn.Conv( UpperCamelCase__ , kernel_size=(1, 1) , padding="""VALID""" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) def __call__( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = 1.0 , UpperCamelCase__ = True , UpperCamelCase__ = False , ) -> Union[FlaxControlNetOutput, Tuple]: '''simple docstring''' A_ = self.controlnet_conditioning_channel_order if channel_order == "bgr": A_ = jnp.flip(UpperCamelCase__ , axis=1 ) # 1. time if not isinstance(UpperCamelCase__ , jnp.ndarray ): A_ = jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(UpperCamelCase__ , jnp.ndarray ) and len(timesteps.shape ) == 0: A_ = timesteps.astype(dtype=jnp.floataa ) A_ = jnp.expand_dims(UpperCamelCase__ , 0 ) A_ = self.time_proj(UpperCamelCase__ ) A_ = self.time_embedding(UpperCamelCase__ ) # 2. pre-process A_ = jnp.transpose(UpperCamelCase__ , (0, 2, 3, 1) ) A_ = self.conv_in(UpperCamelCase__ ) A_ = jnp.transpose(UpperCamelCase__ , (0, 2, 3, 1) ) A_ = self.controlnet_cond_embedding(UpperCamelCase__ ) sample += controlnet_cond # 3. down A_ = (sample,) for down_block in self.down_blocks: if isinstance(UpperCamelCase__ , UpperCamelCase__ ): A_ , A_ = down_block(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , deterministic=not train ) else: A_ , A_ = down_block(UpperCamelCase__ , UpperCamelCase__ , deterministic=not train ) down_block_res_samples += res_samples # 4. mid A_ = self.mid_block(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , deterministic=not train ) # 5. contronet blocks A_ = () for down_block_res_sample, controlnet_block in zip(UpperCamelCase__ , self.controlnet_down_blocks ): A_ = controlnet_block(UpperCamelCase__ ) controlnet_down_block_res_samples += (down_block_res_sample,) A_ = controlnet_down_block_res_samples A_ = self.controlnet_mid_block(UpperCamelCase__ ) # 6. scaling A_ = [sample * conditioning_scale for sample in down_block_res_samples] mid_block_res_sample *= conditioning_scale if not return_dict: return (down_block_res_samples, mid_block_res_sample) return FlaxControlNetOutput( down_block_res_samples=UpperCamelCase__ , mid_block_res_sample=UpperCamelCase__ )
101
'''simple docstring''' def UpperCAmelCase__ ( UpperCAmelCase__ ) -> float: if edge <= 0 or not isinstance(UpperCAmelCase__, UpperCAmelCase__ ): raise ValueError("""Length must be a positive.""" ) return 3 * ((25 + 10 * (5 ** (1 / 2))) ** (1 / 2)) * (edge**2) def UpperCAmelCase__ ( UpperCAmelCase__ ) -> float: if edge <= 0 or not isinstance(UpperCAmelCase__, UpperCAmelCase__ ): raise ValueError("""Length must be a positive.""" ) return ((15 + (7 * (5 ** (1 / 2)))) / 4) * (edge**3) if __name__ == "__main__": import doctest doctest.testmod()
101
1
"""simple docstring""" import re import string import numpy as np import datasets lowerCamelCase__ = """ Returns the rate at which the input predicted strings exactly match their references, ignoring any strings input as part of the regexes_to_ignore list. """ lowerCamelCase__ = """ Args: predictions: List of predicted texts. references: List of reference texts. regexes_to_ignore: List, defaults to None. Regex expressions of characters to ignore when calculating the exact matches. Note: these regexes are removed from the input data before the changes based on the options below (e.g. ignore_case, ignore_punctuation, ignore_numbers) are applied. ignore_case: Boolean, defaults to False. If true, turns everything to lowercase so that capitalization differences are ignored. ignore_punctuation: Boolean, defaults to False. If true, removes all punctuation before comparing predictions and references. ignore_numbers: Boolean, defaults to False. If true, removes all punctuation before comparing predictions and references. Returns: exact_match: Dictionary containing exact_match rate. Possible values are between 0.0 and 100.0, inclusive. Examples: >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"] >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"] >>> results = exact_match.compute(references=refs, predictions=preds) >>> print(round(results[\"exact_match\"], 1)) 25.0 >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"] >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\"], ignore_case=True, ignore_punctuation=True) >>> print(round(results[\"exact_match\"], 1)) 50.0 >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"] >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\", \"YELL\"], ignore_case=True, ignore_punctuation=True) >>> print(round(results[\"exact_match\"], 1)) 75.0 >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"the cat\", \"theater\", \"YELLING\", \"agent007\"] >>> preds = [\"cat?\", \"theater\", \"yelling\", \"agent\"] >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=[\"the \", \"yell\", \"YELL\"], ignore_case=True, ignore_punctuation=True, ignore_numbers=True) >>> print(round(results[\"exact_match\"], 1)) 100.0 >>> exact_match = datasets.load_metric(\"exact_match\") >>> refs = [\"The cat sat on the mat.\", \"Theaters are great.\", \"It's like comparing oranges and apples.\"] >>> preds = [\"The cat sat on the mat?\", \"Theaters are great.\", \"It's like comparing apples and oranges.\"] >>> results = exact_match.compute(references=refs, predictions=preds) >>> print(round(results[\"exact_match\"], 1)) 33.3 """ lowerCamelCase__ = """ """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class A__ ( datasets.Metric): def __lowerCamelCase ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , reference_urls=[] , ) def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=False , ): if regexes_to_ignore is not None: for s in regexes_to_ignore: __lowerCAmelCase : int = np.array([re.sub(_SCREAMING_SNAKE_CASE , '' , _SCREAMING_SNAKE_CASE ) for x in predictions] ) __lowerCAmelCase : List[str] = np.array([re.sub(_SCREAMING_SNAKE_CASE , '' , _SCREAMING_SNAKE_CASE ) for x in references] ) else: __lowerCAmelCase : Union[str, Any] = np.asarray(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[Any] = np.asarray(_SCREAMING_SNAKE_CASE ) if ignore_case: __lowerCAmelCase : str = np.char.lower(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[str] = np.char.lower(_SCREAMING_SNAKE_CASE ) if ignore_punctuation: __lowerCAmelCase : Dict = string.punctuation.maketrans('' , '' , string.punctuation ) __lowerCAmelCase : int = np.char.translate(_SCREAMING_SNAKE_CASE , table=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Tuple = np.char.translate(_SCREAMING_SNAKE_CASE , table=_SCREAMING_SNAKE_CASE ) if ignore_numbers: __lowerCAmelCase : int = string.digits.maketrans('' , '' , string.digits ) __lowerCAmelCase : List[str] = np.char.translate(_SCREAMING_SNAKE_CASE , table=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[str] = np.char.translate(_SCREAMING_SNAKE_CASE , table=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : int = predictions == references return {"exact_match": np.mean(_SCREAMING_SNAKE_CASE ) * 1_00}
86
"""simple docstring""" import tempfile import unittest import numpy as np import transformers from transformers import GPTaTokenizer, GPTJConfig, is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax, tooslow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax import jax.numpy as jnp from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) from transformers.models.gptj.modeling_flax_gptj import FlaxGPTJForCausalLM, FlaxGPTJModel if is_torch_available(): import torch class A__ : def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=14 , _SCREAMING_SNAKE_CASE=7 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=99 , _SCREAMING_SNAKE_CASE=32 , _SCREAMING_SNAKE_CASE=4 , _SCREAMING_SNAKE_CASE=4 , _SCREAMING_SNAKE_CASE=4 , _SCREAMING_SNAKE_CASE=37 , _SCREAMING_SNAKE_CASE="gelu" , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=0.1 , _SCREAMING_SNAKE_CASE=5_12 , _SCREAMING_SNAKE_CASE=0.02 , ): __lowerCAmelCase : Union[str, Any] = parent __lowerCAmelCase : Any = batch_size __lowerCAmelCase : Any = seq_length __lowerCAmelCase : Optional[Any] = is_training __lowerCAmelCase : Any = use_input_mask __lowerCAmelCase : Any = use_token_type_ids __lowerCAmelCase : Tuple = use_labels __lowerCAmelCase : Optional[Any] = vocab_size __lowerCAmelCase : Tuple = hidden_size __lowerCAmelCase : str = rotary_dim __lowerCAmelCase : Union[str, Any] = num_hidden_layers __lowerCAmelCase : Union[str, Any] = num_attention_heads __lowerCAmelCase : int = intermediate_size __lowerCAmelCase : List[str] = hidden_act __lowerCAmelCase : int = hidden_dropout_prob __lowerCAmelCase : Any = attention_probs_dropout_prob __lowerCAmelCase : List[Any] = max_position_embeddings __lowerCAmelCase : Optional[Any] = initializer_range __lowerCAmelCase : Tuple = None __lowerCAmelCase : int = vocab_size - 1 __lowerCAmelCase : Dict = vocab_size - 1 __lowerCAmelCase : int = vocab_size - 1 def __lowerCamelCase ( self ): __lowerCAmelCase : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowerCAmelCase : List[str] = None if self.use_input_mask: __lowerCAmelCase : Dict = random_attention_mask([self.batch_size, self.seq_length] ) __lowerCAmelCase : Optional[int] = GPTJConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , use_cache=_SCREAMING_SNAKE_CASE , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , rotary_dim=self.rotary_dim , ) return (config, input_ids, input_mask) def __lowerCamelCase ( self ): __lowerCAmelCase : Optional[int] = self.prepare_config_and_inputs() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Any = config_and_inputs __lowerCAmelCase : Dict = {'input_ids': input_ids, 'attention_mask': attention_mask} return config, inputs_dict def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): __lowerCAmelCase : List[str] = 20 __lowerCAmelCase : List[str] = model_class_name(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[int] = model.init_cache(input_ids.shape[0] , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[Any] = jnp.ones((input_ids.shape[0], max_decoder_length) , dtype='i4' ) __lowerCAmelCase : Optional[Any] = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) __lowerCAmelCase : Any = model( input_ids[:, :-1] , attention_mask=_SCREAMING_SNAKE_CASE , past_key_values=_SCREAMING_SNAKE_CASE , position_ids=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Any = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype='i4' ) __lowerCAmelCase : int = model( input_ids[:, -1:] , attention_mask=_SCREAMING_SNAKE_CASE , past_key_values=outputs_cache.past_key_values , position_ids=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Any = model(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[str] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f"Max diff is {diff}" ) def __lowerCamelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): __lowerCAmelCase : Tuple = 20 __lowerCAmelCase : List[str] = model_class_name(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Union[str, Any] = jnp.concatenate( [attention_mask, jnp.zeros((attention_mask.shape[0], max_decoder_length - attention_mask.shape[1]) )] , axis=-1 , ) __lowerCAmelCase : List[str] = model.init_cache(input_ids.shape[0] , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : str = jnp.broadcast_to( jnp.arange(input_ids.shape[-1] - 1 )[None, :] , (input_ids.shape[0], input_ids.shape[-1] - 1) ) __lowerCAmelCase : Optional[Any] = model( input_ids[:, :-1] , attention_mask=_SCREAMING_SNAKE_CASE , past_key_values=_SCREAMING_SNAKE_CASE , position_ids=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : str = jnp.array(input_ids.shape[0] * [[input_ids.shape[-1] - 1]] , dtype='i4' ) __lowerCAmelCase : Tuple = model( input_ids[:, -1:] , past_key_values=outputs_cache.past_key_values , attention_mask=_SCREAMING_SNAKE_CASE , position_ids=_SCREAMING_SNAKE_CASE , ) __lowerCAmelCase : Union[str, Any] = model(_SCREAMING_SNAKE_CASE , attention_mask=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[str] = np.max(np.abs((outputs_cache_next[0][:, -1, :5] - outputs[0][:, -1, :5]) ) ) self.parent.assertTrue(diff < 1E-3 , msg=f"Max diff is {diff}" ) @require_flax class A__ ( _lowerCamelCase , _lowerCamelCase , unittest.TestCase): A_ : Tuple = (FlaxGPTJModel, FlaxGPTJForCausalLM) if is_flax_available() else () A_ : str = (FlaxGPTJForCausalLM,) if is_flax_available() else () def __lowerCamelCase ( self ): __lowerCAmelCase : int = FlaxGPTJModelTester(self ) def __lowerCamelCase ( self ): for model_class_name in self.all_model_classes: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) def __lowerCamelCase ( self ): for model_class_name in self.all_model_classes: __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.check_use_cache_forward_with_attn_mask( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @tooslow def __lowerCamelCase ( self ): __lowerCAmelCase : Tuple = GPTaTokenizer.from_pretrained('gpt2' , pad_token='<|endoftext|>' , padding_side='left' ) __lowerCAmelCase : Optional[int] = tokenizer(['Hello this is a long string', 'Hey'] , return_tensors='np' , padding=_SCREAMING_SNAKE_CASE , truncation=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : int = FlaxGPTJForCausalLM.from_pretrained('EleutherAI/gpt-j-6B' ) __lowerCAmelCase : Any = False __lowerCAmelCase : Any = model.config.eos_token_id __lowerCAmelCase : Union[str, Any] = jax.jit(model.generate ) __lowerCAmelCase : Optional[Any] = jit_generate( inputs['input_ids'] , attention_mask=inputs['attention_mask'] , pad_token_id=tokenizer.pad_token_id ).sequences __lowerCAmelCase : str = tokenizer.batch_decode(_SCREAMING_SNAKE_CASE , skip_special_tokens=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[Any] = [ 'Hello this is a long string of text.\n\nI\'m trying to get the text of the', 'Hey, I\'m a little late to the party. I\'m going to', ] self.assertListEqual(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) @is_pt_flax_cross_test def __lowerCamelCase ( self ): __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs __lowerCAmelCase : str = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Union[str, Any] = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class __lowerCAmelCase : Dict = model_class.__name__[4:] # Skip the "Flax" at the beginning __lowerCAmelCase : Optional[int] = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase , __lowerCAmelCase : List[Any] = pt_inputs['input_ids'].shape __lowerCAmelCase : int = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_SCREAMING_SNAKE_CASE ): __lowerCAmelCase : Tuple = 0 __lowerCAmelCase : Tuple = 1 __lowerCAmelCase : List[str] = 0 __lowerCAmelCase : Any = 1 __lowerCAmelCase : Optional[Any] = pt_model_class(_SCREAMING_SNAKE_CASE ).eval() __lowerCAmelCase : Any = model_class(_SCREAMING_SNAKE_CASE , dtype=jnp.floataa ) __lowerCAmelCase : int = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : str = fx_state with torch.no_grad(): __lowerCAmelCase : Union[str, Any] = pt_model(**_SCREAMING_SNAKE_CASE ).to_tuple() __lowerCAmelCase : str = fx_model(**_SCREAMING_SNAKE_CASE ).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , len(_SCREAMING_SNAKE_CASE ) , 'Output lengths differ between Flax and PyTorch' ) for fx_output, pt_output in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Tuple = model_class.from_pretrained(_SCREAMING_SNAKE_CASE , from_pt=_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[Any] = fx_model_loaded(**_SCREAMING_SNAKE_CASE ).to_tuple() self.assertEqual( len(_SCREAMING_SNAKE_CASE ) , len(_SCREAMING_SNAKE_CASE ) , 'Output lengths differ between Flax and PyTorch' ) for fx_output_loaded, pt_output in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): self.assert_almost_equals(fx_output_loaded[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @is_pt_flax_cross_test def __lowerCamelCase ( self ): __lowerCAmelCase , __lowerCAmelCase : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): # prepare inputs __lowerCAmelCase : List[str] = self._prepare_for_class(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : List[str] = {k: torch.tensor(v.tolist() ) for k, v in prepared_inputs_dict.items()} # load corresponding PyTorch class __lowerCAmelCase : Dict = model_class.__name__[4:] # Skip the "Flax" at the beginning __lowerCAmelCase : str = getattr(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Optional[int] = pt_model_class(_SCREAMING_SNAKE_CASE ).eval() __lowerCAmelCase : Tuple = model_class(_SCREAMING_SNAKE_CASE , dtype=jnp.floataa ) __lowerCAmelCase : List[str] = load_flax_weights_in_pytorch_model(_SCREAMING_SNAKE_CASE , fx_model.params ) __lowerCAmelCase , __lowerCAmelCase : int = pt_inputs['input_ids'].shape __lowerCAmelCase : List[str] = np.random.randint(0 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(_SCREAMING_SNAKE_CASE ): __lowerCAmelCase : str = 0 __lowerCAmelCase : Optional[Any] = 1 __lowerCAmelCase : Optional[int] = 0 __lowerCAmelCase : Optional[Any] = 1 # make sure weights are tied in PyTorch pt_model.tie_weights() with torch.no_grad(): __lowerCAmelCase : List[str] = pt_model(**_SCREAMING_SNAKE_CASE ).to_tuple() __lowerCAmelCase : Optional[int] = fx_model(**_SCREAMING_SNAKE_CASE ).to_tuple() self.assertEqual(len(_SCREAMING_SNAKE_CASE ) , len(_SCREAMING_SNAKE_CASE ) , 'Output lengths differ between Flax and PyTorch' ) for fx_output, pt_output in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(_SCREAMING_SNAKE_CASE ) __lowerCAmelCase : Dict = pt_model_class.from_pretrained(_SCREAMING_SNAKE_CASE , from_flax=_SCREAMING_SNAKE_CASE ) with torch.no_grad(): __lowerCAmelCase : Any = pt_model_loaded(**_SCREAMING_SNAKE_CASE ).to_tuple() self.assertEqual( len(_SCREAMING_SNAKE_CASE ) , len(_SCREAMING_SNAKE_CASE ) , 'Output lengths differ between Flax and PyTorch' ) for fx_output, pt_output in zip(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): self.assert_almost_equals(fx_output[:, -1] , pt_output[:, -1].numpy() , 4E-2 ) @tooslow def __lowerCamelCase ( self ): for model_class_name in self.all_model_classes: __lowerCAmelCase : Optional[int] = model_class_name.from_pretrained('EleutherAI/gpt-j-6B' ) __lowerCAmelCase : List[Any] = model(np.ones((1, 1) ) ) self.assertIsNotNone(_SCREAMING_SNAKE_CASE )
86
1
'''simple docstring''' def __UpperCamelCase ( _UpperCAmelCase, _UpperCAmelCase ): __UpperCAmelCase : Union[str, Any] = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def __UpperCamelCase ( _UpperCAmelCase, _UpperCAmelCase, _UpperCAmelCase ): __UpperCAmelCase : Any = 0 while b > 0: if b & 1: __UpperCAmelCase : Union[str, Any] = ((res % c) + (a % c)) % c a += a b >>= 1 return res
353
'''simple docstring''' import unittest from transformers import PegasusTokenizer, PegasusTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin lowerCAmelCase__ : Any = get_tests_dir("fixtures/test_sentencepiece_no_bos.model") @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( snake_case__ ,unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE = PegasusTokenizer SCREAMING_SNAKE_CASE = PegasusTokenizerFast SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True def lowerCamelCase_ ( self : int ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __UpperCAmelCase : Tuple = PegasusTokenizer(UpperCAmelCase_ ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowerCamelCase_ ( self : Dict ): """simple docstring""" return PegasusTokenizer.from_pretrained("google/pegasus-large" ) def lowerCamelCase_ ( self : List[Any] , **UpperCAmelCase_ : List[str] ): """simple docstring""" return PegasusTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase_ ) def lowerCamelCase_ ( self : str , UpperCAmelCase_ : int ): """simple docstring""" return ("This is a test", "This is a test") def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" __UpperCAmelCase : List[str] = "</s>" __UpperCAmelCase : Union[str, Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase_ ) , UpperCAmelCase_ ) def lowerCamelCase_ ( self : Any ): """simple docstring""" __UpperCAmelCase : int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<pad>" ) self.assertEqual(vocab_keys[1] , "</s>" ) self.assertEqual(vocab_keys[-1] , "v" ) self.assertEqual(len(UpperCAmelCase_ ) , 1_103 ) def lowerCamelCase_ ( self : Tuple ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_103 ) def lowerCamelCase_ ( self : str ): """simple docstring""" __UpperCAmelCase : str = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) __UpperCAmelCase : int = self.tokenizer_class.from_pretrained(self.tmpdirname ) __UpperCAmelCase : Tuple = ( "Let's see which <unk> is the better <unk_token_11> one <mask_1> It seems like this <mask_2> was important" " </s> <pad> <pad> <pad>" ) __UpperCAmelCase : List[str] = rust_tokenizer([raw_input_str] , return_tensors=UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ).input_ids[0] __UpperCAmelCase : int = py_tokenizer([raw_input_str] , return_tensors=UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ).input_ids[0] self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def lowerCamelCase_ ( self : Dict ): """simple docstring""" __UpperCAmelCase : Any = self._large_tokenizer # <mask_1> masks whole sentence while <mask_2> masks single word __UpperCAmelCase : Tuple = "<mask_1> To ensure a <mask_2> flow of bank resolutions." __UpperCAmelCase : Optional[Any] = [2, 413, 615, 114, 3, 1_971, 113, 1_679, 10_710, 107, 1] __UpperCAmelCase : Optional[Any] = tokenizer([raw_input_str] , return_tensors=UpperCAmelCase_ ).input_ids[0] self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) def lowerCamelCase_ ( self : Any ): """simple docstring""" __UpperCAmelCase : Dict = self._large_tokenizer # The tracebacks for the following asserts are **better** without messages or self.assertEqual assert tokenizer.vocab_size == 96_103 assert tokenizer.pad_token_id == 0 assert tokenizer.eos_token_id == 1 assert tokenizer.offset == 103 assert tokenizer.unk_token_id == tokenizer.offset + 2 == 105 assert tokenizer.unk_token == "<unk>" assert tokenizer.model_max_length == 1_024 __UpperCAmelCase : Tuple = "To ensure a smooth flow of bank resolutions." __UpperCAmelCase : str = [413, 615, 114, 2_291, 1_971, 113, 1_679, 10_710, 107, 1] __UpperCAmelCase : Union[str, Any] = tokenizer([raw_input_str] , return_tensors=UpperCAmelCase_ ).input_ids[0] self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) assert tokenizer.convert_ids_to_tokens([0, 1, 2, 3] ) == ["<pad>", "</s>", "<mask_1>", "<mask_2>"] @require_torch def lowerCamelCase_ ( self : str ): """simple docstring""" __UpperCAmelCase : List[Any] = ["This is going to be way too long." * 150, "short example"] __UpperCAmelCase : Optional[int] = ["not super long but more than 5 tokens", "tiny"] __UpperCAmelCase : str = self._large_tokenizer(UpperCAmelCase_ , padding=UpperCAmelCase_ , truncation=UpperCAmelCase_ , return_tensors="pt" ) __UpperCAmelCase : Union[str, Any] = self._large_tokenizer( text_target=UpperCAmelCase_ , max_length=5 , padding=UpperCAmelCase_ , truncation=UpperCAmelCase_ , return_tensors="pt" ) assert batch.input_ids.shape == (2, 1_024) assert batch.attention_mask.shape == (2, 1_024) assert targets["input_ids"].shape == (2, 5) assert len(UpperCAmelCase_ ) == 2 # input_ids, attention_mask. @slow def lowerCamelCase_ ( self : Any ): """simple docstring""" # fmt: off __UpperCAmelCase : Tuple = {"input_ids": [[38_979, 143, 18_485, 606, 130, 26_669, 87_686, 121, 54_189, 1_129, 111, 26_669, 87_686, 121, 9_114, 14_787, 121, 13_249, 158, 592, 956, 121, 14_621, 31_576, 143, 62_613, 108, 9_688, 930, 43_430, 11_562, 62_613, 304, 108, 11_443, 897, 108, 9_314, 17_415, 63_399, 108, 11_443, 7_614, 18_316, 118, 4_284, 7_148, 12_430, 143, 1_400, 25_703, 158, 111, 4_284, 7_148, 11_772, 143, 21_297, 1_064, 158, 122, 204, 3_506, 1_754, 1_133, 14_787, 1_581, 115, 33_224, 4_482, 111, 1_355, 110, 29_173, 317, 50_833, 108, 20_147, 94_665, 111, 77_198, 107, 1], [110, 62_613, 117, 638, 112, 1_133, 121, 20_098, 1_355, 79_050, 13_872, 135, 1_596, 53_541, 1_352, 141, 13_039, 5_542, 124, 302, 518, 111, 268, 2_956, 115, 149, 4_427, 107, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [139, 1_235, 2_799, 18_289, 17_780, 204, 109, 9_474, 1_296, 107, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase_ , model_name="google/bigbird-pegasus-large-arxiv" , revision="ba85d0851d708441f91440d509690f1ab6353415" , ) @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( snake_case__ ,unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE = PegasusTokenizer SCREAMING_SNAKE_CASE = PegasusTokenizerFast SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = True def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __UpperCAmelCase : List[str] = PegasusTokenizer(UpperCAmelCase_ , offset=0 , mask_token_sent=UpperCAmelCase_ , mask_token="[MASK]" ) tokenizer.save_pretrained(self.tmpdirname ) @cached_property def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" return PegasusTokenizer.from_pretrained("google/bigbird-pegasus-large-arxiv" ) def lowerCamelCase_ ( self : Union[str, Any] , **UpperCAmelCase_ : int ): """simple docstring""" return PegasusTokenizer.from_pretrained(self.tmpdirname , **UpperCAmelCase_ ) def lowerCamelCase_ ( self : str , UpperCAmelCase_ : Union[str, Any] ): """simple docstring""" return ("This is a test", "This is a test") def lowerCamelCase_ ( self : Optional[Any] ): """simple docstring""" __UpperCAmelCase : Optional[Any] = self.rust_tokenizer_class.from_pretrained(self.tmpdirname ) __UpperCAmelCase : Optional[Any] = self.tokenizer_class.from_pretrained(self.tmpdirname ) __UpperCAmelCase : List[str] = ( "Let's see which <unk> is the better <unk_token> one [MASK] It seems like this [MASK] was important </s>" " <pad> <pad> <pad>" ) __UpperCAmelCase : str = rust_tokenizer([raw_input_str] , return_tensors=UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ).input_ids[0] __UpperCAmelCase : int = py_tokenizer([raw_input_str] , return_tensors=UpperCAmelCase_ , add_special_tokens=UpperCAmelCase_ ).input_ids[0] self.assertListEqual(UpperCAmelCase_ , UpperCAmelCase_ ) @require_torch def lowerCamelCase_ ( self : List[Any] ): """simple docstring""" __UpperCAmelCase : Any = ["This is going to be way too long." * 1_000, "short example"] __UpperCAmelCase : List[Any] = ["not super long but more than 5 tokens", "tiny"] __UpperCAmelCase : int = self._large_tokenizer(UpperCAmelCase_ , padding=UpperCAmelCase_ , truncation=UpperCAmelCase_ , return_tensors="pt" ) __UpperCAmelCase : List[Any] = self._large_tokenizer( text_target=UpperCAmelCase_ , max_length=5 , padding=UpperCAmelCase_ , truncation=UpperCAmelCase_ , return_tensors="pt" ) assert batch.input_ids.shape == (2, 4_096) assert batch.attention_mask.shape == (2, 4_096) assert targets["input_ids"].shape == (2, 5) assert len(UpperCAmelCase_ ) == 2 # input_ids, attention_mask. def lowerCamelCase_ ( self : Tuple ): """simple docstring""" __UpperCAmelCase : List[Any] = ( "This is an example string that is used to test the original TF implementation against the HF" " implementation" ) __UpperCAmelCase : int = self._large_tokenizer(UpperCAmelCase_ ).input_ids self.assertListEqual( UpperCAmelCase_ , [182, 117, 142, 587, 4_211, 120, 117, 263, 112, 804, 109, 856, 25_016, 3_137, 464, 109, 26_955, 3_137, 1] , )
37
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available UpperCamelCase_ = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = ['''GPTSw3Tokenizer'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_gpt_swa import GPTSwaTokenizer else: import sys UpperCamelCase_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
345
import random from typing import Any def lowerCamelCase_ ( _a : list ): '''simple docstring''' for _ in range(len(_a ) ): UpperCAmelCase_ : Tuple = random.randint(0 , len(_a ) - 1 ) UpperCAmelCase_ : List[Any] = random.randint(0 , len(_a ) - 1 ) UpperCAmelCase_ , UpperCAmelCase_ : int = data[b], data[a] return data if __name__ == "__main__": UpperCamelCase_ = [0, 1, 2, 3, 4, 5, 6, 7] UpperCamelCase_ = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
345
1
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available lowerCamelCase__ = { """configuration_efficientnet""": [ """EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """EfficientNetConfig""", """EfficientNetOnnxConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = ["""EfficientNetImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ = [ """EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """EfficientNetForImageClassification""", """EfficientNetModel""", """EfficientNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_efficientnet import ( EFFICIENTNET_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientNetConfig, EfficientNetOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientnet import EfficientNetImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientnet import ( EFFICIENTNET_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientNetForImageClassification, EfficientNetModel, EfficientNetPreTrainedModel, ) else: import sys lowerCamelCase__ = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
352
import numpy as np from cva import COLOR_BGR2GRAY, cvtColor, imread from numpy import array, uinta from PIL import Image from digital_image_processing import change_contrast as cc from digital_image_processing import convert_to_negative as cn from digital_image_processing import sepia as sp from digital_image_processing.dithering import burkes as bs from digital_image_processing.edge_detection import canny from digital_image_processing.filters import convolve as conv from digital_image_processing.filters import gaussian_filter as gg from digital_image_processing.filters import local_binary_pattern as lbp from digital_image_processing.filters import median_filter as med from digital_image_processing.filters import sobel_filter as sob from digital_image_processing.resize import resize as rs lowerCamelCase__ = imread(r"""digital_image_processing/image_data/lena_small.jpg""") lowerCamelCase__ = cvtColor(img, COLOR_BGR2GRAY) def lowerCAmelCase__ ( ) -> Dict: lowerCAmelCase__ : List[Any] = cn.convert_to_negative(SCREAMING_SNAKE_CASE_ ) # assert negative_img array for at least one True assert negative_img.any() def lowerCAmelCase__ ( ) -> Optional[Any]: with Image.open('digital_image_processing/image_data/lena_small.jpg' ) as img: # Work around assertion for response assert str(cc.change_contrast(SCREAMING_SNAKE_CASE_ , 110 ) ).startswith( '<PIL.Image.Image image mode=RGB size=100x100 at' ) def lowerCAmelCase__ ( ) -> Tuple: lowerCAmelCase__ : str = canny.gen_gaussian_kernel(9 , sigma=1.4 ) # Assert ambiguous array assert resp.all() def lowerCAmelCase__ ( ) -> Tuple: lowerCAmelCase__ : Tuple = imread('digital_image_processing/image_data/lena_small.jpg' , 0 ) # assert ambiguous array for all == True assert canny_img.all() lowerCAmelCase__ : Optional[Any] = canny.canny(SCREAMING_SNAKE_CASE_ ) # assert canny array for at least one True assert canny_array.any() def lowerCAmelCase__ ( ) -> Optional[int]: assert gg.gaussian_filter(SCREAMING_SNAKE_CASE_ , 5 , sigma=0.9 ).all() def lowerCAmelCase__ ( ) -> Dict: # laplace diagonals lowerCAmelCase__ : Union[str, Any] = array([[0.25, 0.5, 0.25], [0.5, -3, 0.5], [0.25, 0.5, 0.25]] ) lowerCAmelCase__ : int = conv.img_convolve(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ).astype(SCREAMING_SNAKE_CASE_ ) assert res.any() def lowerCAmelCase__ ( ) -> List[str]: assert med.median_filter(SCREAMING_SNAKE_CASE_ , 3 ).any() def lowerCAmelCase__ ( ) -> Any: lowerCAmelCase__ , lowerCAmelCase__ : str = sob.sobel_filter(SCREAMING_SNAKE_CASE_ ) assert grad.any() and theta.any() def lowerCAmelCase__ ( ) -> Any: lowerCAmelCase__ : int = sp.make_sepia(SCREAMING_SNAKE_CASE_ , 20 ) assert sepia.all() def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = "digital_image_processing/image_data/lena_small.jpg" ) -> Optional[Any]: lowerCAmelCase__ : List[Any] = bs.Burkes(imread(SCREAMING_SNAKE_CASE_ , 1 ) , 120 ) burkes.process() assert burkes.output_img.any() def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_ = "digital_image_processing/image_data/lena_small.jpg" , ) -> Any: lowerCAmelCase__ : Dict = rs.NearestNeighbour(imread(SCREAMING_SNAKE_CASE_ , 1 ) , 400 , 200 ) nn.process() assert nn.output.any() def lowerCAmelCase__ ( ) -> int: lowerCAmelCase__ : int = 'digital_image_processing/image_data/lena.jpg' # Reading the image and converting it to grayscale. lowerCAmelCase__ : List[str] = imread(SCREAMING_SNAKE_CASE_ , 0 ) # Test for get_neighbors_pixel function() return not None lowerCAmelCase__ : str = 0 lowerCAmelCase__ : str = 0 lowerCAmelCase__ : List[str] = image[x_coordinate][y_coordinate] lowerCAmelCase__ : Dict = lbp.get_neighbors_pixel( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) assert neighbors_pixels is not None # Test for local_binary_pattern function() # Create a numpy array as the same height and width of read image lowerCAmelCase__ : List[str] = np.zeros((image.shape[0], image.shape[1]) ) # Iterating through the image and calculating the local binary pattern value # for each pixel. for i in range(0 , image.shape[0] ): for j in range(0 , image.shape[1] ): lowerCAmelCase__ : Dict = lbp.local_binary_value(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) assert lbp_image.any()
307
0
import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def A_ ( _UpperCAmelCase , _UpperCAmelCase ): SCREAMING_SNAKE_CASE_: str = checkpoint SCREAMING_SNAKE_CASE_: Union[str, Any] = {} SCREAMING_SNAKE_CASE_: Any = vae_state_dict["encoder.conv_in.weight"] SCREAMING_SNAKE_CASE_: Tuple = vae_state_dict["encoder.conv_in.bias"] SCREAMING_SNAKE_CASE_: Dict = vae_state_dict["encoder.conv_out.weight"] SCREAMING_SNAKE_CASE_: Optional[Any] = vae_state_dict["encoder.conv_out.bias"] SCREAMING_SNAKE_CASE_: Any = vae_state_dict["encoder.norm_out.weight"] SCREAMING_SNAKE_CASE_: Dict = vae_state_dict["encoder.norm_out.bias"] SCREAMING_SNAKE_CASE_: Tuple = vae_state_dict["decoder.conv_in.weight"] SCREAMING_SNAKE_CASE_: Union[str, Any] = vae_state_dict["decoder.conv_in.bias"] SCREAMING_SNAKE_CASE_: int = vae_state_dict["decoder.conv_out.weight"] SCREAMING_SNAKE_CASE_: Optional[int] = vae_state_dict["decoder.conv_out.bias"] SCREAMING_SNAKE_CASE_: Any = vae_state_dict["decoder.norm_out.weight"] SCREAMING_SNAKE_CASE_: Optional[int] = vae_state_dict["decoder.norm_out.bias"] SCREAMING_SNAKE_CASE_: List[Any] = vae_state_dict["quant_conv.weight"] SCREAMING_SNAKE_CASE_: int = vae_state_dict["quant_conv.bias"] SCREAMING_SNAKE_CASE_: List[Any] = vae_state_dict["post_quant_conv.weight"] SCREAMING_SNAKE_CASE_: Optional[int] = vae_state_dict["post_quant_conv.bias"] # Retrieves the keys for the encoder down blocks only SCREAMING_SNAKE_CASE_: int = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "encoder.down" in layer} ) SCREAMING_SNAKE_CASE_: Any = { layer_id: [key for key in vae_state_dict if f"down.{layer_id}" in key] for layer_id in range(_UpperCAmelCase ) } # Retrieves the keys for the decoder up blocks only SCREAMING_SNAKE_CASE_: Optional[int] = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "decoder.up" in layer} ) SCREAMING_SNAKE_CASE_: List[Any] = { layer_id: [key for key in vae_state_dict if f"up.{layer_id}" in key] for layer_id in range(_UpperCAmelCase ) } for i in range(_UpperCAmelCase ): SCREAMING_SNAKE_CASE_: List[Any] = [key for key in down_blocks[i] if f"down.{i}" in key and f"down.{i}.downsample" not in key] if f"encoder.down.{i}.downsample.conv.weight" in vae_state_dict: SCREAMING_SNAKE_CASE_: str = vae_state_dict.pop( f"encoder.down.{i}.downsample.conv.weight" ) SCREAMING_SNAKE_CASE_: Dict = vae_state_dict.pop( f"encoder.down.{i}.downsample.conv.bias" ) SCREAMING_SNAKE_CASE_: Dict = renew_vae_resnet_paths(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: str = {"old": f"down.{i}.block", "new": f"down_blocks.{i}.resnets"} assign_to_checkpoint(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , additional_replacements=[meta_path] , config=_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] = [key for key in vae_state_dict if "encoder.mid.block" in key] SCREAMING_SNAKE_CASE_: str = 2 for i in range(1 , num_mid_res_blocks + 1 ): SCREAMING_SNAKE_CASE_: Any = [key for key in mid_resnets if f"encoder.mid.block_{i}" in key] SCREAMING_SNAKE_CASE_: List[Any] = renew_vae_resnet_paths(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: int = {"old": f"mid.block_{i}", "new": f"mid_block.resnets.{i - 1}"} assign_to_checkpoint(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , additional_replacements=[meta_path] , config=_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: List[Any] = [key for key in vae_state_dict if "encoder.mid.attn" in key] SCREAMING_SNAKE_CASE_: List[str] = renew_vae_attention_paths(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[Any] = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , additional_replacements=[meta_path] , config=_UpperCAmelCase ) conv_attn_to_linear(_UpperCAmelCase ) for i in range(_UpperCAmelCase ): SCREAMING_SNAKE_CASE_: Union[str, Any] = num_up_blocks - 1 - i SCREAMING_SNAKE_CASE_: Union[str, Any] = [ key for key in up_blocks[block_id] if f"up.{block_id}" in key and f"up.{block_id}.upsample" not in key ] if f"decoder.up.{block_id}.upsample.conv.weight" in vae_state_dict: SCREAMING_SNAKE_CASE_: Dict = vae_state_dict[ f"decoder.up.{block_id}.upsample.conv.weight" ] SCREAMING_SNAKE_CASE_: Optional[Any] = vae_state_dict[ f"decoder.up.{block_id}.upsample.conv.bias" ] SCREAMING_SNAKE_CASE_: Optional[Any] = renew_vae_resnet_paths(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: List[str] = {"old": f"up.{block_id}.block", "new": f"up_blocks.{i}.resnets"} assign_to_checkpoint(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , additional_replacements=[meta_path] , config=_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: List[str] = [key for key in vae_state_dict if "decoder.mid.block" in key] SCREAMING_SNAKE_CASE_: Tuple = 2 for i in range(1 , num_mid_res_blocks + 1 ): SCREAMING_SNAKE_CASE_: Dict = [key for key in mid_resnets if f"decoder.mid.block_{i}" in key] SCREAMING_SNAKE_CASE_: Any = renew_vae_resnet_paths(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[Any] = {"old": f"mid.block_{i}", "new": f"mid_block.resnets.{i - 1}"} assign_to_checkpoint(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , additional_replacements=[meta_path] , config=_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Tuple = [key for key in vae_state_dict if "decoder.mid.attn" in key] SCREAMING_SNAKE_CASE_: Any = renew_vae_attention_paths(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[Any] = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , additional_replacements=[meta_path] , config=_UpperCAmelCase ) conv_attn_to_linear(_UpperCAmelCase ) return new_checkpoint def A_ ( _UpperCAmelCase , _UpperCAmelCase , ): # Only support V1 SCREAMING_SNAKE_CASE_: Tuple = requests.get( " https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml" ) SCREAMING_SNAKE_CASE_: Optional[int] = io.BytesIO(r.content ) SCREAMING_SNAKE_CASE_: Any = OmegaConf.load(_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[Any] = 5_12 SCREAMING_SNAKE_CASE_: Tuple = "cuda" if torch.cuda.is_available() else "cpu" if checkpoint_path.endswith("safetensors" ): from safetensors import safe_open SCREAMING_SNAKE_CASE_: Optional[Any] = {} with safe_open(_UpperCAmelCase , framework="pt" , device="cpu" ) as f: for key in f.keys(): SCREAMING_SNAKE_CASE_: Any = f.get_tensor(_UpperCAmelCase ) else: SCREAMING_SNAKE_CASE_: Optional[int] = torch.load(_UpperCAmelCase , map_location=_UpperCAmelCase )["state_dict"] # Convert the VAE model. SCREAMING_SNAKE_CASE_: Optional[int] = create_vae_diffusers_config(_UpperCAmelCase , image_size=_UpperCAmelCase ) SCREAMING_SNAKE_CASE_: Optional[int] = custom_convert_ldm_vae_checkpoint(_UpperCAmelCase , _UpperCAmelCase ) SCREAMING_SNAKE_CASE_: int = AutoencoderKL(**_UpperCAmelCase ) vae.load_state_dict(_UpperCAmelCase ) vae.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": lowerCAmelCase : Optional[Any] = argparse.ArgumentParser() parser.add_argument("""--vae_pt_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") lowerCAmelCase : Optional[int] = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
13
import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTForImageClassification, ViTForMaskedImageModeling, ViTModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class __lowercase : """simple docstring""" def __init__( self : Any , lowerCAmelCase__ : Union[str, Any] , lowerCAmelCase__ : Any=13 , lowerCAmelCase__ : Tuple=30 , lowerCAmelCase__ : List[str]=2 , lowerCAmelCase__ : int=3 , lowerCAmelCase__ : Optional[int]=True , lowerCAmelCase__ : List[str]=True , lowerCAmelCase__ : str=32 , lowerCAmelCase__ : Any=5 , lowerCAmelCase__ : str=4 , lowerCAmelCase__ : int=37 , lowerCAmelCase__ : Optional[Any]="gelu" , lowerCAmelCase__ : Optional[int]=0.1 , lowerCAmelCase__ : Dict=0.1 , lowerCAmelCase__ : Tuple=10 , lowerCAmelCase__ : Optional[Any]=0.02 , lowerCAmelCase__ : List[str]=None , lowerCAmelCase__ : Union[str, Any]=2 , ): SCREAMING_SNAKE_CASE_: str = parent SCREAMING_SNAKE_CASE_: Optional[Any] = batch_size SCREAMING_SNAKE_CASE_: str = image_size SCREAMING_SNAKE_CASE_: Tuple = patch_size SCREAMING_SNAKE_CASE_: int = num_channels SCREAMING_SNAKE_CASE_: List[str] = is_training SCREAMING_SNAKE_CASE_: str = use_labels SCREAMING_SNAKE_CASE_: int = hidden_size SCREAMING_SNAKE_CASE_: List[Any] = num_hidden_layers SCREAMING_SNAKE_CASE_: Union[str, Any] = num_attention_heads SCREAMING_SNAKE_CASE_: Any = intermediate_size SCREAMING_SNAKE_CASE_: str = hidden_act SCREAMING_SNAKE_CASE_: str = hidden_dropout_prob SCREAMING_SNAKE_CASE_: List[str] = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_: int = type_sequence_label_size SCREAMING_SNAKE_CASE_: Dict = initializer_range SCREAMING_SNAKE_CASE_: Dict = scope SCREAMING_SNAKE_CASE_: Dict = encoder_stride # in ViT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) SCREAMING_SNAKE_CASE_: List[Any] = (image_size // patch_size) ** 2 SCREAMING_SNAKE_CASE_: Dict = num_patches + 1 def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): SCREAMING_SNAKE_CASE_: Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size]) SCREAMING_SNAKE_CASE_: str = None if self.use_labels: SCREAMING_SNAKE_CASE_: Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size) SCREAMING_SNAKE_CASE_: Optional[Any] = self.get_config() return config, pixel_values, labels def _SCREAMING_SNAKE_CASE ( self : Optional[int]): return ViTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=lowerCAmelCase__ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _SCREAMING_SNAKE_CASE ( self : Optional[Any] , lowerCAmelCase__ : int , lowerCAmelCase__ : int , lowerCAmelCase__ : Tuple): SCREAMING_SNAKE_CASE_: Union[str, Any] = ViTModel(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() SCREAMING_SNAKE_CASE_: Optional[int] = model(lowerCAmelCase__) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size)) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] , lowerCAmelCase__ : Optional[Any] , lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Dict): SCREAMING_SNAKE_CASE_: Optional[int] = ViTForMaskedImageModeling(config=lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() SCREAMING_SNAKE_CASE_: str = model(lowerCAmelCase__) self.parent.assertEqual( result.reconstruction.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size)) # test greyscale images SCREAMING_SNAKE_CASE_: Dict = 1 SCREAMING_SNAKE_CASE_: List[str] = ViTForMaskedImageModeling(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() SCREAMING_SNAKE_CASE_: List[str] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size]) SCREAMING_SNAKE_CASE_: str = model(lowerCAmelCase__) self.parent.assertEqual(result.reconstruction.shape , (self.batch_size, 1, self.image_size, self.image_size)) def _SCREAMING_SNAKE_CASE ( self : Any , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Union[str, Any]): SCREAMING_SNAKE_CASE_: Tuple = self.type_sequence_label_size SCREAMING_SNAKE_CASE_: List[str] = ViTForImageClassification(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() SCREAMING_SNAKE_CASE_: Any = model(lowerCAmelCase__ , labels=lowerCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size)) # test greyscale images SCREAMING_SNAKE_CASE_: Union[str, Any] = 1 SCREAMING_SNAKE_CASE_: List[str] = ViTForImageClassification(lowerCAmelCase__) model.to(lowerCAmelCase__) model.eval() SCREAMING_SNAKE_CASE_: Union[str, Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size]) SCREAMING_SNAKE_CASE_: Dict = model(lowerCAmelCase__) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size)) def _SCREAMING_SNAKE_CASE ( self : List[Any]): SCREAMING_SNAKE_CASE_: Union[str, Any] = self.prepare_config_and_inputs() ( ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ( SCREAMING_SNAKE_CASE_ ) , ): List[str] = config_and_inputs SCREAMING_SNAKE_CASE_: Optional[Any] = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __lowercase ( UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): """simple docstring""" _UpperCAmelCase : List[Any] = ( ( ViTModel, ViTForImageClassification, ViTForMaskedImageModeling, ) if is_torch_available() else () ) _UpperCAmelCase : Tuple = ( {'''feature-extraction''': ViTModel, '''image-classification''': ViTForImageClassification} if is_torch_available() else {} ) _UpperCAmelCase : List[str] = True _UpperCAmelCase : List[Any] = False _UpperCAmelCase : Optional[Any] = False _UpperCAmelCase : Tuple = False def _SCREAMING_SNAKE_CASE ( self : Optional[int]): SCREAMING_SNAKE_CASE_: List[str] = ViTModelTester(self) SCREAMING_SNAKE_CASE_: Union[str, Any] = ConfigTester(self , config_class=lowerCAmelCase__ , has_text_modality=lowerCAmelCase__ , hidden_size=37) def _SCREAMING_SNAKE_CASE ( self : Any): self.config_tester.run_common_tests() @unittest.skip(reason="ViT does not use inputs_embeds") def _SCREAMING_SNAKE_CASE ( self : str): pass def _SCREAMING_SNAKE_CASE ( self : str): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_: Dict = model_class(lowerCAmelCase__) self.assertIsInstance(model.get_input_embeddings() , (nn.Module)) SCREAMING_SNAKE_CASE_: List[Any] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCAmelCase__ , nn.Linear)) def _SCREAMING_SNAKE_CASE ( self : Optional[int]): SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_: Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE_: List[Any] = model_class(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: int = inspect.signature(model.forward) # signature.parameters is an OrderedDict => so arg_names order is deterministic SCREAMING_SNAKE_CASE_: Optional[Any] = [*signature.parameters.keys()] SCREAMING_SNAKE_CASE_: Optional[int] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): SCREAMING_SNAKE_CASE_: Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : Optional[Any]): SCREAMING_SNAKE_CASE_: Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*lowerCAmelCase__) def _SCREAMING_SNAKE_CASE ( self : List[str]): SCREAMING_SNAKE_CASE_: int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCAmelCase__) @slow def _SCREAMING_SNAKE_CASE ( self : int): for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE_: Union[str, Any] = ViTModel.from_pretrained(lowerCAmelCase__) self.assertIsNotNone(lowerCAmelCase__) def A_ ( ): SCREAMING_SNAKE_CASE_: List[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class __lowercase ( unittest.TestCase ): """simple docstring""" @cached_property def _SCREAMING_SNAKE_CASE ( self : int): return ViTImageProcessor.from_pretrained("google/vit-base-patch16-224") if is_vision_available() else None @slow def _SCREAMING_SNAKE_CASE ( self : Union[str, Any]): SCREAMING_SNAKE_CASE_: int = ViTForImageClassification.from_pretrained("google/vit-base-patch16-224").to(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Optional[Any] = self.default_image_processor SCREAMING_SNAKE_CASE_: str = prepare_img() SCREAMING_SNAKE_CASE_: Optional[Any] = image_processor(images=lowerCAmelCase__ , return_tensors="pt").to(lowerCAmelCase__) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE_: Optional[int] = model(**lowerCAmelCase__) # verify the logits SCREAMING_SNAKE_CASE_: Any = torch.Size((1, 1000)) self.assertEqual(outputs.logits.shape , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[Any] = torch.tensor([-0.2744, 0.8215, -0.0836]).to(lowerCAmelCase__) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowerCAmelCase__ , atol=1E-4)) @slow def _SCREAMING_SNAKE_CASE ( self : List[Any]): # ViT models have an `interpolate_pos_encoding` argument in their forward method, # allowing to interpolate the pre-trained position embeddings in order to use # the model on higher resolutions. The DINO model by Facebook AI leverages this # to visualize self-attention on higher resolution images. SCREAMING_SNAKE_CASE_: str = ViTModel.from_pretrained("facebook/dino-vits8").to(lowerCAmelCase__) SCREAMING_SNAKE_CASE_: List[Any] = ViTImageProcessor.from_pretrained("facebook/dino-vits8" , size=480) SCREAMING_SNAKE_CASE_: List[Any] = prepare_img() SCREAMING_SNAKE_CASE_: List[Any] = image_processor(images=lowerCAmelCase__ , return_tensors="pt") SCREAMING_SNAKE_CASE_: int = inputs.pixel_values.to(lowerCAmelCase__) # forward pass with torch.no_grad(): SCREAMING_SNAKE_CASE_: Optional[int] = model(lowerCAmelCase__ , interpolate_pos_encoding=lowerCAmelCase__) # verify the logits SCREAMING_SNAKE_CASE_: Tuple = torch.Size((1, 3601, 384)) self.assertEqual(outputs.last_hidden_state.shape , lowerCAmelCase__) SCREAMING_SNAKE_CASE_: Union[str, Any] = torch.tensor( [[4.2340, 4.3906, -6.6692], [4.5463, 1.8928, -6.7257], [4.4429, 0.8496, -5.8585]]).to(lowerCAmelCase__) self.assertTrue(torch.allclose(outputs.last_hidden_state[0, :3, :3] , lowerCAmelCase__ , atol=1E-4)) @slow @require_accelerate @require_torch_gpu def _SCREAMING_SNAKE_CASE ( self : int): SCREAMING_SNAKE_CASE_: Dict = ViTModel.from_pretrained("facebook/dino-vits8" , torch_dtype=torch.floataa , device_map="auto") SCREAMING_SNAKE_CASE_: int = self.default_image_processor SCREAMING_SNAKE_CASE_: Union[str, Any] = prepare_img() SCREAMING_SNAKE_CASE_: Dict = image_processor(images=lowerCAmelCase__ , return_tensors="pt") SCREAMING_SNAKE_CASE_: str = inputs.pixel_values.to(lowerCAmelCase__) # forward pass to make sure inference works in fp16 with torch.no_grad(): SCREAMING_SNAKE_CASE_: str = model(lowerCAmelCase__)
13
1
'''simple docstring''' import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _lowerCAmelCase ( a_, unittest.TestCase ): """simple docstring""" lowerCamelCase = CTRLTokenizer lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> List[Any]: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt A_ : Tuple = ["""adapt""", """re@@""", """a@@""", """apt""", """c@@""", """t""", """<unk>"""] A_ : Optional[int] = dict(zip(_lowerCamelCase , range(len(_lowerCamelCase ) ) ) ) A_ : int = ["""#version: 0.2""", """a p""", """ap t</w>""", """r e""", """a d""", """ad apt</w>""", """"""] A_ : Tuple = {"""unk_token""": """<unk>"""} A_ : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) A_ : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(_lowerCamelCase ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(_lowerCamelCase ) ) def UpperCAmelCase_ ( self , **_lowerCamelCase ) -> List[str]: kwargs.update(self.special_tokens_map ) return CTRLTokenizer.from_pretrained(self.tmpdirname , **_lowerCamelCase ) def UpperCAmelCase_ ( self , _lowerCamelCase ) -> Optional[int]: A_ : Optional[int] = """adapt react readapt apt""" A_ : Optional[Any] = """adapt react readapt apt""" return input_text, output_text def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Any = CTRLTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) A_ : List[Any] = """adapt react readapt apt""" A_ : Optional[int] = """adapt re@@ a@@ c@@ t re@@ adapt apt""".split() A_ : Any = tokenizer.tokenize(_lowerCamelCase ) self.assertListEqual(_lowerCamelCase , _lowerCamelCase ) A_ : Tuple = tokens + [tokenizer.unk_token] A_ : List[Any] = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(_lowerCamelCase ) , _lowerCamelCase )
350
'''simple docstring''' def UpperCAmelCase ( a_ , a_ ) -> int: """simple docstring""" A_ : int = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): A_ : Tuple = n - k # Calculate C(n,k) for i in range(a_ ): result *= n - i result //= i + 1 return result def UpperCAmelCase ( a_ ) -> int: """simple docstring""" return binomial_coefficient(2 * node_count , a_ ) // (node_count + 1) def UpperCAmelCase ( a_ ) -> int: """simple docstring""" if n < 0: raise ValueError("""factorial() not defined for negative values""" ) A_ : Union[str, Any] = 1 for i in range(1 , n + 1 ): result *= i return result def UpperCAmelCase ( a_ ) -> int: """simple docstring""" return catalan_number(a_ ) * factorial(a_ ) if __name__ == "__main__": UpperCamelCase__ : Any = int(input('Enter the number of nodes: ').strip() or 0) if node_count <= 0: raise ValueError('We need some nodes to work with.') print( f'Given {node_count} nodes, there are {binary_tree_count(node_count)} ' f'binary trees and {catalan_number(node_count)} binary search trees.' )
164
0
import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('''3.8'''): import importlib_metadata else: import importlib.metadata as importlib_metadata def _UpperCamelCase ( snake_case__, snake_case__=False ) -> Optional[Any]: try: __UpperCAmelCase : Any = os.environ[key] except KeyError: # KEY isn't set, default to `default`. __UpperCAmelCase : Dict = default else: # KEY is set, convert it to True or False. try: __UpperCAmelCase : Union[str, Any] = strtobool(snake_case__ ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f'''If set, {key} must be yes or no.''' ) return _value _snake_case = parse_flag_from_env('''RUN_SLOW''', default=False) _snake_case = parse_flag_from_env('''RUN_REMOTE''', default=False) _snake_case = parse_flag_from_env('''RUN_LOCAL''', default=True) _snake_case = parse_flag_from_env('''RUN_PACKAGED''', default=True) # Compression _snake_case = pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='''test requires lz4''') _snake_case = pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='''test requires py7zr''') _snake_case = pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='''test requires zstandard''') # Audio _snake_case = pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('''soundfile''') is None or version.parse(importlib_metadata.version('''soundfile''')) < version.parse('''0.12.0'''), reason='''test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ''', ) # Beam _snake_case = pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('''0.3.2'''), reason='''test requires apache-beam and a compatible dill version''', ) # Dill-cloudpickle compatibility _snake_case = pytest.mark.skipif( config.DILL_VERSION <= version.parse('''0.3.2'''), reason='''test requires dill>0.3.2 for cloudpickle compatibility''', ) # Windows _snake_case = pytest.mark.skipif( sys.platform == '''win32''', reason='''test should not be run on Windows''', ) def _UpperCamelCase ( snake_case__ ) -> Tuple: try: import faiss # noqa except ImportError: __UpperCAmelCase : Optional[Any] = unittest.skip("test requires faiss" )(snake_case__ ) return test_case def _UpperCamelCase ( snake_case__ ) -> int: try: import regex # noqa except ImportError: __UpperCAmelCase : Dict = unittest.skip("test requires regex" )(snake_case__ ) return test_case def _UpperCamelCase ( snake_case__ ) -> str: try: import elasticsearch # noqa except ImportError: __UpperCAmelCase : str = unittest.skip("test requires elasticsearch" )(snake_case__ ) return test_case def _UpperCamelCase ( snake_case__ ) -> Optional[int]: try: import sqlalchemy # noqa except ImportError: __UpperCAmelCase : Tuple = unittest.skip("test requires sqlalchemy" )(snake_case__ ) return test_case def _UpperCamelCase ( snake_case__ ) -> Optional[int]: if not config.TORCH_AVAILABLE: __UpperCAmelCase : Union[str, Any] = unittest.skip("test requires PyTorch" )(snake_case__ ) return test_case def _UpperCamelCase ( snake_case__ ) -> Union[str, Any]: if not config.TF_AVAILABLE: __UpperCAmelCase : Union[str, Any] = unittest.skip("test requires TensorFlow" )(snake_case__ ) return test_case def _UpperCamelCase ( snake_case__ ) -> str: if not config.JAX_AVAILABLE: __UpperCAmelCase : Union[str, Any] = unittest.skip("test requires JAX" )(snake_case__ ) return test_case def _UpperCamelCase ( snake_case__ ) -> Dict: if not config.PIL_AVAILABLE: __UpperCAmelCase : Optional[int] = unittest.skip("test requires Pillow" )(snake_case__ ) return test_case def _UpperCamelCase ( snake_case__ ) -> List[str]: try: import transformers # noqa F401 except ImportError: return unittest.skip("test requires transformers" )(snake_case__ ) else: return test_case def _UpperCamelCase ( snake_case__ ) -> str: try: import tiktoken # noqa F401 except ImportError: return unittest.skip("test requires tiktoken" )(snake_case__ ) else: return test_case def _UpperCamelCase ( snake_case__ ) -> List[str]: try: import spacy # noqa F401 except ImportError: return unittest.skip("test requires spacy" )(snake_case__ ) else: return test_case def _UpperCamelCase ( snake_case__ ) -> Tuple: def _require_spacy_model(snake_case__ ): try: import spacy # noqa F401 spacy.load(snake_case__ ) except ImportError: return unittest.skip("test requires spacy" )(snake_case__ ) except OSError: return unittest.skip("test requires spacy model '{}'".format(snake_case__ ) )(snake_case__ ) else: return test_case return _require_spacy_model def _UpperCamelCase ( snake_case__ ) -> Union[str, Any]: try: import pyspark # noqa F401 except ImportError: return unittest.skip("test requires pyspark" )(snake_case__ ) else: return test_case def _UpperCamelCase ( snake_case__ ) -> Tuple: try: import joblibspark # noqa F401 except ImportError: return unittest.skip("test requires joblibspark" )(snake_case__ ) else: return test_case def _UpperCamelCase ( snake_case__ ) -> Optional[Any]: if not _run_slow_tests or _run_slow_tests == 0: __UpperCAmelCase : List[Any] = unittest.skip("test is slow" )(snake_case__ ) return test_case def _UpperCamelCase ( snake_case__ ) -> List[Any]: if not _run_local_tests or _run_local_tests == 0: __UpperCAmelCase : Tuple = unittest.skip("test is local" )(snake_case__ ) return test_case def _UpperCamelCase ( snake_case__ ) -> int: if not _run_packaged_tests or _run_packaged_tests == 0: __UpperCAmelCase : str = unittest.skip("test is packaged" )(snake_case__ ) return test_case def _UpperCamelCase ( snake_case__ ) -> List[Any]: if not _run_remote_tests or _run_remote_tests == 0: __UpperCAmelCase : str = unittest.skip("test requires remote" )(snake_case__ ) return test_case def _UpperCamelCase ( *snake_case__ ) -> Any: def decorate(cls ): for name, fn in cls.__dict__.items(): if callable(snake_case__ ) and name.startswith("test" ): for decorator in decorators: __UpperCAmelCase : Tuple = decorator(snake_case__ ) setattr(cls, snake_case__, snake_case__ ) return cls return decorate class _snake_case ( _lowercase ): pass class _snake_case ( _lowercase ): lowerCamelCase__: List[str] = 0 lowerCamelCase__: Dict = 1 lowerCamelCase__: Tuple = 2 @contextmanager def _UpperCamelCase ( snake_case__=OfflineSimulationMode.CONNECTION_FAILS, snake_case__=1e-1_6 ) -> List[Any]: __UpperCAmelCase : str = requests.Session().request def timeout_request(snake_case__, snake_case__, snake_case__, **snake_case__ ): # Change the url to an invalid url so that the connection hangs __UpperCAmelCase : Optional[int] = "https://10.255.255.1" if kwargs.get("timeout" ) is None: raise RequestWouldHangIndefinitelyError( f'''Tried a call to {url} in offline mode with no timeout set. Please set a timeout.''' ) __UpperCAmelCase : Optional[int] = timeout try: return online_request(snake_case__, snake_case__, **snake_case__ ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier __UpperCAmelCase : Optional[Any] = url __UpperCAmelCase : str = e.args[0] __UpperCAmelCase : str = (max_retry_error.args[0].replace("10.255.255.1", f'''OfflineMock[{url}]''' ),) __UpperCAmelCase : Tuple = (max_retry_error,) raise def raise_connection_error(snake_case__, snake_case__, **snake_case__ ): raise requests.ConnectionError("Offline mode is enabled.", request=snake_case__ ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch("requests.Session.send", snake_case__ ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch("requests.Session.request", snake_case__ ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch("datasets.config.HF_DATASETS_OFFLINE", snake_case__ ): yield else: raise ValueError("Please use a value from the OfflineSimulationMode enum." ) @contextmanager def _UpperCamelCase ( *snake_case__, **snake_case__ ) -> Union[str, Any]: __UpperCAmelCase : Tuple = str(Path().resolve() ) with tempfile.TemporaryDirectory(*snake_case__, **snake_case__ ) as tmp_dir: try: os.chdir(snake_case__ ) yield finally: os.chdir(snake_case__ ) @contextmanager def _UpperCamelCase ( ) -> Tuple: import gc gc.collect() __UpperCAmelCase : Dict = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def _UpperCamelCase ( ) -> Tuple: import gc gc.collect() __UpperCAmelCase : List[str] = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def _UpperCamelCase ( snake_case__, snake_case__ ) -> Tuple: return deepcopy(snake_case__ ).integers(0, 100, 10 ).tolist() == deepcopy(snake_case__ ).integers(0, 100, 10 ).tolist() def _UpperCamelCase ( snake_case__ ) -> Optional[int]: import decorator from requests.exceptions import HTTPError def _wrapper(snake_case__, *snake_case__, **snake_case__ ): try: return func(*snake_case__, **snake_case__ ) except HTTPError as err: if str(snake_case__ ).startswith("500" ) or str(snake_case__ ).startswith("502" ): pytest.xfail(str(snake_case__ ) ) raise err return decorator.decorator(_wrapper, snake_case__ ) class _snake_case : def __init__( self: Optional[int] , __lowerCamelCase: Optional[int] , __lowerCamelCase: List[Any] , __lowerCamelCase: Dict ) -> Dict: __UpperCAmelCase : List[str] = returncode __UpperCAmelCase : str = stdout __UpperCAmelCase : Optional[int] = stderr async def _UpperCamelCase ( snake_case__, snake_case__ ) -> Union[str, Any]: while True: __UpperCAmelCase : List[str] = await stream.readline() if line: callback(snake_case__ ) else: break async def _UpperCamelCase ( snake_case__, snake_case__=None, snake_case__=None, snake_case__=None, snake_case__=False, snake_case__=False ) -> _RunOutput: if echo: print("\nRunning: ", " ".join(snake_case__ ) ) __UpperCAmelCase : List[str] = await asyncio.create_subprocess_exec( cmd[0], *cmd[1:], stdin=snake_case__, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE, env=snake_case__, ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) __UpperCAmelCase : List[Any] = [] __UpperCAmelCase : Union[str, Any] = [] def tee(snake_case__, snake_case__, snake_case__, snake_case__="" ): __UpperCAmelCase : Optional[Any] = line.decode("utf-8" ).rstrip() sink.append(snake_case__ ) if not quiet: print(snake_case__, snake_case__, file=snake_case__ ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout, lambda snake_case__ : tee(snake_case__, snake_case__, sys.stdout, label="stdout:" ) ), _read_stream(p.stderr, lambda snake_case__ : tee(snake_case__, snake_case__, sys.stderr, label="stderr:" ) ), ], timeout=snake_case__, ) return _RunOutput(await p.wait(), snake_case__, snake_case__ ) def _UpperCamelCase ( snake_case__, snake_case__=None, snake_case__=None, snake_case__=180, snake_case__=False, snake_case__=True ) -> _RunOutput: __UpperCAmelCase : Dict = asyncio.get_event_loop() __UpperCAmelCase : Dict = loop.run_until_complete( _stream_subprocess(snake_case__, env=snake_case__, stdin=snake_case__, timeout=snake_case__, quiet=snake_case__, echo=snake_case__ ) ) __UpperCAmelCase : List[Any] = " ".join(snake_case__ ) if result.returncode > 0: __UpperCAmelCase : List[Any] = "\n".join(result.stderr ) raise RuntimeError( f'''\'{cmd_str}\' failed with returncode {result.returncode}\n\n''' f'''The combined stderr from workers follows:\n{stderr}''' ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(f'''\'{cmd_str}\' produced no output.''' ) return result def _UpperCamelCase ( ) -> Tuple: __UpperCAmelCase : Any = os.environ.get("PYTEST_XDIST_WORKER", "gw0" ) __UpperCAmelCase : int = re.sub(r"^gw", "", snake_case__, 0, re.M ) return int(snake_case__ ) def _UpperCamelCase ( ) -> Dict: __UpperCAmelCase : Tuple = 2_9500 __UpperCAmelCase : Union[str, Any] = pytest_xdist_worker_id() return port + uniq_delta
157
import numpy as np import datasets _snake_case = ''' Compute the Mahalanobis Distance Mahalonobis distance is the distance between a point and a distribution. And not between two distinct points. It is effectively a multivariate equivalent of the Euclidean distance. It was introduced by Prof. P. C. Mahalanobis in 1936 and has been used in various statistical applications ever since [source: https://www.machinelearningplus.com/statistics/mahalanobis-distance/] ''' _snake_case = '''\ @article{de2000mahalanobis, title={The mahalanobis distance}, author={De Maesschalck, Roy and Jouan-Rimbaud, Delphine and Massart, D{\'e}sir{\'e} L}, journal={Chemometrics and intelligent laboratory systems}, volume={50}, number={1}, pages={1--18}, year={2000}, publisher={Elsevier} } ''' _snake_case = ''' Args: X: List of datapoints to be compared with the `reference_distribution`. reference_distribution: List of datapoints from the reference distribution we want to compare to. Returns: mahalanobis: The Mahalonobis distance for each datapoint in `X`. Examples: >>> mahalanobis_metric = datasets.load_metric("mahalanobis") >>> results = mahalanobis_metric.compute(reference_distribution=[[0, 1], [1, 0]], X=[[0, 1]]) >>> print(results) {\'mahalanobis\': array([0.5])} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _snake_case ( datasets.Metric ): def _lowerCamelCase ( self: int ) -> int: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "X": datasets.Sequence(datasets.Value("float" , id="sequence" ) , id="X" ), } ) , ) def _lowerCamelCase ( self: Optional[int] , __lowerCamelCase: Union[str, Any] , __lowerCamelCase: str ) -> Tuple: # convert to numpy arrays __UpperCAmelCase : Dict = np.array(__lowerCamelCase ) __UpperCAmelCase : Optional[int] = np.array(__lowerCamelCase ) # Assert that arrays are 2D if len(X.shape ) != 2: raise ValueError("Expected `X` to be a 2D vector" ) if len(reference_distribution.shape ) != 2: raise ValueError("Expected `reference_distribution` to be a 2D vector" ) if reference_distribution.shape[0] < 2: raise ValueError( "Expected `reference_distribution` to be a 2D vector with more than one element in the first dimension" ) # Get mahalanobis distance for each prediction __UpperCAmelCase : List[str] = X - np.mean(__lowerCamelCase ) __UpperCAmelCase : Tuple = np.cov(reference_distribution.T ) try: __UpperCAmelCase : int = np.linalg.inv(__lowerCamelCase ) except np.linalg.LinAlgError: __UpperCAmelCase : List[Any] = np.linalg.pinv(__lowerCamelCase ) __UpperCAmelCase : Union[str, Any] = np.dot(__lowerCamelCase , __lowerCamelCase ) __UpperCAmelCase : Optional[int] = np.dot(__lowerCamelCase , X_minus_mu.T ).diagonal() return {"mahalanobis": mahal_dist}
157
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _SCREAMING_SNAKE_CASE = { """configuration_xlm_roberta""": [ """XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XLMRobertaConfig""", """XLMRobertaOnnxConfig""", ], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ["""XLMRobertaTokenizer"""] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = ["""XLMRobertaTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ """XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """XLMRobertaForCausalLM""", """XLMRobertaForMaskedLM""", """XLMRobertaForMultipleChoice""", """XLMRobertaForQuestionAnswering""", """XLMRobertaForSequenceClassification""", """XLMRobertaForTokenClassification""", """XLMRobertaModel""", """XLMRobertaPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ """TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFXLMRobertaForCausalLM""", """TFXLMRobertaForMaskedLM""", """TFXLMRobertaForMultipleChoice""", """TFXLMRobertaForQuestionAnswering""", """TFXLMRobertaForSequenceClassification""", """TFXLMRobertaForTokenClassification""", """TFXLMRobertaModel""", """TFXLMRobertaPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _SCREAMING_SNAKE_CASE = [ """FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """FlaxXLMRobertaForMaskedLM""", """FlaxXLMRobertaForCausalLM""", """FlaxXLMRobertaForMultipleChoice""", """FlaxXLMRobertaForQuestionAnswering""", """FlaxXLMRobertaForSequenceClassification""", """FlaxXLMRobertaForTokenClassification""", """FlaxXLMRobertaModel""", """FlaxXLMRobertaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaConfig, XLMRobertaOnnxConfig, ) try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta import XLMRobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta_fast import XLMRobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaForCausalLM, XLMRobertaForMaskedLM, XLMRobertaForMultipleChoice, XLMRobertaForQuestionAnswering, XLMRobertaForSequenceClassification, XLMRobertaForTokenClassification, XLMRobertaModel, XLMRobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm_roberta import ( TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMRobertaForCausalLM, TFXLMRobertaForMaskedLM, TFXLMRobertaForMultipleChoice, TFXLMRobertaForQuestionAnswering, TFXLMRobertaForSequenceClassification, TFXLMRobertaForTokenClassification, TFXLMRobertaModel, TFXLMRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xlm_roberta import ( FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxXLMRobertaForCausalLM, FlaxXLMRobertaForMaskedLM, FlaxXLMRobertaForMultipleChoice, FlaxXLMRobertaForQuestionAnswering, FlaxXLMRobertaForSequenceClassification, FlaxXLMRobertaForTokenClassification, FlaxXLMRobertaModel, FlaxXLMRobertaPreTrainedModel, ) else: import sys _SCREAMING_SNAKE_CASE = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
88
from __future__ import annotations import pandas as pd def SCREAMING_SNAKE_CASE__ ( __a , __a , __a ): snake_case_ : Optional[Any] = [0] * no_of_processes snake_case_ : Tuple = [0] * no_of_processes # Copy the burst time into remaining_time[] for i in range(__a ): snake_case_ : Union[str, Any] = burst_time[i] snake_case_ : Optional[Any] = 0 snake_case_ : Dict = 0 snake_case_ : Any = 9_99_99_99_99 snake_case_ : Tuple = 0 snake_case_ : List[Any] = False # Process until all processes are completed while complete != no_of_processes: for j in range(__a ): if arrival_time[j] <= increment_time and remaining_time[j] > 0: if remaining_time[j] < minm: snake_case_ : str = remaining_time[j] snake_case_ : Any = j snake_case_ : List[str] = True if not check: increment_time += 1 continue remaining_time[short] -= 1 snake_case_ : Any = remaining_time[short] if minm == 0: snake_case_ : Dict = 9_99_99_99_99 if remaining_time[short] == 0: complete += 1 snake_case_ : List[str] = False # Find finish time of current process snake_case_ : List[str] = increment_time + 1 # Calculate waiting time snake_case_ : Any = finish_time - arrival_time[short] snake_case_ : Any = finar - burst_time[short] if waiting_time[short] < 0: snake_case_ : Optional[int] = 0 # Increment time increment_time += 1 return waiting_time def SCREAMING_SNAKE_CASE__ ( __a , __a , __a ): snake_case_ : Tuple = [0] * no_of_processes for i in range(__a ): snake_case_ : str = burst_time[i] + waiting_time[i] return turn_around_time def SCREAMING_SNAKE_CASE__ ( __a , __a , __a ): snake_case_ : int = 0 snake_case_ : Optional[Any] = 0 for i in range(__a ): snake_case_ : int = total_waiting_time + waiting_time[i] snake_case_ : Optional[Any] = total_turn_around_time + turn_around_time[i] print(f"""Average waiting time = {total_waiting_time / no_of_processes:.5f}""" ) print('Average turn around time =' , total_turn_around_time / no_of_processes ) if __name__ == "__main__": print("""Enter how many process you want to analyze""") _SCREAMING_SNAKE_CASE = int(input()) _SCREAMING_SNAKE_CASE = [0] * no_of_processes _SCREAMING_SNAKE_CASE = [0] * no_of_processes _SCREAMING_SNAKE_CASE = list(range(1, no_of_processes + 1)) for i in range(no_of_processes): print("""Enter the arrival time and burst time for process:--""" + str(i + 1)) _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = map(int, input().split()) _SCREAMING_SNAKE_CASE = calculate_waitingtime(arrival_time, burst_time, no_of_processes) _SCREAMING_SNAKE_CASE = burst_time _SCREAMING_SNAKE_CASE = no_of_processes _SCREAMING_SNAKE_CASE = waiting_time _SCREAMING_SNAKE_CASE = calculate_turnaroundtime(bt, n, wt) calculate_average_times(waiting_time, turn_around_time, no_of_processes) _SCREAMING_SNAKE_CASE = pd.DataFrame( list(zip(processes, burst_time, arrival_time, waiting_time, turn_around_time)), columns=[ """Process""", """BurstTime""", """ArrivalTime""", """WaitingTime""", """TurnAroundTime""", ], ) # Printing the dataFrame pd.set_option("""display.max_rows""", fcfs.shape[0] + 1) print(fcfs)
88
1
import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers.models.esm.modeling_esmfold import EsmForProteinFolding class _a : def __init__(self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_=13, SCREAMING_SNAKE_CASE_=7, SCREAMING_SNAKE_CASE_=False, SCREAMING_SNAKE_CASE_=True, SCREAMING_SNAKE_CASE_=False, SCREAMING_SNAKE_CASE_=False, SCREAMING_SNAKE_CASE_=19, SCREAMING_SNAKE_CASE_=32, SCREAMING_SNAKE_CASE_=5, SCREAMING_SNAKE_CASE_=4, SCREAMING_SNAKE_CASE_=37, SCREAMING_SNAKE_CASE_="gelu", SCREAMING_SNAKE_CASE_=0.1, SCREAMING_SNAKE_CASE_=0.1, SCREAMING_SNAKE_CASE_=512, SCREAMING_SNAKE_CASE_=16, SCREAMING_SNAKE_CASE_=2, SCREAMING_SNAKE_CASE_=0.0_2, SCREAMING_SNAKE_CASE_=3, SCREAMING_SNAKE_CASE_=4, SCREAMING_SNAKE_CASE_=None, ) -> Any: UpperCAmelCase_: Dict = parent UpperCAmelCase_: Dict = batch_size UpperCAmelCase_: Optional[int] = seq_length UpperCAmelCase_: Dict = is_training UpperCAmelCase_: Union[str, Any] = use_input_mask UpperCAmelCase_: Dict = use_token_type_ids UpperCAmelCase_: Any = use_labels UpperCAmelCase_: Tuple = vocab_size UpperCAmelCase_: Optional[int] = hidden_size UpperCAmelCase_: List[str] = num_hidden_layers UpperCAmelCase_: Optional[int] = num_attention_heads UpperCAmelCase_: Optional[Any] = intermediate_size UpperCAmelCase_: Optional[Any] = hidden_act UpperCAmelCase_: Tuple = hidden_dropout_prob UpperCAmelCase_: List[str] = attention_probs_dropout_prob UpperCAmelCase_: Optional[Any] = max_position_embeddings UpperCAmelCase_: Dict = type_vocab_size UpperCAmelCase_: Tuple = type_sequence_label_size UpperCAmelCase_: Dict = initializer_range UpperCAmelCase_: Optional[int] = num_labels UpperCAmelCase_: List[str] = num_choices UpperCAmelCase_: List[str] = scope def __snake_case (self ) -> str: UpperCAmelCase_: Optional[int] = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) UpperCAmelCase_: Union[str, Any] = None if self.use_input_mask: UpperCAmelCase_: int = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_: Dict = None UpperCAmelCase_: List[Any] = None UpperCAmelCase_: Optional[Any] = None if self.use_labels: UpperCAmelCase_: Union[str, Any] = ids_tensor([self.batch_size], self.type_sequence_label_size ) UpperCAmelCase_: Dict = ids_tensor([self.batch_size, self.seq_length], self.num_labels ) UpperCAmelCase_: List[str] = ids_tensor([self.batch_size], self.num_choices ) UpperCAmelCase_: Optional[int] = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def __snake_case (self ) -> Tuple: UpperCAmelCase_: Optional[Any] = EsmConfig( vocab_size=33, hidden_size=self.hidden_size, pad_token_id=1, num_hidden_layers=self.num_hidden_layers, num_attention_heads=self.num_attention_heads, intermediate_size=self.intermediate_size, hidden_act=self.hidden_act, hidden_dropout_prob=self.hidden_dropout_prob, attention_probs_dropout_prob=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, type_vocab_size=self.type_vocab_size, initializer_range=self.initializer_range, is_folding_model=SCREAMING_SNAKE_CASE_, esmfold_config={"""trunk""": {"""num_blocks""": 2}, """fp16_esm""": False}, ) return config def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Tuple: UpperCAmelCase_: int = EsmForProteinFolding(config=SCREAMING_SNAKE_CASE_ ).float() model.to(SCREAMING_SNAKE_CASE_ ) model.eval() UpperCAmelCase_: Optional[Any] = model(SCREAMING_SNAKE_CASE_, attention_mask=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: List[str] = model(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_: List[str] = model(SCREAMING_SNAKE_CASE_ ) self.parent.assertEqual(result.positions.shape, (8, self.batch_size, self.seq_length, 14, 3) ) self.parent.assertEqual(result.angles.shape, (8, self.batch_size, self.seq_length, 7, 2) ) def __snake_case (self ) -> int: UpperCAmelCase_: Optional[int] = self.prepare_config_and_inputs() ( ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ( UpperCAmelCase_ ) , ): Tuple = config_and_inputs UpperCAmelCase_: Optional[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class _a ( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): A = False A = (EsmForProteinFolding,) if is_torch_available() else () A = () A = {} if is_torch_available() else {} A = False def __snake_case (self ) -> Union[str, Any]: UpperCAmelCase_: Dict = EsmFoldModelTester(self ) UpperCAmelCase_: List[str] = ConfigTester(self, config_class=SCREAMING_SNAKE_CASE_, hidden_size=37 ) def __snake_case (self ) -> Any: self.config_tester.run_common_tests() def __snake_case (self ) -> Dict: UpperCAmelCase_: List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*SCREAMING_SNAKE_CASE_ ) @unittest.skip("""Does not support attention outputs""" ) def __snake_case (self ) -> int: pass @unittest.skip def __snake_case (self ) -> int: pass @unittest.skip("""Esm does not support embedding resizing""" ) def __snake_case (self ) -> List[str]: pass @unittest.skip("""Esm does not support embedding resizing""" ) def __snake_case (self ) -> int: pass @unittest.skip("""ESMFold does not support passing input embeds!""" ) def __snake_case (self ) -> Dict: pass @unittest.skip("""ESMFold does not support head pruning.""" ) def __snake_case (self ) -> Dict: pass @unittest.skip("""ESMFold does not support head pruning.""" ) def __snake_case (self ) -> Optional[Any]: pass @unittest.skip("""ESMFold does not support head pruning.""" ) def __snake_case (self ) -> str: pass @unittest.skip("""ESMFold does not support head pruning.""" ) def __snake_case (self ) -> Any: pass @unittest.skip("""ESMFold does not support head pruning.""" ) def __snake_case (self ) -> List[str]: pass @unittest.skip("""ESMFold does not output hidden states in the normal way.""" ) def __snake_case (self ) -> List[Any]: pass @unittest.skip("""ESMfold does not output hidden states in the normal way.""" ) def __snake_case (self ) -> int: pass @unittest.skip("""ESMFold only has one output format.""" ) def __snake_case (self ) -> List[str]: pass @unittest.skip("""This test doesn't work for ESMFold and doesn't test core functionality""" ) def __snake_case (self ) -> str: pass @unittest.skip("""ESMFold does not support input chunking.""" ) def __snake_case (self ) -> List[Any]: pass @unittest.skip("""ESMFold doesn't respect you and it certainly doesn't respect your initialization arguments.""" ) def __snake_case (self ) -> int: pass @unittest.skip("""ESMFold doesn't support torchscript compilation.""" ) def __snake_case (self ) -> Union[str, Any]: pass @unittest.skip("""ESMFold doesn't support torchscript compilation.""" ) def __snake_case (self ) -> Optional[Any]: pass @unittest.skip("""ESMFold doesn't support torchscript compilation.""" ) def __snake_case (self ) -> str: pass @unittest.skip("""ESMFold doesn't support data parallel.""" ) def __snake_case (self ) -> List[Any]: pass @unittest.skip("""Will be fixed soon by reducing the size of the model used for common tests.""" ) def __snake_case (self ) -> List[str]: pass @require_torch class _a ( _lowerCAmelCase ): @slow def __snake_case (self ) -> List[Any]: UpperCAmelCase_: Tuple = EsmForProteinFolding.from_pretrained("""facebook/esmfold_v1""" ).float() model.eval() UpperCAmelCase_: str = torch.tensor([[0, 6, 4, 13, 5, 4, 16, 12, 11, 7, 2]] ) UpperCAmelCase_: str = model(SCREAMING_SNAKE_CASE_ )["""positions"""] UpperCAmelCase_: Optional[Any] = torch.tensor([2.5_8_2_8, 0.7_9_9_3, -1_0.9_3_3_4], dtype=torch.floataa ) self.assertTrue(torch.allclose(position_outputs[0, 0, 0, 0], SCREAMING_SNAKE_CASE_, atol=1E-4 ) )
147
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) a : List[str] = { 'configuration_vision_encoder_decoder': ['VisionEncoderDecoderConfig', 'VisionEncoderDecoderOnnxConfig'] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : List[Any] = ['VisionEncoderDecoderModel'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Optional[Any] = ['TFVisionEncoderDecoderModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Tuple = ['FlaxVisionEncoderDecoderModel'] if TYPE_CHECKING: from .configuration_vision_encoder_decoder import VisionEncoderDecoderConfig, VisionEncoderDecoderOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vision_encoder_decoder import VisionEncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vision_encoder_decoder import TFVisionEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vision_encoder_decoder import FlaxVisionEncoderDecoderModel else: import sys a : List[str] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
147
1
lowerCAmelCase = '''0.21.0''' from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from .data_loader import skip_first_batches from .launchers import debug_launcher, notebook_launcher from .state import PartialState from .utils import ( DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, FullyShardedDataParallelPlugin, GradScalerKwargs, InitProcessGroupKwargs, find_executable_batch_size, infer_auto_device_map, is_rich_available, load_checkpoint_in_model, synchronize_rng_states, ) if is_rich_available(): from .utils import rich
357
from typing import Any import numpy as np def _lowerCamelCase( lowercase__ ) -> bool: '''simple docstring''' return np.array_equal(lowercase__ , matrix.conjugate().T ) def _lowerCamelCase( lowercase__ , lowercase__ ) -> Any: '''simple docstring''' __lowercase= v.conjugate().T __lowercase= v_star.dot(lowercase__ ) assert isinstance(lowercase__ , np.ndarray ) return (v_star_dot.dot(lowercase__ )) / (v_star.dot(lowercase__ )) def _lowerCamelCase( ) -> None: '''simple docstring''' __lowercase= np.array([[2, 2 + 1j, 4], [2 - 1j, 3, 1j], [4, -1j, 1]] ) __lowercase= np.array([[1], [2], [3]] ) assert is_hermitian(lowercase__ ), F'{a} is not hermitian.' print(rayleigh_quotient(lowercase__ , lowercase__ ) ) __lowercase= np.array([[1, 2, 4], [2, 3, -1], [4, -1, 1]] ) assert is_hermitian(lowercase__ ), F'{a} is not hermitian.' assert rayleigh_quotient(lowercase__ , lowercase__ ) == float(3 ) if __name__ == "__main__": import doctest doctest.testmod() tests()
304
0
import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class __lowerCAmelCase ( unittest.TestCase ): def __init__( self , lowerCAmelCase , lowerCAmelCase=7 , lowerCAmelCase=3 , lowerCAmelCase=18 , lowerCAmelCase=30 , lowerCAmelCase=400 , lowerCAmelCase=True , lowerCAmelCase=None , lowerCAmelCase=True , lowerCAmelCase=None , lowerCAmelCase=True , lowerCAmelCase=[0.5, 0.5, 0.5] , lowerCAmelCase=[0.5, 0.5, 0.5] , lowerCAmelCase=False , ) -> Optional[int]: '''simple docstring''' _lowercase =size if size is not None else {'height': 20, 'width': 20} _lowercase =crop_size if crop_size is not None else {'height': 18, 'width': 18} _lowercase =parent _lowercase =batch_size _lowercase =num_channels _lowercase =image_size _lowercase =min_resolution _lowercase =max_resolution _lowercase =do_resize _lowercase =size _lowercase =do_center_crop _lowercase =crop_size _lowercase =do_normalize _lowercase =image_mean _lowercase =image_std _lowercase =do_reduce_labels def A__ ( self ) -> Optional[Any]: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_reduce_labels": self.do_reduce_labels, } def a ( ) -> Tuple: """simple docstring""" _lowercase =load_dataset('hf-internal-testing/fixtures_ade20k' , split='test' ) _lowercase =Image.open(dataset[0]['file'] ) _lowercase =Image.open(dataset[1]['file'] ) return image, map def a ( ) -> List[str]: """simple docstring""" _lowercase =load_dataset('hf-internal-testing/fixtures_ade20k' , split='test' ) _lowercase =Image.open(ds[0]['file'] ) _lowercase =Image.open(ds[1]['file'] ) _lowercase =Image.open(ds[2]['file'] ) _lowercase =Image.open(ds[3]['file'] ) return [imagea, imagea], [mapa, mapa] @require_torch @require_vision class __lowerCAmelCase ( SCREAMING_SNAKE_CASE , unittest.TestCase ): _a = BeitImageProcessor if is_vision_available() else None def A__ ( self ) -> str: '''simple docstring''' _lowercase =BeitImageProcessingTester(self ) @property def A__ ( self ) -> Tuple: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def A__ ( self ) -> Union[str, Any]: '''simple docstring''' _lowercase =self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowerCAmelCase , 'do_resize' ) ) self.assertTrue(hasattr(lowerCAmelCase , 'size' ) ) self.assertTrue(hasattr(lowerCAmelCase , 'do_center_crop' ) ) self.assertTrue(hasattr(lowerCAmelCase , 'center_crop' ) ) self.assertTrue(hasattr(lowerCAmelCase , 'do_normalize' ) ) self.assertTrue(hasattr(lowerCAmelCase , 'image_mean' ) ) self.assertTrue(hasattr(lowerCAmelCase , 'image_std' ) ) def A__ ( self ) -> List[str]: '''simple docstring''' _lowercase =self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'height': 20, 'width': 20} ) self.assertEqual(image_processor.crop_size , {'height': 18, 'width': 18} ) self.assertEqual(image_processor.do_reduce_labels , lowerCAmelCase ) _lowercase =self.image_processing_class.from_dict( self.image_processor_dict , size=42 , crop_size=84 , reduce_labels=lowerCAmelCase ) self.assertEqual(image_processor.size , {'height': 42, 'width': 42} ) self.assertEqual(image_processor.crop_size , {'height': 84, 'width': 84} ) self.assertEqual(image_processor.do_reduce_labels , lowerCAmelCase ) def A__ ( self ) -> Dict: '''simple docstring''' pass def A__ ( self ) -> List[Any]: '''simple docstring''' _lowercase =self.image_processing_class(**self.image_processor_dict ) # create random PIL images _lowercase =prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase , Image.Image ) # Test not batched input _lowercase =image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched _lowercase =image_processing(lowerCAmelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def A__ ( self ) -> List[Any]: '''simple docstring''' _lowercase =self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _lowercase =prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase , numpify=lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase , np.ndarray ) # Test not batched input _lowercase =image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched _lowercase =image_processing(lowerCAmelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def A__ ( self ) -> List[Any]: '''simple docstring''' _lowercase =self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowercase =prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase , torchify=lowerCAmelCase ) for image in image_inputs: self.assertIsInstance(lowerCAmelCase , torch.Tensor ) # Test not batched input _lowercase =image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) # Test batched _lowercase =image_processing(lowerCAmelCase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) def A__ ( self ) -> Optional[Any]: '''simple docstring''' _lowercase =self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _lowercase =prepare_image_inputs(self.image_processor_tester , equal_resolution=lowerCAmelCase , torchify=lowerCAmelCase ) _lowercase =[] for image in image_inputs: self.assertIsInstance(lowerCAmelCase , torch.Tensor ) maps.append(torch.zeros(image.shape[-2:] ).long() ) # Test not batched input _lowercase =image_processing(image_inputs[0] , maps[0] , return_tensors='pt' ) self.assertEqual( encoding['pixel_values'].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual( encoding['labels'].shape , ( 1, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual(encoding['labels'].dtype , torch.long ) self.assertTrue(encoding['labels'].min().item() >= 0 ) self.assertTrue(encoding['labels'].max().item() <= 255 ) # Test batched _lowercase =image_processing(lowerCAmelCase , lowerCAmelCase , return_tensors='pt' ) self.assertEqual( encoding['pixel_values'].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual( encoding['labels'].shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual(encoding['labels'].dtype , torch.long ) self.assertTrue(encoding['labels'].min().item() >= 0 ) self.assertTrue(encoding['labels'].max().item() <= 255 ) # Test not batched input (PIL images) _lowercase , _lowercase =prepare_semantic_single_inputs() _lowercase =image_processing(lowerCAmelCase , lowerCAmelCase , return_tensors='pt' ) self.assertEqual( encoding['pixel_values'].shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual( encoding['labels'].shape , ( 1, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual(encoding['labels'].dtype , torch.long ) self.assertTrue(encoding['labels'].min().item() >= 0 ) self.assertTrue(encoding['labels'].max().item() <= 255 ) # Test batched input (PIL images) _lowercase , _lowercase =prepare_semantic_batch_inputs() _lowercase =image_processing(lowerCAmelCase , lowerCAmelCase , return_tensors='pt' ) self.assertEqual( encoding['pixel_values'].shape , ( 2, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual( encoding['labels'].shape , ( 2, self.image_processor_tester.crop_size['height'], self.image_processor_tester.crop_size['width'], ) , ) self.assertEqual(encoding['labels'].dtype , torch.long ) self.assertTrue(encoding['labels'].min().item() >= 0 ) self.assertTrue(encoding['labels'].max().item() <= 255 ) def A__ ( self ) -> List[Any]: '''simple docstring''' _lowercase =self.image_processing_class(**self.image_processor_dict ) # ADE20k has 150 classes, and the background is included, so labels should be between 0 and 150 _lowercase , _lowercase =prepare_semantic_single_inputs() _lowercase =image_processing(lowerCAmelCase , lowerCAmelCase , return_tensors='pt' ) self.assertTrue(encoding['labels'].min().item() >= 0 ) self.assertTrue(encoding['labels'].max().item() <= 150 ) _lowercase =True _lowercase =image_processing(lowerCAmelCase , lowerCAmelCase , return_tensors='pt' ) self.assertTrue(encoding['labels'].min().item() >= 0 ) self.assertTrue(encoding['labels'].max().item() <= 255 )
205
import argparse import os import re lowercase_ = 'src/transformers' # Pattern that looks at the indentation in a line. lowercase_ = re.compile(R'^(\s*)\S') # Pattern that matches `"key":" and puts `key` in group 0. lowercase_ = re.compile(R'^\s*"([^"]+)":') # Pattern that matches `_import_structure["key"]` and puts `key` in group 0. lowercase_ = re.compile(R'^\s*_import_structure\["([^"]+)"\]') # Pattern that matches `"key",` and puts `key` in group 0. lowercase_ = re.compile(R'^\s*"([^"]+)",\s*$') # Pattern that matches any `[stuff]` and puts `stuff` in group 0. lowercase_ = re.compile(R'\[([^\]]+)\]') def a ( A__ : Dict ) -> Optional[Any]: """simple docstring""" _lowercase =_re_indent.search(A__ ) return "" if search is None else search.groups()[0] def a ( A__ : Optional[Any] , A__ : Dict="" , A__ : Union[str, Any]=None , A__ : Tuple=None ) -> Dict: """simple docstring""" _lowercase =0 _lowercase =code.split('\n' ) if start_prompt is not None: while not lines[index].startswith(A__ ): index += 1 _lowercase =['\n'.join(lines[:index] )] else: _lowercase =[] # We split into blocks until we get to the `end_prompt` (or the end of the block). _lowercase =[lines[index]] index += 1 while index < len(A__ ) and (end_prompt is None or not lines[index].startswith(A__ )): if len(lines[index] ) > 0 and get_indent(lines[index] ) == indent_level: if len(A__ ) > 0 and get_indent(current_block[-1] ).startswith(indent_level + ' ' ): current_block.append(lines[index] ) blocks.append('\n'.join(A__ ) ) if index < len(A__ ) - 1: _lowercase =[lines[index + 1]] index += 1 else: _lowercase =[] else: blocks.append('\n'.join(A__ ) ) _lowercase =[lines[index]] else: current_block.append(lines[index] ) index += 1 # Adds current block if it's nonempty. if len(A__ ) > 0: blocks.append('\n'.join(A__ ) ) # Add final block after end_prompt if provided. if end_prompt is not None and index < len(A__ ): blocks.append('\n'.join(lines[index:] ) ) return blocks def a ( A__ : int ) -> Union[str, Any]: """simple docstring""" def _inner(A__ : Any ): return key(A__ ).lower().replace('_' , '' ) return _inner def a ( A__ : Any , A__ : Union[str, Any]=None ) -> int: """simple docstring""" def noop(A__ : Optional[int] ): return x if key is None: _lowercase =noop # Constants are all uppercase, they go first. _lowercase =[obj for obj in objects if key(A__ ).isupper()] # Classes are not all uppercase but start with a capital, they go second. _lowercase =[obj for obj in objects if key(A__ )[0].isupper() and not key(A__ ).isupper()] # Functions begin with a lowercase, they go last. _lowercase =[obj for obj in objects if not key(A__ )[0].isupper()] _lowercase =ignore_underscore(A__ ) return sorted(A__ , key=A__ ) + sorted(A__ , key=A__ ) + sorted(A__ , key=A__ ) def a ( A__ : Union[str, Any] ) -> Tuple: """simple docstring""" def _replace(A__ : Optional[int] ): _lowercase =match.groups()[0] if "," not in imports: return F'''[{imports}]''' _lowercase =[part.strip().replace('"' , '' ) for part in imports.split(',' )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: _lowercase =keys[:-1] return "[" + ", ".join([F'''"{k}"''' for k in sort_objects(A__ )] ) + "]" _lowercase =import_statement.split('\n' ) if len(A__ ) > 3: # Here we have to sort internal imports that are on several lines (one per name): # key: [ # "object1", # "object2", # ... # ] # We may have to ignore one or two lines on each side. _lowercase =2 if lines[1].strip() == '[' else 1 _lowercase =[(i, _re_strip_line.search(A__ ).groups()[0]) for i, line in enumerate(lines[idx:-idx] )] _lowercase =sort_objects(A__ , key=lambda A__ : x[1] ) _lowercase =[lines[x[0] + idx] for x in sorted_indices] return "\n".join(lines[:idx] + sorted_lines + lines[-idx:] ) elif len(A__ ) == 3: # Here we have to sort internal imports that are on one separate line: # key: [ # "object1", "object2", ... # ] if _re_bracket_content.search(lines[1] ) is not None: _lowercase =_re_bracket_content.sub(_replace , lines[1] ) else: _lowercase =[part.strip().replace('"' , '' ) for part in lines[1].split(',' )] # We will have a final empty element if the line finished with a comma. if len(keys[-1] ) == 0: _lowercase =keys[:-1] _lowercase =get_indent(lines[1] ) + ', '.join([F'''"{k}"''' for k in sort_objects(A__ )] ) return "\n".join(A__ ) else: # Finally we have to deal with imports fitting on one line _lowercase =_re_bracket_content.sub(_replace , A__ ) return import_statement def a ( A__ : Dict , A__ : int=True ) -> Optional[Any]: """simple docstring""" with open(A__ , encoding='utf-8' ) as f: _lowercase =f.read() if "_import_structure" not in code: return # Blocks of indent level 0 _lowercase =split_code_in_indented_blocks( A__ , start_prompt='_import_structure = {' , end_prompt='if TYPE_CHECKING:' ) # We ignore block 0 (everything untils start_prompt) and the last block (everything after end_prompt). for block_idx in range(1 , len(A__ ) - 1 ): # Check if the block contains some `_import_structure`s thingy to sort. _lowercase =main_blocks[block_idx] _lowercase =block.split('\n' ) # Get to the start of the imports. _lowercase =0 while line_idx < len(A__ ) and "_import_structure" not in block_lines[line_idx]: # Skip dummy import blocks if "import dummy" in block_lines[line_idx]: _lowercase =len(A__ ) else: line_idx += 1 if line_idx >= len(A__ ): continue # Ignore beginning and last line: they don't contain anything. _lowercase ='\n'.join(block_lines[line_idx:-1] ) _lowercase =get_indent(block_lines[1] ) # Slit the internal block into blocks of indent level 1. _lowercase =split_code_in_indented_blocks(A__ , indent_level=A__ ) # We have two categories of import key: list or _import_structure[key].append/extend _lowercase =_re_direct_key if '_import_structure = {' in block_lines[0] else _re_indirect_key # Grab the keys, but there is a trap: some lines are empty or just comments. _lowercase =[(pattern.search(A__ ).groups()[0] if pattern.search(A__ ) is not None else None) for b in internal_blocks] # We only sort the lines with a key. _lowercase =[(i, key) for i, key in enumerate(A__ ) if key is not None] _lowercase =[x[0] for x in sorted(A__ , key=lambda A__ : x[1] )] # We reorder the blocks by leaving empty lines/comments as they were and reorder the rest. _lowercase =0 _lowercase =[] for i in range(len(A__ ) ): if keys[i] is None: reorderded_blocks.append(internal_blocks[i] ) else: _lowercase =sort_objects_in_import(internal_blocks[sorted_indices[count]] ) reorderded_blocks.append(A__ ) count += 1 # And we put our main block back together with its first and last line. _lowercase ='\n'.join(block_lines[:line_idx] + reorderded_blocks + [block_lines[-1]] ) if code != "\n".join(A__ ): if check_only: return True else: print(F'''Overwriting {file}.''' ) with open(A__ , 'w' , encoding='utf-8' ) as f: f.write('\n'.join(A__ ) ) def a ( A__ : List[Any]=True ) -> List[str]: """simple docstring""" _lowercase =[] for root, _, files in os.walk(A__ ): if "__init__.py" in files: _lowercase =sort_imports(os.path.join(A__ , '__init__.py' ) , check_only=A__ ) if result: _lowercase =[os.path.join(A__ , '__init__.py' )] if len(A__ ) > 0: raise ValueError(F'''Would overwrite {len(A__ )} files, run `make style`.''' ) if __name__ == "__main__": lowercase_ = argparse.ArgumentParser() parser.add_argument('--check_only', action='store_true', help='Whether to only check or fix style.') lowercase_ = parser.parse_args() sort_imports_in_all_inits(check_only=args.check_only)
205
1
from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef import datasets _UpperCAmelCase : int = """\ @inproceedings{wang2019glue, title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding}, author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.}, note={In the Proceedings of ICLR.}, year={2019} } """ _UpperCAmelCase : Tuple = """\ GLUE, the General Language Understanding Evaluation benchmark (https://gluebenchmark.com/) is a collection of resources for training, evaluating, and analyzing natural language understanding systems. """ _UpperCAmelCase : List[Any] = """ Compute GLUE evaluation metric associated to each GLUE dataset. Args: predictions: list of predictions to score. Each translation should be tokenized into a list of tokens. references: list of lists of references for each translation. Each reference should be tokenized into a list of tokens. Returns: depending on the GLUE subset, one or several of: \"accuracy\": Accuracy \"f1\": F1 score \"pearson\": Pearson Correlation \"spearmanr\": Spearman Correlation \"matthews_correlation\": Matthew Correlation Examples: >>> glue_metric = datasets.load_metric('glue', 'sst2') # 'sst2' or any of [\"mnli\", \"mnli_mismatched\", \"mnli_matched\", \"qnli\", \"rte\", \"wnli\", \"hans\"] >>> references = [0, 1] >>> predictions = [0, 1] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0} >>> glue_metric = datasets.load_metric('glue', 'mrpc') # 'mrpc' or 'qqp' >>> references = [0, 1] >>> predictions = [0, 1] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'accuracy': 1.0, 'f1': 1.0} >>> glue_metric = datasets.load_metric('glue', 'stsb') >>> references = [0., 1., 2., 3., 4., 5.] >>> predictions = [0., 1., 2., 3., 4., 5.] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print({\"pearson\": round(results[\"pearson\"], 2), \"spearmanr\": round(results[\"spearmanr\"], 2)}) {'pearson': 1.0, 'spearmanr': 1.0} >>> glue_metric = datasets.load_metric('glue', 'cola') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = glue_metric.compute(predictions=predictions, references=references) >>> print(results) {'matthews_correlation': 1.0} """ def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Optional[Any]: return float((preds == labels).mean() ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> List[Any]: lowerCamelCase__ : Optional[int] = simple_accuracy(_UpperCAmelCase , _UpperCAmelCase ) lowerCamelCase__ : Tuple = float(fa_score(y_true=_UpperCAmelCase , y_pred=_UpperCAmelCase ) ) return { "accuracy": acc, "f1": fa, } def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase ) -> Optional[int]: lowerCamelCase__ : List[str] = float(pearsonr(_UpperCAmelCase , _UpperCAmelCase )[0] ) lowerCamelCase__ : Optional[int] = float(spearmanr(_UpperCAmelCase , _UpperCAmelCase )[0] ) return { "pearson": pearson_corr, "spearmanr": spearman_corr, } @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCAmelCase ( datasets.Metric ): def A_ ( self : str ) -> Tuple: if self.config_name not in [ "sst2", "mnli", "mnli_mismatched", "mnli_matched", "cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans", ]: raise KeyError( 'You should supply a configuration name selected in ' '["sst2", "mnli", "mnli_mismatched", "mnli_matched", ' '"cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans"]' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('int64' if self.config_name != 'stsb' else 'float32' ), 'references': datasets.Value('int64' if self.config_name != 'stsb' else 'float32' ), } ) , codebase_urls=[] , reference_urls=[] , format='numpy' , ) def A_ ( self : int , UpperCAmelCase : List[str] , UpperCAmelCase : Dict ) -> Tuple: if self.config_name == "cola": return {"matthews_correlation": matthews_corrcoef(UpperCAmelCase , UpperCAmelCase )} elif self.config_name == "stsb": return pearson_and_spearman(UpperCAmelCase , UpperCAmelCase ) elif self.config_name in ["mrpc", "qqp"]: return acc_and_fa(UpperCAmelCase , UpperCAmelCase ) elif self.config_name in ["sst2", "mnli", "mnli_mismatched", "mnli_matched", "qnli", "rte", "wnli", "hans"]: return {"accuracy": simple_accuracy(UpperCAmelCase , UpperCAmelCase )} else: raise KeyError( 'You should supply a configuration name selected in ' '["sst2", "mnli", "mnli_mismatched", "mnli_matched", ' '"cola", "stsb", "mrpc", "qqp", "qnli", "rte", "wnli", "hans"]' )
365
from collections import deque class lowerCAmelCase : def __init__( self : str , UpperCAmelCase : str , UpperCAmelCase : int , UpperCAmelCase : int ) -> None: lowerCamelCase__ : Optional[int] = process_name # process name lowerCamelCase__ : Optional[int] = arrival_time # arrival time of the process # completion time of finished process or last interrupted time lowerCamelCase__ : str = arrival_time lowerCamelCase__ : List[Any] = burst_time # remaining burst time lowerCamelCase__ : Any = 0 # total time of the process wait in ready queue lowerCamelCase__ : Tuple = 0 # time from arrival time to completion time class lowerCAmelCase : def __init__( self : List[str] , UpperCAmelCase : int , UpperCAmelCase : list[int] , UpperCAmelCase : deque[Process] , UpperCAmelCase : int , ) -> None: # total number of mlfq's queues lowerCamelCase__ : Optional[int] = number_of_queues # time slice of queues that round robin algorithm applied lowerCamelCase__ : List[str] = time_slices # unfinished process is in this ready_queue lowerCamelCase__ : List[str] = queue # current time lowerCamelCase__ : Optional[Any] = current_time # finished process is in this sequence queue lowerCamelCase__ : deque[Process] = deque() def A_ ( self : Tuple ) -> list[str]: lowerCamelCase__ : Union[str, Any] = [] for i in range(len(self.finish_queue ) ): sequence.append(self.finish_queue[i].process_name ) return sequence def A_ ( self : Tuple , UpperCAmelCase : list[Process] ) -> list[int]: lowerCamelCase__ : Tuple = [] for i in range(len(UpperCAmelCase ) ): waiting_times.append(queue[i].waiting_time ) return waiting_times def A_ ( self : Union[str, Any] , UpperCAmelCase : list[Process] ) -> list[int]: lowerCamelCase__ : int = [] for i in range(len(UpperCAmelCase ) ): turnaround_times.append(queue[i].turnaround_time ) return turnaround_times def A_ ( self : Optional[int] , UpperCAmelCase : list[Process] ) -> list[int]: lowerCamelCase__ : Tuple = [] for i in range(len(UpperCAmelCase ) ): completion_times.append(queue[i].stop_time ) return completion_times def A_ ( self : str , UpperCAmelCase : deque[Process] ) -> list[int]: return [q.burst_time for q in queue] def A_ ( self : int , UpperCAmelCase : Process ) -> int: process.waiting_time += self.current_time - process.stop_time return process.waiting_time def A_ ( self : Optional[int] , UpperCAmelCase : deque[Process] ) -> deque[Process]: lowerCamelCase__ : deque[Process] = deque() # sequence deque of finished process while len(UpperCAmelCase ) != 0: lowerCamelCase__ : List[Any] = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of current process self.update_waiting_time(UpperCAmelCase ) # update current time self.current_time += cp.burst_time # finish the process and set the process's burst-time 0 lowerCamelCase__ : Optional[int] = 0 # set the process's turnaround time because it is finished lowerCamelCase__ : Union[str, Any] = self.current_time - cp.arrival_time # set the completion time lowerCamelCase__ : Any = self.current_time # add the process to queue that has finished queue finished.append(UpperCAmelCase ) self.finish_queue.extend(UpperCAmelCase ) # add finished process to finish queue # FCFS will finish all remaining processes return finished def A_ ( self : str , UpperCAmelCase : deque[Process] , UpperCAmelCase : int ) -> tuple[deque[Process], deque[Process]]: lowerCamelCase__ : deque[Process] = deque() # sequence deque of terminated process # just for 1 cycle and unfinished processes will go back to queue for _ in range(len(UpperCAmelCase ) ): lowerCamelCase__ : Dict = ready_queue.popleft() # current process # if process's arrival time is later than current time, update current time if self.current_time < cp.arrival_time: self.current_time += cp.arrival_time # update waiting time of unfinished processes self.update_waiting_time(UpperCAmelCase ) # if the burst time of process is bigger than time-slice if cp.burst_time > time_slice: # use CPU for only time-slice self.current_time += time_slice # update remaining burst time cp.burst_time -= time_slice # update end point time lowerCamelCase__ : List[str] = self.current_time # locate the process behind the queue because it is not finished ready_queue.append(UpperCAmelCase ) else: # use CPU for remaining burst time self.current_time += cp.burst_time # set burst time 0 because the process is finished lowerCamelCase__ : Any = 0 # set the finish time lowerCamelCase__ : int = self.current_time # update the process' turnaround time because it is finished lowerCamelCase__ : Dict = self.current_time - cp.arrival_time # add the process to queue that has finished queue finished.append(UpperCAmelCase ) self.finish_queue.extend(UpperCAmelCase ) # add finished process to finish queue # return finished processes queue and remaining processes queue return finished, ready_queue def A_ ( self : Dict ) -> deque[Process]: # all queues except last one have round_robin algorithm for i in range(self.number_of_queues - 1 ): lowerCamelCase__ , lowerCamelCase__ : Any = self.round_robin( self.ready_queue , self.time_slices[i] ) # the last queue has first_come_first_served algorithm self.first_come_first_served(self.ready_queue ) return self.finish_queue if __name__ == "__main__": import doctest _UpperCAmelCase : List[str] = Process("""P1""", 0, 53) _UpperCAmelCase : Union[str, Any] = Process("""P2""", 0, 17) _UpperCAmelCase : int = Process("""P3""", 0, 68) _UpperCAmelCase : str = Process("""P4""", 0, 24) _UpperCAmelCase : Optional[int] = 3 _UpperCAmelCase : Optional[Any] = [17, 25] _UpperCAmelCase : Optional[int] = deque([Pa, Pa, Pa, Pa]) if len(time_slices) != number_of_queues - 1: raise SystemExit(0) doctest.testmod(extraglobs={"""queue""": deque([Pa, Pa, Pa, Pa])}) _UpperCAmelCase : Tuple = Process("""P1""", 0, 53) _UpperCAmelCase : Any = Process("""P2""", 0, 17) _UpperCAmelCase : Any = Process("""P3""", 0, 68) _UpperCAmelCase : List[Any] = Process("""P4""", 0, 24) _UpperCAmelCase : List[str] = 3 _UpperCAmelCase : Optional[int] = [17, 25] _UpperCAmelCase : Optional[int] = deque([Pa, Pa, Pa, Pa]) _UpperCAmelCase : Union[str, Any] = MLFQ(number_of_queues, time_slices, queue, 0) _UpperCAmelCase : Dict = mlfq.multi_level_feedback_queue() # print total waiting times of processes(P1, P2, P3, P4) print( F"""waiting time:\ \t\t\t{MLFQ.calculate_waiting_time(mlfq, [Pa, Pa, Pa, Pa])}""" ) # print completion times of processes(P1, P2, P3, P4) print( F"""completion time:\ \t\t{MLFQ.calculate_completion_time(mlfq, [Pa, Pa, Pa, Pa])}""" ) # print total turnaround times of processes(P1, P2, P3, P4) print( F"""turnaround time:\ \t\t{MLFQ.calculate_turnaround_time(mlfq, [Pa, Pa, Pa, Pa])}""" ) # print sequence of finished processes print( F"""sequence of finished processes:\ {mlfq.calculate_sequence_of_finish_queue()}""" )
45
0
import math import os import unittest from transformers import MegatronBertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MODEL_FOR_PRETRAINING_MAPPING, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, ) class lowercase : def __init__( self , _a , _a=13 , _a=7 , _a=True , _a=True , _a=True , _a=True , _a=99 , _a=64 , _a=32 , _a=5 , _a=4 , _a=37 , _a="gelu" , _a=0.1 , _a=0.1 , _a=512 , _a=16 , _a=2 , _a=0.02 , _a=3 , _a=4 , _a=None , ) -> Optional[int]: _A : Optional[int] = parent _A : Optional[Any] = batch_size _A : Tuple = seq_length _A : Optional[Any] = is_training _A : List[Any] = use_input_mask _A : Any = use_token_type_ids _A : List[Any] = use_labels _A : Optional[int] = vocab_size _A : str = hidden_size _A : List[str] = embedding_size _A : List[str] = num_hidden_layers _A : List[str] = num_attention_heads _A : Dict = intermediate_size _A : Dict = hidden_act _A : Any = hidden_dropout_prob _A : List[Any] = attention_probs_dropout_prob _A : str = max_position_embeddings _A : List[str] = type_vocab_size _A : int = type_sequence_label_size _A : Optional[Any] = initializer_range _A : Union[str, Any] = num_labels _A : Any = num_choices _A : List[str] = scope def a__ ( self ) -> Union[str, Any]: _A : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _A : List[str] = None if self.use_input_mask: _A : Tuple = random_attention_mask([self.batch_size, self.seq_length] ) _A : Union[str, Any] = None if self.use_token_type_ids: _A : Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _A : Any = None _A : List[Any] = None _A : Optional[Any] = None if self.use_labels: _A : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _A : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _A : List[str] = ids_tensor([self.batch_size] , self.num_choices ) _A : List[Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def a__ ( self ) -> Optional[int]: return MegatronBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , embedding_size=self.embedding_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=_a , initializer_range=self.initializer_range , ) def a__ ( self , _a , _a , _a , _a , _a , _a , _a ) -> Any: _A : Tuple = MegatronBertModel(config=_a ) model.to(_a ) model.eval() _A : str = model(_a , attention_mask=_a , token_type_ids=_a ) _A : Any = model(_a , token_type_ids=_a ) _A : int = model(_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def a__ ( self , _a , _a , _a , _a , _a , _a , _a ) -> List[str]: _A : List[Any] = MegatronBertForMaskedLM(config=_a ) model.to(_a ) model.eval() _A : List[Any] = model(_a , attention_mask=_a , token_type_ids=_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a__ ( self , _a , _a , _a , _a , _a , _a , _a ) -> Union[str, Any]: _A : Optional[int] = MegatronBertForCausalLM(config=_a ) model.to(_a ) model.eval() _A : List[Any] = model(_a , attention_mask=_a , token_type_ids=_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def a__ ( self , _a , _a , _a , _a , _a , _a , _a ) -> Union[str, Any]: _A : List[str] = MegatronBertForNextSentencePrediction(config=_a ) model.to(_a ) model.eval() _A : List[str] = model( _a , attention_mask=_a , token_type_ids=_a , labels=_a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def a__ ( self , _a , _a , _a , _a , _a , _a , _a ) -> List[str]: _A : Dict = MegatronBertForPreTraining(config=_a ) model.to(_a ) model.eval() _A : str = model( _a , attention_mask=_a , token_type_ids=_a , labels=_a , next_sentence_label=_a , ) self.parent.assertEqual(result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def a__ ( self , _a , _a , _a , _a , _a , _a , _a ) -> List[str]: _A : Dict = MegatronBertForQuestionAnswering(config=_a ) model.to(_a ) model.eval() _A : Dict = model( _a , attention_mask=_a , token_type_ids=_a , start_positions=_a , end_positions=_a , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def a__ ( self , _a , _a , _a , _a , _a , _a , _a ) -> Tuple: _A : Tuple = self.num_labels _A : Any = MegatronBertForSequenceClassification(_a ) model.to(_a ) model.eval() _A : str = model(_a , attention_mask=_a , token_type_ids=_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def a__ ( self , _a , _a , _a , _a , _a , _a , _a ) -> int: _A : Union[str, Any] = self.num_labels _A : Optional[int] = MegatronBertForTokenClassification(config=_a ) model.to(_a ) model.eval() _A : Dict = model(_a , attention_mask=_a , token_type_ids=_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def a__ ( self , _a , _a , _a , _a , _a , _a , _a ) -> Union[str, Any]: _A : str = self.num_choices _A : Optional[Any] = MegatronBertForMultipleChoice(config=_a ) model.to(_a ) model.eval() _A : List[str] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A : int = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A : Any = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _A : Optional[int] = model( _a , attention_mask=_a , token_type_ids=_a , labels=_a , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def a__ ( self ) -> int: _A : Optional[int] = self.prepare_config_and_inputs() ( ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ( _A ) , ) : int = config_and_inputs _A : Optional[int] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class lowercase ( UpperCamelCase__,UpperCamelCase__,unittest.TestCase ): _a = ( ( MegatronBertModel, MegatronBertForMaskedLM, MegatronBertForCausalLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, ) if is_torch_available() else () ) _a = ( { "feature-extraction": MegatronBertModel, "fill-mask": MegatronBertForMaskedLM, "question-answering": MegatronBertForQuestionAnswering, "text-classification": MegatronBertForSequenceClassification, "text-generation": MegatronBertForCausalLM, "token-classification": MegatronBertForTokenClassification, "zero-shot": MegatronBertForSequenceClassification, } if is_torch_available() else {} ) _a = True # test_resize_embeddings = False _a = False def a__ ( self , _a , _a , _a=False ) -> Optional[int]: _A : List[str] = super()._prepare_for_class(_a , _a , return_labels=_a ) if return_labels: if model_class in get_values(_a ): _A : Any = torch.zeros( (self.model_tester.batch_size, self.model_tester.seq_length) , dtype=torch.long , device=_a ) _A : Any = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=_a ) return inputs_dict def a__ ( self ) -> Dict: _A : Dict = MegatronBertModelTester(self ) _A : str = ConfigTester(self , config_class=_a , hidden_size=37 ) def a__ ( self ) -> str: self.config_tester.run_common_tests() def a__ ( self ) -> Optional[int]: _A : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_model(*_a ) def a__ ( self ) -> List[Any]: _A : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_masked_lm(*_a ) def a__ ( self ) -> Dict: _A : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_multiple_choice(*_a ) def a__ ( self ) -> int: _A : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_next_sequence_prediction(*_a ) def a__ ( self ) -> List[str]: _A : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_pretraining(*_a ) def a__ ( self ) -> int: _A : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_question_answering(*_a ) def a__ ( self ) -> Dict: _A : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_sequence_classification(*_a ) def a__ ( self ) -> int: _A : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_megatron_bert_for_token_classification(*_a ) def lowerCAmelCase_ ( snake_case_ ): return torch.tensor( snake_case_,dtype=torch.long,device=snake_case_,) _snake_case = 1e-4 @require_torch @require_sentencepiece @require_tokenizers class lowercase ( unittest.TestCase ): @slow @unittest.skip("""Model is not available.""" ) def a__ ( self ) -> List[Any]: _A : Optional[Any] = """nvidia/megatron-bert-uncased-345m""" if "MYDIR" in os.environ: _A : List[str] = os.path.join(os.environ["""MYDIR"""] , _a ) _A : Any = MegatronBertModel.from_pretrained(_a ) model.to(_a ) model.half() _A : Optional[int] = _long_tensor([[101, 7110, 1005, 1056, 2023, 1_1333, 1_7413, 1029, 102]] ) with torch.no_grad(): _A : Dict = model(_a )[0] _A : List[str] = torch.Size((1, 9, 1024) ) self.assertEqual(output.shape , _a ) _A : Tuple = [-0.6040, -0.2517, -0.1025, 0.3420, -0.6758, -0.0017, -0.1089, -0.1990, 0.5728] for ii in range(3 ): for jj in range(3 ): _A : Any = output[0, ii, jj] _A : Optional[int] = expected[3 * ii + jj] _A : Tuple = """ii={} jj={} a={} b={}""".format(_a , _a , _a , _a ) self.assertTrue(math.isclose(_a , _a , rel_tol=_a , abs_tol=_a ) , msg=_a )
26
"""simple docstring""" from dataclasses import dataclass from typing import Dict, Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .attention_processor import AttentionProcessor, AttnProcessor from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, DiagonalGaussianDistribution, Encoder @dataclass class snake_case__ ( snake_case_ ): _snake_case : "DiagonalGaussianDistribution" class snake_case__ ( snake_case_, snake_case_ ): _snake_case : Optional[Any] = True @register_to_config def __init__( self , lowerCamelCase = 3 , lowerCamelCase = 3 , lowerCamelCase = ("DownEncoderBlock2D",) , lowerCamelCase = ("UpDecoderBlock2D",) , lowerCamelCase = (64,) , lowerCamelCase = 1 , lowerCamelCase = "silu" , lowerCamelCase = 4 , lowerCamelCase = 32 , lowerCamelCase = 32 , lowerCamelCase = 0.1_8215 , ): super().__init__() # pass init params to Encoder __a = Encoder( in_channels=lowerCamelCase , out_channels=lowerCamelCase , down_block_types=lowerCamelCase , block_out_channels=lowerCamelCase , layers_per_block=lowerCamelCase , act_fn=lowerCamelCase , norm_num_groups=lowerCamelCase , double_z=lowerCamelCase , ) # pass init params to Decoder __a = Decoder( in_channels=lowerCamelCase , out_channels=lowerCamelCase , up_block_types=lowerCamelCase , block_out_channels=lowerCamelCase , layers_per_block=lowerCamelCase , norm_num_groups=lowerCamelCase , act_fn=lowerCamelCase , ) __a = nn.Convad(2 * latent_channels , 2 * latent_channels , 1 ) __a = nn.Convad(lowerCamelCase , lowerCamelCase , 1 ) __a = False __a = False # only relevant if vae tiling is enabled __a = self.config.sample_size __a = ( self.config.sample_size[0] if isinstance(self.config.sample_size , (list, tuple) ) else self.config.sample_size ) __a = int(sample_size / (2 ** (len(self.config.block_out_channels ) - 1)) ) __a = 0.25 def a__ ( self , lowerCamelCase , lowerCamelCase=False ): if isinstance(lowerCamelCase , (Encoder, Decoder) ): __a = value def a__ ( self , lowerCamelCase = True ): __a = use_tiling def a__ ( self ): self.enable_tiling(lowerCamelCase ) def a__ ( self ): __a = True def a__ ( self ): __a = False @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def a__ ( self ): __a = {} def fn_recursive_add_processors(lowerCamelCase , lowerCamelCase , lowerCamelCase ): if hasattr(lowerCamelCase , "set_processor" ): __a = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(F"{name}.{sub_name}" , lowerCamelCase , lowerCamelCase ) return processors for name, module in self.named_children(): fn_recursive_add_processors(lowerCamelCase , lowerCamelCase , lowerCamelCase ) return processors def a__ ( self , lowerCamelCase ): __a = len(self.attn_processors.keys() ) if isinstance(lowerCamelCase , lowerCamelCase ) and len(lowerCamelCase ) != count: raise ValueError( F"A dict of processors was passed, but the number of processors {len(lowerCamelCase )} does not match the" F" number of attention layers: {count}. Please make sure to pass {count} processor classes." ) def fn_recursive_attn_processor(lowerCamelCase , lowerCamelCase , lowerCamelCase ): if hasattr(lowerCamelCase , "set_processor" ): if not isinstance(lowerCamelCase , lowerCamelCase ): module.set_processor(lowerCamelCase ) else: module.set_processor(processor.pop(F"{name}.processor" ) ) for sub_name, child in module.named_children(): fn_recursive_attn_processor(F"{name}.{sub_name}" , lowerCamelCase , lowerCamelCase ) for name, module in self.named_children(): fn_recursive_attn_processor(lowerCamelCase , lowerCamelCase , lowerCamelCase ) def a__ ( self ): self.set_attn_processor(AttnProcessor() ) @apply_forward_hook def a__ ( self , lowerCamelCase , lowerCamelCase = True ): if self.use_tiling and (x.shape[-1] > self.tile_sample_min_size or x.shape[-2] > self.tile_sample_min_size): return self.tiled_encode(lowerCamelCase , return_dict=lowerCamelCase ) if self.use_slicing and x.shape[0] > 1: __a = [self.encoder(lowerCamelCase ) for x_slice in x.split(1 )] __a = torch.cat(lowerCamelCase ) else: __a = self.encoder(lowerCamelCase ) __a = self.quant_conv(lowerCamelCase ) __a = DiagonalGaussianDistribution(lowerCamelCase ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = True ): if self.use_tiling and (z.shape[-1] > self.tile_latent_min_size or z.shape[-2] > self.tile_latent_min_size): return self.tiled_decode(lowerCamelCase , return_dict=lowerCamelCase ) __a = self.post_quant_conv(lowerCamelCase ) __a = self.decoder(lowerCamelCase ) if not return_dict: return (dec,) return DecoderOutput(sample=lowerCamelCase ) @apply_forward_hook def a__ ( self , lowerCamelCase , lowerCamelCase = True ): if self.use_slicing and z.shape[0] > 1: __a = [self._decode(lowerCamelCase ).sample for z_slice in z.split(1 )] __a = torch.cat(lowerCamelCase ) else: __a = self._decode(lowerCamelCase ).sample if not return_dict: return (decoded,) return DecoderOutput(sample=lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = min(a.shape[2] , b.shape[2] , lowerCamelCase ) for y in range(lowerCamelCase ): __a = a[:, :, -blend_extent + y, :] * (1 - y / blend_extent) + b[:, :, y, :] * (y / blend_extent) return b def a__ ( self , lowerCamelCase , lowerCamelCase , lowerCamelCase ): __a = min(a.shape[3] , b.shape[3] , lowerCamelCase ) for x in range(lowerCamelCase ): __a = a[:, :, :, -blend_extent + x] * (1 - x / blend_extent) + b[:, :, :, x] * (x / blend_extent) return b def a__ ( self , lowerCamelCase , lowerCamelCase = True ): __a = int(self.tile_sample_min_size * (1 - self.tile_overlap_factor) ) __a = int(self.tile_latent_min_size * self.tile_overlap_factor ) __a = self.tile_latent_min_size - blend_extent # Split the image into 512x512 tiles and encode them separately. __a = [] for i in range(0 , x.shape[2] , lowerCamelCase ): __a = [] for j in range(0 , x.shape[3] , lowerCamelCase ): __a = x[:, :, i : i + self.tile_sample_min_size, j : j + self.tile_sample_min_size] __a = self.encoder(lowerCamelCase ) __a = self.quant_conv(lowerCamelCase ) row.append(lowerCamelCase ) rows.append(lowerCamelCase ) __a = [] for i, row in enumerate(lowerCamelCase ): __a = [] for j, tile in enumerate(lowerCamelCase ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: __a = self.blend_v(rows[i - 1][j] , lowerCamelCase , lowerCamelCase ) if j > 0: __a = self.blend_h(row[j - 1] , lowerCamelCase , lowerCamelCase ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(lowerCamelCase , dim=3 ) ) __a = torch.cat(lowerCamelCase , dim=2 ) __a = DiagonalGaussianDistribution(lowerCamelCase ) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = True ): __a = int(self.tile_latent_min_size * (1 - self.tile_overlap_factor) ) __a = int(self.tile_sample_min_size * self.tile_overlap_factor ) __a = self.tile_sample_min_size - blend_extent # Split z into overlapping 64x64 tiles and decode them separately. # The tiles have an overlap to avoid seams between tiles. __a = [] for i in range(0 , z.shape[2] , lowerCamelCase ): __a = [] for j in range(0 , z.shape[3] , lowerCamelCase ): __a = z[:, :, i : i + self.tile_latent_min_size, j : j + self.tile_latent_min_size] __a = self.post_quant_conv(lowerCamelCase ) __a = self.decoder(lowerCamelCase ) row.append(lowerCamelCase ) rows.append(lowerCamelCase ) __a = [] for i, row in enumerate(lowerCamelCase ): __a = [] for j, tile in enumerate(lowerCamelCase ): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: __a = self.blend_v(rows[i - 1][j] , lowerCamelCase , lowerCamelCase ) if j > 0: __a = self.blend_h(row[j - 1] , lowerCamelCase , lowerCamelCase ) result_row.append(tile[:, :, :row_limit, :row_limit] ) result_rows.append(torch.cat(lowerCamelCase , dim=3 ) ) __a = torch.cat(lowerCamelCase , dim=2 ) if not return_dict: return (dec,) return DecoderOutput(sample=lowerCamelCase ) def a__ ( self , lowerCamelCase , lowerCamelCase = False , lowerCamelCase = True , lowerCamelCase = None , ): __a = sample __a = self.encode(lowerCamelCase ).latent_dist if sample_posterior: __a = posterior.sample(generator=lowerCamelCase ) else: __a = posterior.mode() __a = self.decode(lowerCamelCase ).sample if not return_dict: return (dec,) return DecoderOutput(sample=lowerCamelCase )
261
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCAmelCase_ = { 'configuration_xlm': ['XLM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMConfig', 'XLMOnnxConfig'], 'tokenization_xlm': ['XLMTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ 'XLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMForMultipleChoice', 'XLMForQuestionAnswering', 'XLMForQuestionAnsweringSimple', 'XLMForSequenceClassification', 'XLMForTokenClassification', 'XLMModel', 'XLMPreTrainedModel', 'XLMWithLMHeadModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ 'TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLMForMultipleChoice', 'TFXLMForQuestionAnsweringSimple', 'TFXLMForSequenceClassification', 'TFXLMForTokenClassification', 'TFXLMMainLayer', 'TFXLMModel', 'TFXLMPreTrainedModel', 'TFXLMWithLMHeadModel', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
358
from string import ascii_uppercase lowerCAmelCase_ = {char: i for i, char in enumerate(ascii_uppercase)} lowerCAmelCase_ = dict(enumerate(ascii_uppercase)) def snake_case( __magic_name__ , __magic_name__ ) -> str: '''simple docstring''' lowercase : Optional[Any] = len(__magic_name__ ) lowercase : Any = 0 while True: if x == i: lowercase : Any = 0 if len(__magic_name__ ) == len(__magic_name__ ): break key += key[i] i += 1 return key def snake_case( __magic_name__ , __magic_name__ ) -> str: '''simple docstring''' lowercase : str = '''''' lowercase : Dict = 0 for letter in message: if letter == " ": cipher_text += " " else: lowercase : Dict = (dicta[letter] - dicta[key_new[i]]) % 26 i += 1 cipher_text += dicta[x] return cipher_text def snake_case( __magic_name__ , __magic_name__ ) -> str: '''simple docstring''' lowercase : Any = '''''' lowercase : str = 0 for letter in cipher_text: if letter == " ": or_txt += " " else: lowercase : Any = (dicta[letter] + dicta[key_new[i]] + 26) % 26 i += 1 or_txt += dicta[x] return or_txt def snake_case( ) -> None: '''simple docstring''' lowercase : Dict = '''THE GERMAN ATTACK''' lowercase : Dict = '''SECRET''' lowercase : Union[str, Any] = generate_key(__magic_name__ , __magic_name__ ) lowercase : List[str] = cipher_text(__magic_name__ , __magic_name__ ) print(F"""Encrypted Text = {s}""" ) print(F"""Original Text = {original_text(__magic_name__ , __magic_name__ )}""" ) if __name__ == "__main__": import doctest doctest.testmod() main()
116
0
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.activations import gelu_new, gelu_python, get_activation @require_torch class lowercase_ ( unittest.TestCase ): '''simple docstring''' def __lowerCAmelCase ( self : List[str] ) ->List[str]: """simple docstring""" a = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) a = get_activation('''gelu''' ) self.assertTrue(torch.allclose(gelu_python(__UpperCAmelCase ) , torch_builtin(__UpperCAmelCase ) ) ) self.assertFalse(torch.allclose(gelu_python(__UpperCAmelCase ) , gelu_new(__UpperCAmelCase ) ) ) def __lowerCAmelCase ( self : Optional[Any] ) ->List[Any]: """simple docstring""" a = torch.tensor([-100, -1, -0.1, 0, 0.1, 1.0, 100] ) a = get_activation('''gelu''' ) a = get_activation('''gelu_10''' ) a = torch_builtin(__UpperCAmelCase ) a = geluaa(__UpperCAmelCase ) a = torch.where(y_gelu_aa < 10.0 , 1 , 0 ) self.assertTrue(torch.max(__UpperCAmelCase ).item() == 10.0 ) self.assertTrue(torch.allclose(y_gelu * clipped_mask , y_gelu_aa * clipped_mask ) ) def __lowerCAmelCase ( self : List[Any] ) ->Optional[int]: """simple docstring""" get_activation('''gelu''' ) get_activation('''gelu_10''' ) get_activation('''gelu_fast''' ) get_activation('''gelu_new''' ) get_activation('''gelu_python''' ) get_activation('''gelu_pytorch_tanh''' ) get_activation('''linear''' ) get_activation('''mish''' ) get_activation('''quick_gelu''' ) get_activation('''relu''' ) get_activation('''sigmoid''' ) get_activation('''silu''' ) get_activation('''swish''' ) get_activation('''tanh''' ) with self.assertRaises(__UpperCAmelCase ): get_activation('''bogus''' ) with self.assertRaises(__UpperCAmelCase ): get_activation(__UpperCAmelCase ) def __lowerCAmelCase ( self : Any ) ->str: """simple docstring""" a = get_activation('''gelu''' ) a = 1 a = get_activation('''gelu''' ) self.assertEqual(acta.a , 1 ) with self.assertRaises(__UpperCAmelCase ): a = acta.a
0
from ..utils import DummyObject, requires_backends class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> str: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> str: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) def lowerCamelCase__ ( *__lowerCAmelCase : Union[str, Any] , **__lowerCAmelCase : Dict ): """simple docstring""" requires_backends(__lowerCAmelCase , ["torch"] ) def lowerCamelCase__ ( *__lowerCAmelCase : Optional[int] , **__lowerCAmelCase : int ): """simple docstring""" requires_backends(__lowerCAmelCase , ["torch"] ) def lowerCamelCase__ ( *__lowerCAmelCase : List[str] , **__lowerCAmelCase : int ): """simple docstring""" requires_backends(__lowerCAmelCase , ["torch"] ) def lowerCamelCase__ ( *__lowerCAmelCase : Union[str, Any] , **__lowerCAmelCase : Optional[Any] ): """simple docstring""" requires_backends(__lowerCAmelCase , ["torch"] ) def lowerCamelCase__ ( *__lowerCAmelCase : Dict , **__lowerCAmelCase : Any ): """simple docstring""" requires_backends(__lowerCAmelCase , ["torch"] ) def lowerCamelCase__ ( *__lowerCAmelCase : Optional[int] , **__lowerCAmelCase : Dict ): """simple docstring""" requires_backends(__lowerCAmelCase , ["torch"] ) def lowerCamelCase__ ( *__lowerCAmelCase : int , **__lowerCAmelCase : Any ): """simple docstring""" requires_backends(__lowerCAmelCase , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Union[str, Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Union[str, Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> str: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Union[str, Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Union[str, Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> str: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> str: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Union[str, Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Union[str, Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> int: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Any: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> Optional[int]: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Union[str, Any]: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Tuple: requires_backends(cls , ["torch"] ) class _lowerCAmelCase ( metaclass=__a ): _lowercase =['''torch'''] def __init__( self , *_UpperCamelCase , **_UpperCamelCase ) -> str: requires_backends(self , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> Dict: requires_backends(cls , ["torch"] ) @classmethod def __a ( cls , *_UpperCamelCase , **_UpperCamelCase ) -> List[str]: requires_backends(cls , ["torch"] )
231
0
import warnings from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding class _lowercase ( snake_case_ ): lowercase = ['image_processor', 'tokenizer'] lowercase = 'CLIPImageProcessor' lowercase = ('XLMRobertaTokenizer', 'XLMRobertaTokenizerFast') def __init__( self : str , snake_case : List[str]=None , snake_case : Dict=None , **snake_case : Union[str, Any] ) -> Dict: """simple docstring""" UpperCamelCase_ : Dict = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , snake_case , ) UpperCamelCase_ : Optional[int] = kwargs.pop('feature_extractor' ) UpperCamelCase_ : str = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(snake_case , snake_case ) def __call__( self : Optional[int] , snake_case : Dict=None , snake_case : Optional[Any]=None , snake_case : Union[str, Any]=None , **snake_case : List[str] ) -> Union[str, Any]: """simple docstring""" if text is None and images is None: raise ValueError('You have to specify either text or images. Both cannot be none.' ) if text is not None: UpperCamelCase_ : List[str] = self.tokenizer(snake_case , return_tensors=snake_case , **snake_case ) if images is not None: UpperCamelCase_ : Optional[Any] = self.image_processor(snake_case , return_tensors=snake_case , **snake_case ) if text is not None and images is not None: UpperCamelCase_ : Union[str, Any] = image_features.pixel_values return encoding elif text is not None: return encoding else: return BatchEncoding(data=dict(**snake_case ) , tensor_type=snake_case ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , *snake_case : Dict , **snake_case : int ) -> int: """simple docstring""" return self.tokenizer.batch_decode(*snake_case , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self : int , *snake_case : List[Any] , **snake_case : Union[str, Any] ) -> Optional[Any]: """simple docstring""" return self.tokenizer.decode(*snake_case , **snake_case ) @property def SCREAMING_SNAKE_CASE__ ( self : Dict ) -> Tuple: """simple docstring""" UpperCamelCase_ : int = self.tokenizer.model_input_names UpperCamelCase_ : Optional[int] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
50
from typing import Any class _lowercase : def __init__( self : Optional[Any] , snake_case : Any ) -> Any: """simple docstring""" UpperCamelCase_ : Union[str, Any] = data UpperCamelCase_ : Any = None def __repr__( self : int ) -> str: """simple docstring""" return f"Node({self.data})" class _lowercase : def __init__( self : str ) -> int: """simple docstring""" UpperCamelCase_ : int = None def __iter__( self : Optional[Any] ) -> Any: """simple docstring""" UpperCamelCase_ : Tuple = self.head while node: yield node.data UpperCamelCase_ : Dict = node.next def __len__( self : int ) -> int: """simple docstring""" return sum(1 for _ in self ) def __repr__( self : List[Any] ) -> str: """simple docstring""" return "->".join([str(snake_case ) for item in self] ) def __getitem__( self : Union[str, Any] , snake_case : int ) -> Any: """simple docstring""" if not 0 <= index < len(self ): raise ValueError('list index out of range.' ) for i, node in enumerate(self ): if i == index: return node return None def __setitem__( self : Any , snake_case : int , snake_case : Any ) -> None: """simple docstring""" if not 0 <= index < len(self ): raise ValueError('list index out of range.' ) UpperCamelCase_ : int = self.head for _ in range(snake_case ): UpperCamelCase_ : Union[str, Any] = current.next UpperCamelCase_ : Any = data def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] , snake_case : Any ) -> None: """simple docstring""" self.insert_nth(len(self ) , snake_case ) def SCREAMING_SNAKE_CASE__ ( self : Dict , snake_case : Any ) -> None: """simple docstring""" self.insert_nth(0 , snake_case ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , snake_case : int , snake_case : Any ) -> None: """simple docstring""" if not 0 <= index <= len(self ): raise IndexError('list index out of range' ) UpperCamelCase_ : Union[str, Any] = Node(snake_case ) if self.head is None: UpperCamelCase_ : Union[str, Any] = new_node elif index == 0: UpperCamelCase_ : int = self.head # link new_node to head UpperCamelCase_ : List[str] = new_node else: UpperCamelCase_ : List[str] = self.head for _ in range(index - 1 ): UpperCamelCase_ : Union[str, Any] = temp.next UpperCamelCase_ : Dict = temp.next UpperCamelCase_ : Optional[Any] = new_node def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> None: # print every node data """simple docstring""" print(self ) def SCREAMING_SNAKE_CASE__ ( self : int ) -> Any: """simple docstring""" return self.delete_nth(0 ) def SCREAMING_SNAKE_CASE__ ( self : int ) -> Any: # delete from tail """simple docstring""" return self.delete_nth(len(self ) - 1 ) def SCREAMING_SNAKE_CASE__ ( self : Any , snake_case : int = 0 ) -> Any: """simple docstring""" if not 0 <= index <= len(self ) - 1: # test if index is valid raise IndexError('List index out of range.' ) UpperCamelCase_ : str = self.head # default first node if index == 0: UpperCamelCase_ : List[Any] = self.head.next else: UpperCamelCase_ : Dict = self.head for _ in range(index - 1 ): UpperCamelCase_ : Tuple = temp.next UpperCamelCase_ : List[Any] = temp.next UpperCamelCase_ : Dict = temp.next.next return delete_node.data def SCREAMING_SNAKE_CASE__ ( self : Dict ) -> bool: """simple docstring""" return self.head is None def SCREAMING_SNAKE_CASE__ ( self : str ) -> None: """simple docstring""" UpperCamelCase_ : str = None UpperCamelCase_ : int = self.head while current: # Store the current node's next node. UpperCamelCase_ : Tuple = current.next # Make the current node's next point backwards UpperCamelCase_ : Tuple = prev # Make the previous node be the current node UpperCamelCase_ : List[Any] = current # Make the current node the next node (to progress iteration) UpperCamelCase_ : Dict = next_node # Return prev in order to put the head at the end UpperCamelCase_ : Union[str, Any] = prev def __lowercase ( ): UpperCamelCase_ : Union[str, Any] = LinkedList() assert linked_list.is_empty() is True assert str(lowerCamelCase ) == "" try: linked_list.delete_head() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. try: linked_list.delete_tail() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. for i in range(10 ): assert len(lowerCamelCase ) == i linked_list.insert_nth(lowerCamelCase , i + 1 ) assert str(lowerCamelCase ) == "->".join(str(lowerCamelCase ) for i in range(1 , 11 ) ) linked_list.insert_head(0 ) linked_list.insert_tail(11 ) assert str(lowerCamelCase ) == "->".join(str(lowerCamelCase ) for i in range(0 , 12 ) ) assert linked_list.delete_head() == 0 assert linked_list.delete_nth(9 ) == 10 assert linked_list.delete_tail() == 11 assert len(lowerCamelCase ) == 9 assert str(lowerCamelCase ) == "->".join(str(lowerCamelCase ) for i in range(1 , 10 ) ) assert all(linked_list[i] == i + 1 for i in range(0 , 9 ) ) is True for i in range(0 , 9 ): UpperCamelCase_ : Optional[int] = -i assert all(linked_list[i] == -i for i in range(0 , 9 ) ) is True linked_list.reverse() assert str(lowerCamelCase ) == "->".join(str(lowerCamelCase ) for i in range(-8 , 1 ) ) def __lowercase ( ): UpperCamelCase_ : List[str] = [ -9, 100, Node(77345112 ), 'dlrow olleH', 7, 5555, 0, -1_9_2.5_5_5_5_5, 'Hello, world!', 7_7.9, Node(10 ), None, None, 1_2.2_0, ] UpperCamelCase_ : Optional[int] = LinkedList() for i in test_input: linked_list.insert_tail(lowerCamelCase ) # Check if it's empty or not assert linked_list.is_empty() is False assert ( str(lowerCamelCase ) == "-9->100->Node(77345112)->dlrow olleH->7->5555->0->" "-192.55555->Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the head UpperCamelCase_ : List[Any] = linked_list.delete_head() assert result == -9 assert ( str(lowerCamelCase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the tail UpperCamelCase_ : Tuple = linked_list.delete_tail() assert result == 1_2.2 assert ( str(lowerCamelCase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None" ) # Delete a node in specific location in linked list UpperCamelCase_ : Optional[Any] = linked_list.delete_nth(10 ) assert result is None assert ( str(lowerCamelCase ) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None" ) # Add a Node instance to its head linked_list.insert_head(Node('Hello again, world!' ) ) assert ( str(lowerCamelCase ) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None" ) # Add None to its tail linked_list.insert_tail(lowerCamelCase ) assert ( str(lowerCamelCase ) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None->None" ) # Reverse the linked list linked_list.reverse() assert ( str(lowerCamelCase ) == "None->None->Node(10)->77.9->Hello, world!->-192.55555->0->5555->" "7->dlrow olleH->Node(77345112)->100->Node(Hello again, world!)" ) def __lowercase ( ): from doctest import testmod testmod() UpperCamelCase_ : Dict = LinkedList() linked_list.insert_head(input('Inserting 1st at head ' ).strip() ) linked_list.insert_head(input('Inserting 2nd at head ' ).strip() ) print('\nPrint list:' ) linked_list.print_list() linked_list.insert_tail(input('\nInserting 1st at tail ' ).strip() ) linked_list.insert_tail(input('Inserting 2nd at tail ' ).strip() ) print('\nPrint list:' ) linked_list.print_list() print('\nDelete head' ) linked_list.delete_head() print('Delete tail' ) linked_list.delete_tail() print('\nPrint list:' ) linked_list.print_list() print('\nReverse linked list' ) linked_list.reverse() print('\nPrint list:' ) linked_list.print_list() print('\nString representation of linked list:' ) print(lowerCamelCase ) print('\nReading/changing Node data using indexing:' ) print(F"Element at Position 1: {linked_list[1]}" ) UpperCamelCase_ : Optional[Any] = input('Enter New Value: ' ).strip() print('New list:' ) print(lowerCamelCase ) print(F"length of linked_list is : {len(lowerCamelCase )}" ) if __name__ == "__main__": main()
50
1
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_VISION_2_SEQ_MAPPING if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_VISION_2_SEQ_MAPPING lowerCamelCase_ = logging.get_logger(__name__) @add_end_docstrings(__A ) class UpperCamelCase_ (__A ): def __init__( self : List[Any] , *lowerCAmelCase_ : List[str] , **lowerCAmelCase_ : Dict ) -> Optional[Any]: super().__init__(*lowerCAmelCase_ , **lowerCAmelCase_ ) requires_backends(self , "vision" ) self.check_model_type( TF_MODEL_FOR_VISION_2_SEQ_MAPPING if self.framework == "tf" else MODEL_FOR_VISION_2_SEQ_MAPPING ) def _SCREAMING_SNAKE_CASE ( self : Tuple , lowerCAmelCase_ : List[str]=None , lowerCAmelCase_ : int=None , lowerCAmelCase_ : str=None ) -> str: UpperCAmelCase_ : Any = {} UpperCAmelCase_ : List[Any] = {} if prompt is not None: UpperCAmelCase_ : List[Any] = prompt if generate_kwargs is not None: UpperCAmelCase_ : str = generate_kwargs if max_new_tokens is not None: if "generate_kwargs" not in forward_kwargs: UpperCAmelCase_ : List[Any] = {} if "max_new_tokens" in forward_kwargs["generate_kwargs"]: raise ValueError( "'max_new_tokens' is defined twice, once in 'generate_kwargs' and once as a direct parameter," " please use only one" ) UpperCAmelCase_ : Union[str, Any] = max_new_tokens return preprocess_params, forward_kwargs, {} def __call__( self : Tuple , lowerCAmelCase_ : Union[str, List[str], "Image.Image", List["Image.Image"]] , **lowerCAmelCase_ : int ) -> List[Any]: return super().__call__(lowerCAmelCase_ , **lowerCAmelCase_ ) def _SCREAMING_SNAKE_CASE ( self : Dict , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[str]=None ) -> Union[str, Any]: UpperCAmelCase_ : Union[str, Any] = load_image(lowerCAmelCase_ ) if prompt is not None: if not isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): raise ValueError( f"""Received an invalid text input, got - {type(lowerCAmelCase_ )} - but expected a single string. """ "Note also that one single text can be provided for conditional image to text generation." ) UpperCAmelCase_ : List[str] = self.model.config.model_type if model_type == "git": UpperCAmelCase_ : Optional[int] = self.image_processor(images=lowerCAmelCase_ , return_tensors=self.framework ) UpperCAmelCase_ : Any = self.tokenizer(text=lowerCAmelCase_ , add_special_tokens=lowerCAmelCase_ ).input_ids UpperCAmelCase_ : Optional[int] = [self.tokenizer.cls_token_id] + input_ids UpperCAmelCase_ : Dict = torch.tensor(lowerCAmelCase_ ).unsqueeze(0 ) model_inputs.update({"input_ids": input_ids} ) elif model_type == "pix2struct": UpperCAmelCase_ : Union[str, Any] = self.image_processor(images=lowerCAmelCase_ , header_text=lowerCAmelCase_ , return_tensors=self.framework ) elif model_type != "vision-encoder-decoder": # vision-encoder-decoder does not support conditional generation UpperCAmelCase_ : Optional[int] = self.image_processor(images=lowerCAmelCase_ , return_tensors=self.framework ) UpperCAmelCase_ : str = self.tokenizer(lowerCAmelCase_ , return_tensors=self.framework ) model_inputs.update(lowerCAmelCase_ ) else: raise ValueError(f"""Model type {model_type} does not support conditional text generation""" ) else: UpperCAmelCase_ : Dict = self.image_processor(images=lowerCAmelCase_ , return_tensors=self.framework ) if self.model.config.model_type == "git" and prompt is None: UpperCAmelCase_ : Optional[Any] = None return model_inputs def _SCREAMING_SNAKE_CASE ( self : Any , lowerCAmelCase_ : Union[str, Any] , lowerCAmelCase_ : List[str]=None ) -> Dict: # Git model sets `model_inputs["input_ids"] = None` in `preprocess` (when `prompt=None`). In batch model, the # pipeline will group them into a list of `None`, which fail `_forward`. Avoid this by checking it first. if ( "input_ids" in model_inputs and isinstance(model_inputs["input_ids"] , lowerCAmelCase_ ) and all(x is None for x in model_inputs["input_ids"] ) ): UpperCAmelCase_ : str = None if generate_kwargs is None: UpperCAmelCase_ : Any = {} # FIXME: We need to pop here due to a difference in how `generation.py` and `generation.tf_utils.py` # parse inputs. In the Tensorflow version, `generate` raises an error if we don't use `input_ids` whereas # the PyTorch version matches it with `self.model.main_input_name` or `self.model.encoder.main_input_name` # in the `_prepare_model_inputs` method. UpperCAmelCase_ : Optional[Any] = model_inputs.pop(self.model.main_input_name ) UpperCAmelCase_ : List[Any] = self.model.generate(lowerCAmelCase_ , **lowerCAmelCase_ , **lowerCAmelCase_ ) return model_outputs def _SCREAMING_SNAKE_CASE ( self : Tuple , lowerCAmelCase_ : List[Any] ) -> int: UpperCAmelCase_ : List[Any] = [] for output_ids in model_outputs: UpperCAmelCase_ : int = { "generated_text": self.tokenizer.decode( lowerCAmelCase_ , skip_special_tokens=lowerCAmelCase_ , ) } records.append(lowerCAmelCase_ ) return records
268
"""simple docstring""" def snake_case ( A__ ,A__ ,A__ ,A__ ,A__ ): if index == number_of_items: return 0 UpperCAmelCase_ : str = 0 UpperCAmelCase_ : Dict = 0 UpperCAmelCase_ : Tuple = knapsack(A__ ,A__ ,A__ ,A__ ,index + 1 ) if weights[index] <= max_weight: UpperCAmelCase_ : Union[str, Any] = values[index] + knapsack( A__ ,A__ ,A__ ,max_weight - weights[index] ,index + 1 ) return max(A__ ,A__ ) if __name__ == "__main__": import doctest doctest.testmod()
268
1
import inspect import unittest from transformers import ConvNextVaConfig from transformers.models.auto import get_values from transformers.models.auto.modeling_auto import MODEL_FOR_BACKBONE_MAPPING_NAMES, MODEL_MAPPING_NAMES from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextVaBackbone, ConvNextVaForImageClassification, ConvNextVaModel from transformers.models.convnextva.modeling_convnextva import CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _snake_case : def __init__( self: Tuple , __lowerCamelCase: Optional[int] , __lowerCamelCase: Optional[Any]=13 , __lowerCamelCase: Optional[int]=32 , __lowerCamelCase: List[str]=3 , __lowerCamelCase: Dict=4 , __lowerCamelCase: Optional[Any]=[10, 20, 30, 40] , __lowerCamelCase: int=[2, 2, 3, 2] , __lowerCamelCase: Union[str, Any]=True , __lowerCamelCase: Union[str, Any]=True , __lowerCamelCase: Tuple=37 , __lowerCamelCase: Tuple="gelu" , __lowerCamelCase: List[Any]=10 , __lowerCamelCase: Optional[int]=0.02 , __lowerCamelCase: Optional[Any]=["stage2", "stage3", "stage4"] , __lowerCamelCase: Optional[int]=[2, 3, 4] , __lowerCamelCase: int=None , ) -> List[str]: __UpperCAmelCase : Union[str, Any] = parent __UpperCAmelCase : List[str] = batch_size __UpperCAmelCase : Optional[int] = image_size __UpperCAmelCase : List[str] = num_channels __UpperCAmelCase : Union[str, Any] = num_stages __UpperCAmelCase : List[str] = hidden_sizes __UpperCAmelCase : Any = depths __UpperCAmelCase : Optional[int] = is_training __UpperCAmelCase : List[Any] = use_labels __UpperCAmelCase : Optional[int] = intermediate_size __UpperCAmelCase : Optional[Any] = hidden_act __UpperCAmelCase : Union[str, Any] = num_labels __UpperCAmelCase : Any = initializer_range __UpperCAmelCase : List[str] = out_features __UpperCAmelCase : Tuple = out_indices __UpperCAmelCase : List[Any] = scope def _lowerCamelCase ( self: List[Any] ) -> Optional[int]: __UpperCAmelCase : Dict = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __UpperCAmelCase : List[str] = None if self.use_labels: __UpperCAmelCase : List[Any] = ids_tensor([self.batch_size] , self.num_labels ) __UpperCAmelCase : Optional[Any] = self.get_config() return config, pixel_values, labels def _lowerCamelCase ( self: Tuple ) -> List[Any]: return ConvNextVaConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=__lowerCamelCase , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def _lowerCamelCase ( self: List[Any] , __lowerCamelCase: int , __lowerCamelCase: int , __lowerCamelCase: Optional[int] ) -> Union[str, Any]: __UpperCAmelCase : Optional[Any] = ConvNextVaModel(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : List[str] = model(__lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _lowerCamelCase ( self: Optional[Any] , __lowerCamelCase: Optional[Any] , __lowerCamelCase: Any , __lowerCamelCase: Tuple ) -> Tuple: __UpperCAmelCase : Union[str, Any] = ConvNextVaForImageClassification(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : Optional[int] = model(__lowerCamelCase , labels=__lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _lowerCamelCase ( self: int , __lowerCamelCase: Any , __lowerCamelCase: Optional[int] , __lowerCamelCase: Optional[Any] ) -> Optional[int]: __UpperCAmelCase : List[str] = ConvNextVaBackbone(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : Any = model(__lowerCamelCase ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None __UpperCAmelCase : List[Any] = None __UpperCAmelCase : List[str] = ConvNextVaBackbone(config=__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() __UpperCAmelCase : Any = model(__lowerCamelCase ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def _lowerCamelCase ( self: int ) -> List[str]: __UpperCAmelCase : int = self.prepare_config_and_inputs() __UpperCAmelCase : Union[str, Any] = config_and_inputs __UpperCAmelCase : str = {"pixel_values": pixel_values} return config, inputs_dict def _lowerCamelCase ( self: List[Any] ) -> List[Any]: __UpperCAmelCase : Optional[int] = self.prepare_config_and_inputs() __UpperCAmelCase : Tuple = config_and_inputs __UpperCAmelCase : Dict = {"pixel_values": pixel_values, "labels": labels} return config, inputs_dict @require_torch class _snake_case ( _lowercase , _lowercase , unittest.TestCase ): lowerCamelCase__: Dict = ( ( ConvNextVaModel, ConvNextVaForImageClassification, ConvNextVaBackbone, ) if is_torch_available() else () ) lowerCamelCase__: str = ( {"feature-extraction": ConvNextVaModel, "image-classification": ConvNextVaForImageClassification} if is_torch_available() else {} ) lowerCamelCase__: Tuple = False lowerCamelCase__: int = False lowerCamelCase__: Dict = False lowerCamelCase__: int = False lowerCamelCase__: Any = False def _lowerCamelCase ( self: Union[str, Any] ) -> Union[str, Any]: __UpperCAmelCase : Union[str, Any] = ConvNextVaModelTester(self ) __UpperCAmelCase : str = ConfigTester(self , config_class=__lowerCamelCase , has_text_modality=__lowerCamelCase , hidden_size=37 ) def _lowerCamelCase ( self: Dict ) -> Tuple: self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _lowerCamelCase ( self: List[Any] ) -> int: return @unittest.skip(reason="ConvNextV2 does not use inputs_embeds" ) def _lowerCamelCase ( self: Optional[Any] ) -> Optional[int]: pass @unittest.skip(reason="ConvNextV2 does not support input and output embeddings" ) def _lowerCamelCase ( self: Any ) -> Any: pass @unittest.skip(reason="ConvNextV2 does not use feedforward chunking" ) def _lowerCamelCase ( self: str ) -> Optional[Any]: pass def _lowerCamelCase ( self: List[Any] ) -> int: if not self.model_tester.is_training: return for model_class in self.all_model_classes: __UpperCAmelCase : str = self.model_tester.prepare_config_and_inputs_with_labels() __UpperCAmelCase : Optional[Any] = True if model_class.__name__ in [ *get_values(__lowerCamelCase ), *get_values(__lowerCamelCase ), ]: continue __UpperCAmelCase : Optional[Any] = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.train() __UpperCAmelCase : Any = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) __UpperCAmelCase : Any = model(**__lowerCamelCase ).loss loss.backward() def _lowerCamelCase ( self: Optional[int] ) -> Dict: if not self.model_tester.is_training: return for model_class in self.all_model_classes: __UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_with_labels() __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = True if ( model_class.__name__ in [*get_values(__lowerCamelCase ), *get_values(__lowerCamelCase )] or not model_class.supports_gradient_checkpointing ): continue __UpperCAmelCase : int = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.gradient_checkpointing_enable() model.train() __UpperCAmelCase : List[Any] = self._prepare_for_class(__lowerCamelCase , __lowerCamelCase , return_labels=__lowerCamelCase ) __UpperCAmelCase : Any = model(**__lowerCamelCase ).loss loss.backward() def _lowerCamelCase ( self: List[str] ) -> Dict: __UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : str = model_class(__lowerCamelCase ) __UpperCAmelCase : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCAmelCase : List[Any] = [*signature.parameters.keys()] __UpperCAmelCase : int = ["pixel_values"] self.assertListEqual(arg_names[:1] , __lowerCamelCase ) def _lowerCamelCase ( self: str ) -> List[Any]: __UpperCAmelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCamelCase ) def _lowerCamelCase ( self: Union[str, Any] ) -> Dict: def check_hidden_states_output(__lowerCamelCase: Any , __lowerCamelCase: Tuple , __lowerCamelCase: str ): __UpperCAmelCase : Any = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() with torch.no_grad(): __UpperCAmelCase : Tuple = model(**self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) ) __UpperCAmelCase : List[Any] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __UpperCAmelCase : Optional[int] = self.model_tester.num_stages self.assertEqual(len(__lowerCamelCase ) , expected_num_stages + 1 ) # ConvNextV2's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase : Optional[int] = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCAmelCase : Any = True check_hidden_states_output(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) def _lowerCamelCase ( self: Optional[Any] ) -> Optional[int]: __UpperCAmelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__lowerCamelCase ) @slow def _lowerCamelCase ( self: Dict ) -> List[Any]: for model_name in CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCAmelCase : Optional[int] = ConvNextVaModel.from_pretrained(__lowerCamelCase ) self.assertIsNotNone(__lowerCamelCase ) def _UpperCamelCase ( ) -> List[Any]: __UpperCAmelCase : List[str] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class _snake_case ( unittest.TestCase ): @cached_property def _lowerCamelCase ( self: Optional[int] ) -> Dict: return AutoImageProcessor.from_pretrained("facebook/convnextv2-tiny-1k-224" ) if is_vision_available() else None @slow def _lowerCamelCase ( self: List[Any] ) -> Tuple: __UpperCAmelCase : List[Any] = ConvNextVaForImageClassification.from_pretrained("facebook/convnextv2-tiny-1k-224" ).to(__lowerCamelCase ) __UpperCAmelCase : List[str] = self.default_image_processor __UpperCAmelCase : Optional[Any] = prepare_img() __UpperCAmelCase : int = preprocessor(images=__lowerCamelCase , return_tensors="pt" ).to(__lowerCamelCase ) # forward pass with torch.no_grad(): __UpperCAmelCase : str = model(**__lowerCamelCase ) # verify the logits __UpperCAmelCase : Dict = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , __lowerCamelCase ) __UpperCAmelCase : str = torch.tensor([0.99_96, 0.19_66, -0.43_86] ).to(__lowerCamelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCamelCase , atol=1e-4 ) )
368
from typing import Optional, Tuple import jax import jax.numpy as jnp from flax import linen as nn from flax.core.frozen_dict import FrozenDict from transformers import CLIPConfig, FlaxPreTrainedModel from transformers.models.clip.modeling_flax_clip import FlaxCLIPVisionModule def _UpperCamelCase ( snake_case__, snake_case__, snake_case__=1e-1_2 ) -> str: __UpperCAmelCase : Any = jnp.divide(emb_a.T, jnp.clip(jnp.linalg.norm(snake_case__, axis=1 ), a_min=snake_case__ ) ).T __UpperCAmelCase : int = jnp.divide(emb_a.T, jnp.clip(jnp.linalg.norm(snake_case__, axis=1 ), a_min=snake_case__ ) ).T return jnp.matmul(snake_case__, norm_emb_a.T ) class _snake_case ( nn.Module ): lowerCamelCase__: CLIPConfig lowerCamelCase__: jnp.dtype = jnp.floataa def _lowerCamelCase ( self: Any ) -> Tuple: __UpperCAmelCase : List[str] = FlaxCLIPVisionModule(self.config.vision_config ) __UpperCAmelCase : Any = nn.Dense(self.config.projection_dim , use_bias=__lowerCamelCase , dtype=self.dtype ) __UpperCAmelCase : int = self.param("concept_embeds" , jax.nn.initializers.ones , (17, self.config.projection_dim) ) __UpperCAmelCase : int = self.param( "special_care_embeds" , jax.nn.initializers.ones , (3, self.config.projection_dim) ) __UpperCAmelCase : Tuple = self.param("concept_embeds_weights" , jax.nn.initializers.ones , (17,) ) __UpperCAmelCase : str = self.param("special_care_embeds_weights" , jax.nn.initializers.ones , (3,) ) def __call__( self: List[Any] , __lowerCamelCase: Dict ) -> Dict: __UpperCAmelCase : Optional[int] = self.vision_model(__lowerCamelCase )[1] __UpperCAmelCase : List[str] = self.visual_projection(__lowerCamelCase ) __UpperCAmelCase : Optional[int] = jax_cosine_distance(__lowerCamelCase , self.special_care_embeds ) __UpperCAmelCase : Optional[Any] = jax_cosine_distance(__lowerCamelCase , self.concept_embeds ) # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign image inputs __UpperCAmelCase : List[str] = 0.0 __UpperCAmelCase : Tuple = special_cos_dist - self.special_care_embeds_weights[None, :] + adjustment __UpperCAmelCase : List[str] = jnp.round(__lowerCamelCase , 3 ) __UpperCAmelCase : Any = jnp.any(special_scores > 0 , axis=1 , keepdims=__lowerCamelCase ) # Use a lower threshold if an image has any special care concept __UpperCAmelCase : List[Any] = is_special_care * 0.01 __UpperCAmelCase : Any = cos_dist - self.concept_embeds_weights[None, :] + special_adjustment __UpperCAmelCase : List[str] = jnp.round(__lowerCamelCase , 3 ) __UpperCAmelCase : Any = jnp.any(concept_scores > 0 , axis=1 ) return has_nsfw_concepts class _snake_case ( _lowercase ): lowerCamelCase__: int = CLIPConfig lowerCamelCase__: Tuple = "clip_input" lowerCamelCase__: str = FlaxStableDiffusionSafetyCheckerModule def __init__( self: Union[str, Any] , __lowerCamelCase: CLIPConfig , __lowerCamelCase: Optional[Tuple] = None , __lowerCamelCase: int = 0 , __lowerCamelCase: jnp.dtype = jnp.floataa , __lowerCamelCase: bool = True , **__lowerCamelCase: Optional[int] , ) -> int: if input_shape is None: __UpperCAmelCase : Dict = (1, 2_24, 2_24, 3) __UpperCAmelCase : Tuple = self.module_class(config=__lowerCamelCase , dtype=__lowerCamelCase , **__lowerCamelCase ) super().__init__(__lowerCamelCase , __lowerCamelCase , input_shape=__lowerCamelCase , seed=__lowerCamelCase , dtype=__lowerCamelCase , _do_init=_do_init ) def _lowerCamelCase ( self: Dict , __lowerCamelCase: jax.random.KeyArray , __lowerCamelCase: Tuple , __lowerCamelCase: FrozenDict = None ) -> FrozenDict: # init input tensor __UpperCAmelCase : Tuple = jax.random.normal(__lowerCamelCase , __lowerCamelCase ) __UpperCAmelCase , __UpperCAmelCase : Dict = jax.random.split(__lowerCamelCase ) __UpperCAmelCase : Optional[int] = {"params": params_rng, "dropout": dropout_rng} __UpperCAmelCase : str = self.module.init(__lowerCamelCase , __lowerCamelCase )["params"] return random_params def __call__( self: Union[str, Any] , __lowerCamelCase: Optional[Any] , __lowerCamelCase: dict = None , ) -> List[Any]: __UpperCAmelCase : int = jnp.transpose(__lowerCamelCase , (0, 2, 3, 1) ) return self.module.apply( {"params": params or self.params} , jnp.array(__lowerCamelCase , dtype=jnp.floataa ) , rngs={} , )
342
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging a_ : List[Any] = logging.get_logger(__name__) a_ : int = { """weiweishi/roc-bert-base-zh""": """https://huggingface.co/weiweishi/roc-bert-base-zh/resolve/main/config.json""", } class __UpperCamelCase ( lowerCamelCase__ ): lowercase : int ='roc_bert' def __init__( self, lowerCAmelCase=30_522, lowerCAmelCase=768, lowerCAmelCase=12, lowerCAmelCase=12, lowerCAmelCase=3_072, lowerCAmelCase="gelu", lowerCAmelCase=0.1, lowerCAmelCase=0.1, lowerCAmelCase=512, lowerCAmelCase=2, lowerCAmelCase=0.0_2, lowerCAmelCase=1e-12, lowerCAmelCase=True, lowerCAmelCase=0, lowerCAmelCase="absolute", lowerCAmelCase=None, lowerCAmelCase=True, lowerCAmelCase=True, lowerCAmelCase=768, lowerCAmelCase=910, lowerCAmelCase=512, lowerCAmelCase=24_858, lowerCAmelCase=True, **lowerCAmelCase, ): """simple docstring""" lowerCamelCase_ =vocab_size lowerCamelCase_ =max_position_embeddings lowerCamelCase_ =hidden_size lowerCamelCase_ =num_hidden_layers lowerCamelCase_ =num_attention_heads lowerCamelCase_ =intermediate_size lowerCamelCase_ =hidden_act lowerCamelCase_ =hidden_dropout_prob lowerCamelCase_ =attention_probs_dropout_prob lowerCamelCase_ =initializer_range lowerCamelCase_ =type_vocab_size lowerCamelCase_ =layer_norm_eps lowerCamelCase_ =use_cache lowerCamelCase_ =enable_pronunciation lowerCamelCase_ =enable_shape lowerCamelCase_ =pronunciation_embed_dim lowerCamelCase_ =pronunciation_vocab_size lowerCamelCase_ =shape_embed_dim lowerCamelCase_ =shape_vocab_size lowerCamelCase_ =concat_input lowerCamelCase_ =position_embedding_type lowerCamelCase_ =classifier_dropout super().__init__(pad_token_id=lowerCAmelCase, **lowerCAmelCase )
75
import argparse import os import re import packaging.version _A : Optional[int] = 'examples/' _A : str = { 'examples': (re.compile(r'^check_min_version\("[^"]+"\)\s*$', re.MULTILINE), 'check_min_version("VERSION")\n'), 'init': (re.compile(r'^__version__\s+=\s+"([^"]+)"\s*$', re.MULTILINE), '__version__ = "VERSION"\n'), 'setup': (re.compile(r'^(\s*)version\s*=\s*"[^"]+",', re.MULTILINE), r'\1version="VERSION",'), 'doc': (re.compile(r'^(\s*)release\s*=\s*"[^"]+"$', re.MULTILINE), 'release = "VERSION"\n'), } _A : Any = { 'init': 'src/diffusers/__init__.py', 'setup': 'setup.py', } _A : List[str] = 'README.md' def _a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> Any: """simple docstring""" with open(UpperCAmelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCamelCase__ : Tuple = f.read() lowerCamelCase__ , lowerCamelCase__ : Optional[int] = REPLACE_PATTERNS[pattern] lowerCamelCase__ : Union[str, Any] = replace.replace('''VERSION''' , UpperCAmelCase ) lowerCamelCase__ : Optional[Any] = re_pattern.sub(UpperCAmelCase , UpperCAmelCase ) with open(UpperCAmelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.write(UpperCAmelCase ) def _a ( UpperCAmelCase ) -> Dict: """simple docstring""" for folder, directories, fnames in os.walk(UpperCAmelCase ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove('''research_projects''' ) if "legacy" in directories: directories.remove('''legacy''' ) for fname in fnames: if fname.endswith('''.py''' ): update_version_in_file(os.path.join(UpperCAmelCase , UpperCAmelCase ) , UpperCAmelCase , pattern='''examples''' ) def _a ( UpperCAmelCase , UpperCAmelCase=False ) -> Dict: """simple docstring""" for pattern, fname in REPLACE_FILES.items(): update_version_in_file(UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) if not patch: update_version_in_examples(UpperCAmelCase ) def _a ( ) -> Union[str, Any]: """simple docstring""" lowerCamelCase__ : Any = '''🤗 Transformers currently provides the following architectures''' lowerCamelCase__ : Dict = '''1. Want to contribute a new model?''' with open(UpperCAmelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: lowerCamelCase__ : str = f.readlines() # Find the start of the list. lowerCamelCase__ : int = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 lowerCamelCase__ : str = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith('''1.''' ): lowerCamelCase__ : Any = lines[index].replace( '''https://huggingface.co/docs/diffusers/main/model_doc''' , '''https://huggingface.co/docs/diffusers/model_doc''' , ) index += 1 with open(UpperCAmelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(UpperCAmelCase ) def _a ( ) -> Any: """simple docstring""" with open(REPLACE_FILES['''init'''] , '''r''' ) as f: lowerCamelCase__ : List[str] = f.read() lowerCamelCase__ : Any = REPLACE_PATTERNS['''init'''][0].search(UpperCAmelCase ).groups()[0] return packaging.version.parse(UpperCAmelCase ) def _a ( UpperCAmelCase=False ) -> str: """simple docstring""" lowerCamelCase__ : List[Any] = get_version() if patch and default_version.is_devrelease: raise ValueError('''Can\'t create a patch version from the dev branch, checkout a released version!''' ) if default_version.is_devrelease: lowerCamelCase__ : Union[str, Any] = default_version.base_version elif patch: lowerCamelCase__ : str = f"{default_version.major}.{default_version.minor}.{default_version.micro + 1}" else: lowerCamelCase__ : Dict = f"{default_version.major}.{default_version.minor + 1}.0" # Now let's ask nicely if that's the right one. lowerCamelCase__ : str = input(f"Which version are you releasing? [{default_version}]" ) if len(UpperCAmelCase ) == 0: lowerCamelCase__ : int = default_version print(f"Updating version to {version}." ) global_version_update(UpperCAmelCase , patch=UpperCAmelCase ) def _a ( ) -> List[Any]: """simple docstring""" lowerCamelCase__ : List[str] = get_version() lowerCamelCase__ : Optional[int] = f"{current_version.major}.{current_version.minor + 1}.0.dev0" lowerCamelCase__ : Union[str, Any] = current_version.base_version # Check with the user we got that right. lowerCamelCase__ : Dict = input(f"Which version are we developing now? [{dev_version}]" ) if len(UpperCAmelCase ) == 0: lowerCamelCase__ : Tuple = dev_version print(f"Updating version to {version}." ) global_version_update(UpperCAmelCase ) # print("Cleaning main README, don't forget to run `make fix-copies`.") # clean_main_ref_in_model_list() if __name__ == "__main__": _A : Any = argparse.ArgumentParser() parser.add_argument('--post_release', action='store_true', help='Whether this is pre or post release.') parser.add_argument('--patch', action='store_true', help='Whether or not this is a patch release.') _A : Optional[Any] = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('Nothing to do after a patch :-)') else: post_release_work()
142
0
"""simple docstring""" import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel _lowercase : Dict = False _lowercase : str = True _lowercase : str = False if __name__ == "__main__": _lowercase : Optional[Any] = argparse.ArgumentParser() parser.add_argument( "--repo_path", default=None, type=str, required=True, help="The config json file corresponding to the architecture.", ) parser.add_argument("--dump_path", default=None, type=str, required=True, help="Path to the output model.") _lowercase : Optional[int] = parser.parse_args() _lowercase : int = { "image_size": "sample_size", "num_res_blocks": "layers_per_block", "block_channels": "block_out_channels", "down_blocks": "down_block_types", "up_blocks": "up_block_types", "downscale_freq_shift": "freq_shift", "resnet_num_groups": "norm_num_groups", "resnet_act_fn": "act_fn", "resnet_eps": "norm_eps", "num_head_channels": "attention_head_dim", } _lowercase : List[Any] = { "time_steps": "time_proj", "mid": "mid_block", "downsample_blocks": "down_blocks", "upsample_blocks": "up_blocks", } _lowercase : Optional[int] = "" if has_file(args.repo_path, "config.json") else "unet" with open(os.path.join(args.repo_path, subfolder, "config.json"), "r", encoding="utf-8") as reader: _lowercase : Tuple = reader.read() _lowercase : int = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, "config.json"): _lowercase : str = UNetaDModel(**config) else: _lowercase : Tuple = UNetaDConditionModel if "ldm-text2im-large-256" in args.repo_path else UNetaDModel _lowercase : str = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) _lowercase : Optional[Any] = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: _lowercase : str = config[key] del config[key] _lowercase : str = [k.replace("UNetRes", "") for k in config["down_block_types"]] _lowercase : Union[str, Any] = [k.replace("UNetRes", "") for k in config["up_block_types"]] if do_only_weights: _lowercase : str = torch.load(os.path.join(args.repo_path, subfolder, "diffusion_pytorch_model.bin")) _lowercase : List[Any] = {} for param_key, param_value in state_dict.items(): if param_key.endswith(".op.bias") or param_key.endswith(".op.weight"): continue _lowercase : str = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split(".")[0] == key: _lowercase : Union[str, Any] = param_value _lowercase : Tuple = True if not has_changed: _lowercase : Dict = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
272
"""simple docstring""" from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging _lowercase : List[str] = logging.get_logger(__name__) def snake_case__ ( __lowerCamelCase : Union[tf.Tensor, np.ndarray] ): """simple docstring""" if isinstance(__lowerCamelCase , np.ndarray ): return list(tensor.shape ) lowerCamelCase__ : List[Any] =tf.shape(__lowerCamelCase ) if tensor.shape == tf.TensorShape(__lowerCamelCase ): return dynamic lowerCamelCase__ : List[str] =tensor.shape.as_list() return [dynamic[i] if s is None else s for i, s in enumerate(__lowerCamelCase )] def snake_case__ ( __lowerCamelCase : tf.Tensor , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : Optional[str] = None ): """simple docstring""" return tf.nn.softmax(logits=logits + 1e-9 , axis=__lowerCamelCase , name=__lowerCamelCase ) def snake_case__ ( __lowerCamelCase : Optional[Any] , __lowerCamelCase : str , __lowerCamelCase : Optional[int] , __lowerCamelCase : Any=1e-5 , __lowerCamelCase : str=-1 ): """simple docstring""" # This is a very simplified functional layernorm, designed to duplicate # the functionality of PyTorch nn.functional.layer_norm when this is needed to port # models in Transformers. if weight.shape.rank != 1 or bias.shape.rank != 1 or not isinstance(__lowerCamelCase , __lowerCamelCase ): raise NotImplementedError('''Only 1D weight and bias tensors are supported for now, with only a single axis.''' ) # Get mean and variance on the axis to be normalized lowerCamelCase__ , lowerCamelCase__ : Optional[int] =tf.nn.moments(__lowerCamelCase , axes=[axis] , keepdims=__lowerCamelCase ) if axis != -1: # Reshape scale and weight to have the same rank as inputs, but with 1 dimensions # on every dimension except axis lowerCamelCase__ : Optional[Any] =[1] * inputs.shape.rank lowerCamelCase__ : Union[str, Any] =shape_list(__lowerCamelCase )[axis] lowerCamelCase__ : Optional[Any] =tf.reshape(__lowerCamelCase , __lowerCamelCase ) lowerCamelCase__ : Any =tf.reshape(__lowerCamelCase , __lowerCamelCase ) # Compute layer normalization using the batch_normalization # function. lowerCamelCase__ : List[str] =tf.nn.batch_normalization( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , offset=__lowerCamelCase , scale=__lowerCamelCase , variance_epsilon=__lowerCamelCase , ) return outputs def snake_case__ ( __lowerCamelCase : Any , __lowerCamelCase : List[Any]=0 , __lowerCamelCase : int=-1 ): """simple docstring""" # Replicates the behavior of torch.flatten in TF # If end_dim or start_dim is negative, count them from the end if end_dim < 0: end_dim += input.shape.rank if start_dim < 0: start_dim += input.shape.rank if start_dim == end_dim: return input lowerCamelCase__ : Optional[int] =tf.shape(__lowerCamelCase ) lowerCamelCase__ : Optional[int] =tf.math.reduce_prod(in_shape[start_dim : end_dim + 1] ) lowerCamelCase__ : List[Any] =tf.concat([in_shape[:start_dim], [flattened_dim], in_shape[end_dim + 1 :]] , axis=0 ) return tf.reshape(__lowerCamelCase , __lowerCamelCase ) def snake_case__ ( __lowerCamelCase : tf.Tensor ): """simple docstring""" if not isinstance(__lowerCamelCase , tf.Tensor ): lowerCamelCase__ : List[Any] =tf.convert_to_tensor(__lowerCamelCase ) # Catches stray NumPy inputs if encoder_attention_mask.shape.rank == 3: lowerCamelCase__ : int =encoder_attention_mask[:, None, :, :] if encoder_attention_mask.shape.rank == 2: lowerCamelCase__ : int =encoder_attention_mask[:, None, None, :] # T5 has a mask that can compare sequence ids, we can simulate this here with this transposition # Cf. https://github.com/tensorflow/mesh/blob/8d2465e9bc93129b913b5ccc6a59aa97abd96ec6/mesh_tensorflow # /transformer/transformer_layers.py#L270 # encoder_extended_attention_mask = (encoder_extended_attention_mask == # encoder_extended_attention_mask.transpose(-1, -2)) lowerCamelCase__ : str =( tf.cast(1 , encoder_attention_mask.dtype ) - encoder_extended_attention_mask ) * encoder_extended_attention_mask.dtype.min return encoder_extended_attention_mask def snake_case__ ( __lowerCamelCase : tf.Tensor , __lowerCamelCase : int , __lowerCamelCase : str = "input_ids" ): """simple docstring""" tf.debugging.assert_less( __lowerCamelCase , tf.cast(__lowerCamelCase , dtype=tensor.dtype ) , message=( f'''The maximum value of {tensor_name} ({tf.math.reduce_max(__lowerCamelCase )}) must be smaller than the embedding ''' f'''layer\'s input dimension ({embed_dim}). The likely cause is some problem at tokenization time.''' ) , ) def snake_case__ ( __lowerCamelCase : Tuple , __lowerCamelCase : List[str] , __lowerCamelCase : Dict ): """simple docstring""" lowerCamelCase__ : Any =64512 # Check that no item in `data` is larger than `HDF5_OBJECT_HEADER_LIMIT` # because in that case even chunking the array would not make the saving # possible. lowerCamelCase__ : Tuple =[x for x in data if len(__lowerCamelCase ) > HDF5_OBJECT_HEADER_LIMIT] # Expecting this to never be true. if bad_attributes: raise RuntimeError( '''The following attributes cannot be saved to HDF5 file because ''' f'''they are larger than {HDF5_OBJECT_HEADER_LIMIT} ''' f'''bytes: {bad_attributes}''' ) lowerCamelCase__ : Optional[Any] =np.asarray(__lowerCamelCase ) lowerCamelCase__ : str =1 lowerCamelCase__ : List[Any] =np.array_split(__lowerCamelCase , __lowerCamelCase ) # This will never loop forever thanks to the test above. while any(x.nbytes > HDF5_OBJECT_HEADER_LIMIT for x in chunked_data ): num_chunks += 1 lowerCamelCase__ : Union[str, Any] =np.array_split(__lowerCamelCase , __lowerCamelCase ) if num_chunks > 1: for chunk_id, chunk_data in enumerate(__lowerCamelCase ): lowerCamelCase__ : List[str] =chunk_data else: lowerCamelCase__ : Dict =data def snake_case__ ( __lowerCamelCase : Dict , __lowerCamelCase : Optional[Any] ): """simple docstring""" if name in group.attrs: lowerCamelCase__ : Optional[int] =[n.decode('''utf8''' ) if hasattr(__lowerCamelCase , '''decode''' ) else n for n in group.attrs[name]] else: lowerCamelCase__ : str =[] lowerCamelCase__ : str =0 while "%s%d" % (name, chunk_id) in group.attrs: data.extend( [n.decode('''utf8''' ) if hasattr(__lowerCamelCase , '''decode''' ) else n for n in group.attrs['''%s%d''' % (name, chunk_id)]] ) chunk_id += 1 return data def snake_case__ ( __lowerCamelCase : Dict ): """simple docstring""" def _expand_single_ad_tensor(__lowerCamelCase : List[Any] ): if isinstance(__lowerCamelCase , tf.Tensor ) and t.shape.rank == 1: return tf.expand_dims(__lowerCamelCase , axis=-1 ) return t return tf.nest.map_structure(_expand_single_ad_tensor , __lowerCamelCase )
272
1
'''simple docstring''' import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType a__ : List[Any] =logging.get_logger(__name__) class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple ="vision-encoder-decoder" SCREAMING_SNAKE_CASE_ : Tuple =True def __init__( self : List[str] , **__A : List[str] ): super().__init__(**__A ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( f'''A configuraton of type {self.model_type} cannot be instantiated because ''' f'''not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}''' ) __UpperCamelCase = kwargs.pop('encoder' ) __UpperCamelCase = encoder_config.pop('model_type' ) __UpperCamelCase = kwargs.pop('decoder' ) __UpperCamelCase = decoder_config.pop('model_type' ) __UpperCamelCase = AutoConfig.for_model(__A , **__A ) __UpperCamelCase = AutoConfig.for_model(__A , **__A ) __UpperCamelCase = True @classmethod def _lowerCamelCase ( cls : Any , __A : PretrainedConfig , __A : PretrainedConfig , **__A : Union[str, Any] ): logger.info('Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config' ) __UpperCamelCase = True __UpperCamelCase = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **__A ) def _lowerCamelCase ( self : str ): __UpperCamelCase = copy.deepcopy(self.__dict__ ) __UpperCamelCase = self.encoder.to_dict() __UpperCamelCase = self.decoder.to_dict() __UpperCamelCase = self.__class__.model_type return output class snake_case ( __lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Tuple =version.parse("1.11" ) @property def _lowerCamelCase ( self : str ): return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def _lowerCamelCase ( self : Dict ): return 1e-4 @property def _lowerCamelCase ( self : List[str] ): return OrderedDict({'last_hidden_state': {0: 'batch', 1: 'encoder_sequence'}} ) class snake_case ( __lowerCamelCase ): """simple docstring""" @property def _lowerCamelCase ( self : int ): __UpperCamelCase = OrderedDict() __UpperCamelCase = {0: 'batch', 1: 'past_decoder_sequence + sequence'} __UpperCamelCase = {0: 'batch', 1: 'past_decoder_sequence + sequence'} __UpperCamelCase = {0: 'batch', 1: 'encoder_sequence'} return common_inputs def _lowerCamelCase ( self : Union[str, Any] , __A : "PreTrainedTokenizerBase" , __A : int = -1 , __A : int = -1 , __A : bool = False , __A : Optional["TensorType"] = None , ): import torch __UpperCamelCase = OrderedDict() __UpperCamelCase = super().generate_dummy_inputs( __A , batch_size=__A , seq_length=__A , is_pair=__A , framework=__A ) __UpperCamelCase , __UpperCamelCase = dummy_input['input_ids'].shape __UpperCamelCase = (batch, encoder_sequence, self._config.encoder_hidden_size) __UpperCamelCase = dummy_input.pop('input_ids' ) __UpperCamelCase = dummy_input.pop('attention_mask' ) __UpperCamelCase = torch.zeros(__A ) return common_inputs class snake_case ( __lowerCamelCase ): """simple docstring""" @property def _lowerCamelCase ( self : str ): pass def _lowerCamelCase ( self : Optional[int] , __A : PretrainedConfig ): return VisionEncoderDecoderEncoderOnnxConfig(__A ) def _lowerCamelCase ( self : Union[str, Any] , __A : PretrainedConfig , __A : PretrainedConfig , __A : str = "default" ): __UpperCamelCase = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(__A , __A )
53
'''simple docstring''' import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def lowercase__ ( __lowercase : Any ) -> Union[str, Any]: """simple docstring""" __UpperCamelCase = [ 'encoder.version', 'decoder.version', 'model.encoder.version', 'model.decoder.version', '_float_tensor', 'decoder.output_projection.weight', ] for k in ignore_keys: state_dict.pop(__lowercase , __lowercase ) def lowercase__ ( __lowercase : Tuple ) -> int: """simple docstring""" __UpperCamelCase , __UpperCamelCase = emb.weight.shape __UpperCamelCase = nn.Linear(__lowercase , __lowercase , bias=__lowercase ) __UpperCamelCase = emb.weight.data return lin_layer def lowercase__ ( __lowercase : int , __lowercase : List[str]="facebook/mbart-large-en-ro" , __lowercase : str=False , __lowercase : List[Any]=False ) -> int: """simple docstring""" __UpperCamelCase = torch.load(__lowercase , map_location='cpu' )['model'] remove_ignore_keys_(__lowercase ) __UpperCamelCase = state_dict['encoder.embed_tokens.weight'].shape[0] __UpperCamelCase = MBartConfig.from_pretrained(__lowercase , vocab_size=__lowercase ) if mbart_aa and finetuned: __UpperCamelCase = 'relu' __UpperCamelCase = state_dict['decoder.embed_tokens.weight'] __UpperCamelCase = MBartForConditionalGeneration(__lowercase ) model.model.load_state_dict(__lowercase ) if finetuned: __UpperCamelCase = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": a__ : Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( '''fairseq_path''', type=str, help='''bart.large, bart.large.cnn or a path to a model.pt on local filesystem.''' ) parser.add_argument('''pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument( '''--hf_config''', default='''facebook/mbart-large-cc25''', type=str, help='''Which huggingface architecture to use: mbart-large''', ) parser.add_argument('''--mbart_50''', action='''store_true''', help='''whether the model is mMART-50 checkpoint''') parser.add_argument('''--finetuned''', action='''store_true''', help='''whether the model is a fine-tuned checkpoint''') a__ : Union[str, Any] =parser.parse_args() a__ : str =convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
53
1
'''simple docstring''' import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase : Tuple = logging.get_logger(__name__) def __lowerCamelCase ( A__ , A__=False ) -> Tuple: """simple docstring""" UpperCamelCase = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('cls_token', 'vit.embeddings.cls_token'), ('patch_embed.proj.weight', 'vit.embeddings.patch_embeddings.projection.weight'), ('patch_embed.proj.bias', 'vit.embeddings.patch_embeddings.projection.bias'), ('pos_embed', 'vit.embeddings.position_embeddings'), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('norm.weight', 'layernorm.weight'), ('norm.bias', 'layernorm.bias'), ('pre_logits.fc.weight', 'pooler.dense.weight'), ('pre_logits.fc.bias', 'pooler.dense.bias'), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" UpperCamelCase = [(pair[0], pair[1][4:]) if pair[1].startswith('vit' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('norm.weight', 'vit.layernorm.weight'), ('norm.bias', 'vit.layernorm.bias'), ('head.weight', 'classifier.weight'), ('head.bias', 'classifier.bias'), ] ) return rename_keys def __lowerCamelCase ( A__ , A__ , A__=False ) -> Union[str, Any]: """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: UpperCamelCase = '' else: UpperCamelCase = 'vit.' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCamelCase = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) UpperCamelCase = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCamelCase = in_proj_weight[ : config.hidden_size, : ] UpperCamelCase = in_proj_bias[: config.hidden_size] UpperCamelCase = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCamelCase = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCamelCase = in_proj_weight[ -config.hidden_size :, : ] UpperCamelCase = in_proj_bias[-config.hidden_size :] def __lowerCamelCase ( A__ ) -> Tuple: """simple docstring""" UpperCamelCase = ['head.weight', 'head.bias'] for k in ignore_keys: state_dict.pop(A__ , A__ ) def __lowerCamelCase ( A__ , A__ , A__ ) -> Union[str, Any]: """simple docstring""" UpperCamelCase = dct.pop(A__ ) UpperCamelCase = val def __lowerCamelCase ( ) -> int: """simple docstring""" UpperCamelCase = 'http://images.cocodataset.org/val2017/000000039769.jpg' UpperCamelCase = Image.open(requests.get(A__ , stream=A__ ).raw ) return im @torch.no_grad() def __lowerCamelCase ( A__ , A__ ) -> List[str]: """simple docstring""" UpperCamelCase = ViTConfig() UpperCamelCase = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": UpperCamelCase = True UpperCamelCase = int(vit_name[-12:-10] ) UpperCamelCase = int(vit_name[-9:-6] ) else: UpperCamelCase = 1_000 UpperCamelCase = 'huggingface/label-files' UpperCamelCase = 'imagenet-1k-id2label.json' UpperCamelCase = json.load(open(hf_hub_download(A__ , A__ , repo_type='dataset' ) , 'r' ) ) UpperCamelCase = {int(A__ ): v for k, v in idalabel.items()} UpperCamelCase = idalabel UpperCamelCase = {v: k for k, v in idalabel.items()} UpperCamelCase = int(vit_name[-6:-4] ) UpperCamelCase = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith('tiny' ): UpperCamelCase = 192 UpperCamelCase = 768 UpperCamelCase = 12 UpperCamelCase = 3 elif vit_name[9:].startswith('small' ): UpperCamelCase = 384 UpperCamelCase = 1_536 UpperCamelCase = 12 UpperCamelCase = 6 else: pass else: if vit_name[4:].startswith('small' ): UpperCamelCase = 768 UpperCamelCase = 2_304 UpperCamelCase = 8 UpperCamelCase = 8 elif vit_name[4:].startswith('base' ): pass elif vit_name[4:].startswith('large' ): UpperCamelCase = 1_024 UpperCamelCase = 4_096 UpperCamelCase = 24 UpperCamelCase = 16 elif vit_name[4:].startswith('huge' ): UpperCamelCase = 1_280 UpperCamelCase = 5_120 UpperCamelCase = 32 UpperCamelCase = 16 # load original model from timm UpperCamelCase = timm.create_model(A__ , pretrained=A__ ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCamelCase = timm_model.state_dict() if base_model: remove_classification_head_(A__ ) UpperCamelCase = create_rename_keys(A__ , A__ ) for src, dest in rename_keys: rename_key(A__ , A__ , A__ ) read_in_q_k_v(A__ , A__ , A__ ) # load HuggingFace model if vit_name[-5:] == "in21k": UpperCamelCase = ViTModel(A__ ).eval() else: UpperCamelCase = ViTForImageClassification(A__ ).eval() model.load_state_dict(A__ ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: UpperCamelCase = DeiTImageProcessor(size=config.image_size ) else: UpperCamelCase = ViTImageProcessor(size=config.image_size ) UpperCamelCase = image_processor(images=prepare_img() , return_tensors='pt' ) UpperCamelCase = encoding['pixel_values'] UpperCamelCase = model(A__ ) if base_model: UpperCamelCase = timm_model.forward_features(A__ ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(A__ , outputs.pooler_output , atol=1e-3 ) else: UpperCamelCase = timm_model(A__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(A__ , outputs.logits , atol=1e-3 ) Path(A__ ).mkdir(exist_ok=A__ ) print(F"""Saving model {vit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(A__ ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(A__ ) if __name__ == "__main__": _lowerCamelCase : str = argparse.ArgumentParser() # Required parameters parser.add_argument( "--vit_name", default="vit_base_patch16_224", type=str, help="Name of the ViT timm model you'd like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) _lowerCamelCase : Union[str, Any] = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
249
'''simple docstring''' from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class SCREAMING_SNAKE_CASE : """simple docstring""" _SCREAMING_SNAKE_CASE = XGLMConfig _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = """gelu""" def __init__( self : List[str] , UpperCamelCase__ : int , UpperCamelCase__ : List[Any]=1_4 , UpperCamelCase__ : int=7 , UpperCamelCase__ : List[Any]=True , UpperCamelCase__ : Optional[Any]=True , UpperCamelCase__ : str=True , UpperCamelCase__ : List[str]=9_9 , UpperCamelCase__ : str=3_2 , UpperCamelCase__ : Any=2 , UpperCamelCase__ : Optional[int]=4 , UpperCamelCase__ : str=3_7 , UpperCamelCase__ : Tuple="gelu" , UpperCamelCase__ : str=0.1 , UpperCamelCase__ : List[str]=0.1 , UpperCamelCase__ : Union[str, Any]=5_1_2 , UpperCamelCase__ : Optional[Any]=0.0_2 , ): """simple docstring""" UpperCamelCase = parent UpperCamelCase = batch_size UpperCamelCase = seq_length UpperCamelCase = is_training UpperCamelCase = use_input_mask UpperCamelCase = use_labels UpperCamelCase = vocab_size UpperCamelCase = d_model UpperCamelCase = num_hidden_layers UpperCamelCase = num_attention_heads UpperCamelCase = ffn_dim UpperCamelCase = activation_function UpperCamelCase = activation_dropout UpperCamelCase = attention_dropout UpperCamelCase = max_position_embeddings UpperCamelCase = initializer_range UpperCamelCase = None UpperCamelCase = 0 UpperCamelCase = 2 UpperCamelCase = 1 def A ( self : Union[str, Any] ): """simple docstring""" return XGLMConfig.from_pretrained('facebook/xglm-564M' ) def A ( self : List[Any] ): """simple docstring""" UpperCamelCase = tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) , clip_value_min=0 , clip_value_max=3 ) UpperCamelCase = None if self.use_input_mask: UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase = self.get_config() UpperCamelCase = floats_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, ) def A ( self : Union[str, Any] ): """simple docstring""" return XGLMConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , num_layers=self.num_hidden_layers , attention_heads=self.num_attention_heads , ffn_dim=self.ffn_dim , activation_function=self.activation_function , activation_dropout=self.activation_dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , use_cache=UpperCamelCase__ , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , return_dict=UpperCamelCase__ , ) def A ( self : Tuple ): """simple docstring""" UpperCamelCase = self.prepare_config_and_inputs() ( ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ( UpperCamelCase ) , ) = config_and_inputs UpperCamelCase = { 'input_ids': input_ids, 'head_mask': head_mask, } return config, inputs_dict @require_tf class SCREAMING_SNAKE_CASE ( _a , _a , unittest.TestCase ): """simple docstring""" _SCREAMING_SNAKE_CASE = (TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () _SCREAMING_SNAKE_CASE = (TFXGLMForCausalLM,) if is_tf_available() else () _SCREAMING_SNAKE_CASE = ( {"""feature-extraction""": TFXGLMModel, """text-generation""": TFXGLMForCausalLM} if is_tf_available() else {} ) _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = False _SCREAMING_SNAKE_CASE = False def A ( self : Dict ): """simple docstring""" UpperCamelCase = TFXGLMModelTester(self ) UpperCamelCase = ConfigTester(self , config_class=UpperCamelCase__ , n_embd=3_7 ) def A ( self : Optional[Any] ): """simple docstring""" self.config_tester.run_common_tests() @slow def A ( self : List[str] ): """simple docstring""" for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase = TFXGLMModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) @unittest.skip(reason='Currently, model embeddings are going to undergo a major refactor.' ) def A ( self : Dict ): """simple docstring""" super().test_resize_token_embeddings() @require_tf class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" @slow def A ( self : Optional[int] , UpperCamelCase__ : Tuple=True ): """simple docstring""" UpperCamelCase = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) UpperCamelCase = tf.convert_to_tensor([[2, 2_6_8, 9_8_6_5]] , dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off UpperCamelCase = [2, 2_6_8, 9_8_6_5, 6_7, 1_1, 1_9_8_8, 5_7_2_5_2, 9_8_6_5, 5, 9_8_4, 6_7, 1_9_8_8, 2_1_3_8_3_8, 1_6_5_8, 5_3, 7_0_4_4_6, 3_3, 6_6_5_7, 2_7_8, 1_5_8_1] # fmt: on UpperCamelCase = model.generate(UpperCamelCase__ , do_sample=UpperCamelCase__ , num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist() , UpperCamelCase__ ) @slow def A ( self : List[Any] ): """simple docstring""" UpperCamelCase = XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) UpperCamelCase = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) tf.random.set_seed(0 ) UpperCamelCase = tokenizer('Today is a nice day and' , return_tensors='tf' ) UpperCamelCase = tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(':/CPU:0' ): UpperCamelCase = model.generate(UpperCamelCase__ , do_sample=UpperCamelCase__ , seed=[7, 0] ) UpperCamelCase = tokenizer.decode(output_ids[0] , skip_special_tokens=UpperCamelCase__ ) UpperCamelCase = ( 'Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due' ) self.assertEqual(UpperCamelCase__ , UpperCamelCase__ ) @slow def A ( self : Dict ): """simple docstring""" UpperCamelCase = TFXGLMForCausalLM.from_pretrained('facebook/xglm-564M' ) UpperCamelCase = XGLMTokenizer.from_pretrained('facebook/xglm-564M' ) UpperCamelCase = 'left' # use different length sentences to test batching UpperCamelCase = [ 'This is an extremelly long sentence that only exists to test the ability of the model to cope with ' 'left-padding, such as in batched generation. The output for the sequence below should be the same ' 'regardless of whether left padding is applied or not. When', 'Hello, my dog is a little', ] UpperCamelCase = tokenizer(UpperCamelCase__ , return_tensors='tf' , padding=UpperCamelCase__ ) UpperCamelCase = inputs['input_ids'] UpperCamelCase = model.generate(input_ids=UpperCamelCase__ , attention_mask=inputs['attention_mask'] , max_new_tokens=1_2 ) UpperCamelCase = tokenizer(sentences[0] , return_tensors='tf' ).input_ids UpperCamelCase = model.generate(input_ids=UpperCamelCase__ , max_new_tokens=1_2 ) UpperCamelCase = tokenizer(sentences[1] , return_tensors='tf' ).input_ids UpperCamelCase = model.generate(input_ids=UpperCamelCase__ , max_new_tokens=1_2 ) UpperCamelCase = tokenizer.batch_decode(UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ ) UpperCamelCase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=UpperCamelCase__ ) UpperCamelCase = tokenizer.decode(output_padded[0] , skip_special_tokens=UpperCamelCase__ ) UpperCamelCase = [ 'This is an extremelly long sentence that only exists to test the ability of the model to cope with ' 'left-padding, such as in batched generation. The output for the sequence below should be the same ' 'regardless of whether left padding is applied or not. When left padding is applied, the sequence will be ' 'a single', 'Hello, my dog is a little bit of a shy one, but he is very friendly', ] self.assertListEqual(UpperCamelCase__ , UpperCamelCase__ ) self.assertListEqual(UpperCamelCase__ , [non_padded_sentence, padded_sentence] )
249
1
"""simple docstring""" import cva import numpy as np class __lowerCAmelCase : def __init__( self , __UpperCAmelCase , __UpperCAmelCase ): '''simple docstring''' if k in (0.0_4, 0.0_6): __UpperCamelCase = k __UpperCamelCase = window_size else: raise ValueError('invalid k value' ) def __str__( self ): '''simple docstring''' return str(self.k ) def UpperCAmelCase ( self , __UpperCAmelCase ): '''simple docstring''' __UpperCamelCase = cva.imread(__UpperCAmelCase , 0 ) __UpperCamelCase , __UpperCamelCase = img.shape __UpperCamelCase = [] __UpperCamelCase = img.copy() __UpperCamelCase = cva.cvtColor(__UpperCAmelCase , cva.COLOR_GRAY2RGB ) __UpperCamelCase , __UpperCamelCase = np.gradient(__UpperCAmelCase ) __UpperCamelCase = dx**2 __UpperCamelCase = dy**2 __UpperCamelCase = dx * dy __UpperCamelCase = 0.0_4 __UpperCamelCase = self.window_size // 2 for y in range(__UpperCAmelCase , h - offset ): for x in range(__UpperCAmelCase , w - offset ): __UpperCamelCase = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() __UpperCamelCase = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() __UpperCamelCase = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() __UpperCamelCase = (wxx * wyy) - (wxy**2) __UpperCamelCase = wxx + wyy __UpperCamelCase = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": UpperCamelCase : int = HarrisCorner(0.04, 3) UpperCamelCase , UpperCamelCase : Any = edge_detect.detect("path_to_image") cva.imwrite("detect.png", color_img)
316
"""simple docstring""" import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation UpperCamelCase : Any = logging.get_logger(__name__) UpperCamelCase : Any = {"vocab_file": "vocab.json", "merges_file": "merges.txt", "tokenizer_file": "tokenizer.json"} UpperCamelCase : Dict = { "vocab_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/vocab.json", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/vocab.json", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/vocab.json", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/vocab.json", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/vocab.json", }, "merges_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/merges.txt", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/merges.txt", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/merges.txt", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/merges.txt", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/merges.txt", }, "tokenizer_file": { "gpt2": "https://huggingface.co/gpt2/resolve/main/tokenizer.json", "gpt2-medium": "https://huggingface.co/gpt2-medium/resolve/main/tokenizer.json", "gpt2-large": "https://huggingface.co/gpt2-large/resolve/main/tokenizer.json", "gpt2-xl": "https://huggingface.co/gpt2-xl/resolve/main/tokenizer.json", "distilgpt2": "https://huggingface.co/distilgpt2/resolve/main/tokenizer.json", }, } UpperCamelCase : Dict = { "gpt2": 1_0_2_4, "gpt2-medium": 1_0_2_4, "gpt2-large": 1_0_2_4, "gpt2-xl": 1_0_2_4, "distilgpt2": 1_0_2_4, } class __lowerCAmelCase ( __SCREAMING_SNAKE_CASE ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = ["input_ids", "attention_mask"] lowercase = GPTaTokenizer def __init__( self , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase=None , __UpperCAmelCase="<|endoftext|>" , __UpperCAmelCase="<|endoftext|>" , __UpperCAmelCase="<|endoftext|>" , __UpperCAmelCase=False , **__UpperCAmelCase , ): '''simple docstring''' super().__init__( __UpperCAmelCase , __UpperCAmelCase , tokenizer_file=__UpperCAmelCase , unk_token=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , add_prefix_space=__UpperCAmelCase , **__UpperCAmelCase , ) __UpperCamelCase = kwargs.pop('add_bos_token' , __UpperCAmelCase ) __UpperCamelCase = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , __UpperCAmelCase ) != add_prefix_space: __UpperCamelCase = getattr(__UpperCAmelCase , pre_tok_state.pop('type' ) ) __UpperCamelCase = add_prefix_space __UpperCamelCase = pre_tok_class(**__UpperCAmelCase ) __UpperCamelCase = add_prefix_space def UpperCAmelCase ( self , *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' __UpperCamelCase = kwargs.get('is_split_into_words' , __UpperCAmelCase ) assert self.add_prefix_space or not is_split_into_words, ( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*__UpperCAmelCase , **__UpperCAmelCase ) def UpperCAmelCase ( self , *__UpperCAmelCase , **__UpperCAmelCase ): '''simple docstring''' __UpperCamelCase = kwargs.get('is_split_into_words' , __UpperCAmelCase ) assert self.add_prefix_space or not is_split_into_words, ( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs." ) return super()._encode_plus(*__UpperCAmelCase , **__UpperCAmelCase ) def UpperCAmelCase ( self , __UpperCAmelCase , __UpperCAmelCase = None ): '''simple docstring''' __UpperCamelCase = self._tokenizer.model.save(__UpperCAmelCase , name=__UpperCAmelCase ) return tuple(__UpperCAmelCase ) def UpperCAmelCase ( self , __UpperCAmelCase ): '''simple docstring''' __UpperCamelCase = [] for is_user, text in conversation.iter_texts(): input_ids.extend(self.encode(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) + [self.eos_token_id] ) if len(__UpperCAmelCase ) > self.model_max_length: __UpperCamelCase = input_ids[-self.model_max_length :] return input_ids
316
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __A : List[Any] = { 'configuration_xlm': ['XLM_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMConfig', 'XLMOnnxConfig'], 'tokenization_xlm': ['XLMTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : str = [ 'XLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'XLMForMultipleChoice', 'XLMForQuestionAnswering', 'XLMForQuestionAnsweringSimple', 'XLMForSequenceClassification', 'XLMForTokenClassification', 'XLMModel', 'XLMPreTrainedModel', 'XLMWithLMHeadModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A : str = [ 'TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFXLMForMultipleChoice', 'TFXLMForQuestionAnsweringSimple', 'TFXLMForSequenceClassification', 'TFXLMForTokenClassification', 'TFXLMMainLayer', 'TFXLMModel', 'TFXLMPreTrainedModel', 'TFXLMWithLMHeadModel', ] if TYPE_CHECKING: from .configuration_xlm import XLM_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMConfig, XLMOnnxConfig from .tokenization_xlm import XLMTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm import ( XLM_PRETRAINED_MODEL_ARCHIVE_LIST, XLMForMultipleChoice, XLMForQuestionAnswering, XLMForQuestionAnsweringSimple, XLMForSequenceClassification, XLMForTokenClassification, XLMModel, XLMPreTrainedModel, XLMWithLMHeadModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm import ( TF_XLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMForMultipleChoice, TFXLMForQuestionAnsweringSimple, TFXLMForSequenceClassification, TFXLMForTokenClassification, TFXLMMainLayer, TFXLMModel, TFXLMPreTrainedModel, TFXLMWithLMHeadModel, ) else: import sys __A : Optional[Any] = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
354
'''simple docstring''' from __future__ import annotations from typing import Dict from ...configuration_utils import PretrainedConfig __A : Dict = { 'susnato/ernie-m-base_pytorch': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/config.json', 'susnato/ernie-m-large_pytorch': 'https://huggingface.co/susnato/ernie-m-large_pytorch/blob/main/config.json', } class __UpperCamelCase ( lowercase__ ): lowercase : Optional[int] = 'ernie_m' lowercase : Dict[str, str] = {"dropout": "classifier_dropout", "num_classes": "num_labels"} def __init__( self :Optional[Any] ,_UpperCamelCase :int = 2_5_0_0_0_2 ,_UpperCamelCase :int = 7_6_8 ,_UpperCamelCase :int = 1_2 ,_UpperCamelCase :int = 1_2 ,_UpperCamelCase :int = 3_0_7_2 ,_UpperCamelCase :str = "gelu" ,_UpperCamelCase :float = 0.1 ,_UpperCamelCase :float = 0.1 ,_UpperCamelCase :int = 5_1_4 ,_UpperCamelCase :float = 0.02 ,_UpperCamelCase :int = 1 ,_UpperCamelCase :float = 1E-0_5 ,_UpperCamelCase :List[Any]=None ,_UpperCamelCase :List[str]=False ,_UpperCamelCase :Optional[int]=0.0 ,**_UpperCamelCase :List[Any] ,): super().__init__(pad_token_id=_UpperCamelCase ,**_UpperCamelCase ) snake_case_ : Optional[int] = vocab_size snake_case_ : Any = hidden_size snake_case_ : Union[str, Any] = num_hidden_layers snake_case_ : Union[str, Any] = num_attention_heads snake_case_ : Any = intermediate_size snake_case_ : Any = hidden_act snake_case_ : Tuple = hidden_dropout_prob snake_case_ : Union[str, Any] = attention_probs_dropout_prob snake_case_ : str = max_position_embeddings snake_case_ : int = initializer_range snake_case_ : Optional[Any] = layer_norm_eps snake_case_ : Union[str, Any] = classifier_dropout snake_case_ : Tuple = is_decoder snake_case_ : int = act_dropout
8
0
import warnings from typing import Dict, List, Optional, Tuple from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _UpperCAmelCase : List[str] = logging.get_logger(__name__) class __lowerCAmelCase ( lowerCAmelCase): _a = ['''input_ids''', '''attention_mask'''] def __init__( self: Dict , _lowerCAmelCase: List[Any]="</s>" , _lowerCAmelCase: Dict="<unk>" , _lowerCAmelCase: Optional[Any]="<pad>" , _lowerCAmelCase: Dict=1_25 , _lowerCAmelCase: int=None , **_lowerCAmelCase: List[str] , ): # Add extra_ids to the special token list if extra_ids > 0 and additional_special_tokens is None: lowercase :int = [F"<extra_id_{i}>" for i in range(_lowerCAmelCase )] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens lowercase :Optional[Any] = len(set(filter(lambda _lowerCAmelCase : bool("extra_id" in str(_lowerCAmelCase ) ) , _lowerCAmelCase ) ) ) if extra_tokens != extra_ids: raise ValueError( F"Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are" " provided to ByT5Tokenizer. In this case the additional_special_tokens must include the" " extra_ids tokens" ) lowercase :Optional[Any] = AddedToken(_lowerCAmelCase , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else pad_token lowercase :Optional[Any] = AddedToken(_lowerCAmelCase , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else eos_token lowercase :Union[str, Any] = AddedToken(_lowerCAmelCase , lstrip=_lowerCAmelCase , rstrip=_lowerCAmelCase ) if isinstance(_lowerCAmelCase , _lowerCAmelCase ) else unk_token super().__init__( eos_token=_lowerCAmelCase , unk_token=_lowerCAmelCase , pad_token=_lowerCAmelCase , extra_ids=_lowerCAmelCase , additional_special_tokens=_lowerCAmelCase , **_lowerCAmelCase , ) lowercase :List[str] = extra_ids lowercase :Optional[int] = 2**8 # utf is 8 bits # define special tokens dict lowercase :Dict[int, str] = { self.pad_token: 0, self.eos_token: 1, self.unk_token: 2, } lowercase :List[str] = len(self.special_tokens_encoder ) lowercase :List[str] = len(_lowerCAmelCase ) for i, token in enumerate(_lowerCAmelCase ): lowercase :Optional[Any] = self.vocab_size + i - n lowercase :Dict[str, int] = {v: k for k, v in self.special_tokens_encoder.items()} @property def SCREAMING_SNAKE_CASE ( self: Dict ): return self._utf_vocab_size + self._num_special_tokens + self._extra_ids def SCREAMING_SNAKE_CASE ( self: List[Any] , _lowerCAmelCase: List[int] , _lowerCAmelCase: Optional[List[int]] = None , _lowerCAmelCase: bool = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_lowerCAmelCase , token_ids_a=_lowerCAmelCase , already_has_special_tokens=_lowerCAmelCase ) # normal case: some special tokens if token_ids_a is None: return ([0] * len(_lowerCAmelCase )) + [1] return ([0] * len(_lowerCAmelCase )) + [1] + ([0] * len(_lowerCAmelCase )) + [1] def SCREAMING_SNAKE_CASE ( self: Dict , _lowerCAmelCase: List[int] ): if len(_lowerCAmelCase ) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( F"This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated" " eos tokens being added." ) return token_ids else: return token_ids + [self.eos_token_id] def SCREAMING_SNAKE_CASE ( self: Any , _lowerCAmelCase: List[int] , _lowerCAmelCase: Optional[List[int]] = None ): lowercase :int = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos ) * [0] return len(token_ids_a + eos + token_ids_a + eos ) * [0] def SCREAMING_SNAKE_CASE ( self: Dict , _lowerCAmelCase: List[int] , _lowerCAmelCase: Optional[List[int]] = None ): lowercase :List[str] = self._add_eos_if_not_present(_lowerCAmelCase ) if token_ids_a is None: return token_ids_a else: lowercase :List[str] = self._add_eos_if_not_present(_lowerCAmelCase ) return token_ids_a + token_ids_a def SCREAMING_SNAKE_CASE ( self: Optional[Any] , _lowerCAmelCase: str ): lowercase :Tuple = [chr(_lowerCAmelCase ) for i in text.encode("utf-8" )] return tokens def SCREAMING_SNAKE_CASE ( self: Tuple , _lowerCAmelCase: int ): if token in self.special_tokens_encoder: lowercase :int = self.special_tokens_encoder[token] elif token in self.added_tokens_encoder: lowercase :Dict = self.added_tokens_encoder[token] elif len(_lowerCAmelCase ) != 1: lowercase :List[str] = self.unk_token_id else: lowercase :Tuple = ord(_lowerCAmelCase ) + self._num_special_tokens return token_id def SCREAMING_SNAKE_CASE ( self: int , _lowerCAmelCase: Tuple ): if index in self.special_tokens_decoder: lowercase :List[str] = self.special_tokens_decoder[index] else: lowercase :List[str] = chr(index - self._num_special_tokens ) return token def SCREAMING_SNAKE_CASE ( self: Any , _lowerCAmelCase: str ): lowercase :Optional[Any] = b"" for token in tokens: if token in self.special_tokens_decoder: lowercase :int = self.special_tokens_decoder[token].encode("utf-8" ) elif token in self.added_tokens_decoder: lowercase :Optional[int] = self.special_tokens_decoder[token].encode("utf-8" ) elif token in self.special_tokens_encoder: lowercase :str = token.encode("utf-8" ) elif token in self.added_tokens_encoder: lowercase :List[str] = token.encode("utf-8" ) else: lowercase :Optional[Any] = bytes([ord(_lowerCAmelCase )] ) bstring += tok_string lowercase :Dict = bstring.decode("utf-8" , errors="ignore" ) return string def SCREAMING_SNAKE_CASE ( self: str , _lowerCAmelCase: str , _lowerCAmelCase: Optional[str] = None ): return ()
236
import math import sys def UpperCAmelCase__ ( lowerCamelCase ): if number != int(lowerCamelCase ): raise ValueError("the value of input must be a natural number" ) if number < 0: raise ValueError("the value of input must not be a negative number" ) if number == 0: return 1 lowercase :Dict = [-1] * (number + 1) lowercase :Optional[Any] = 0 for i in range(1, number + 1 ): lowercase :Optional[int] = sys.maxsize lowercase :Any = int(math.sqrt(lowerCamelCase ) ) for j in range(1, root + 1 ): lowercase :Any = 1 + answers[i - (j**2)] lowercase :Dict = min(lowerCamelCase, lowerCamelCase ) lowercase :Union[str, Any] = answer return answers[number] if __name__ == "__main__": import doctest doctest.testmod()
236
1
from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" def __init__( self : Dict , lowerCAmelCase_ : pyspark.sql.DataFrame , lowerCAmelCase_ : Optional[NamedSplit] = None , lowerCAmelCase_ : Optional[Features] = None , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : str = None , lowerCAmelCase_ : bool = False , lowerCAmelCase_ : str = None , lowerCAmelCase_ : bool = True , lowerCAmelCase_ : str = "arrow" , **lowerCAmelCase_ : Optional[int] , ) -> List[Any]: super().__init__( split=lowerCAmelCase_ , features=lowerCAmelCase_ , cache_dir=lowerCAmelCase_ , keep_in_memory=lowerCAmelCase_ , streaming=lowerCAmelCase_ , **lowerCAmelCase_ , ) __lowerCAmelCase = load_from_cache_file __lowerCAmelCase = file_format __lowerCAmelCase = Spark( df=lowerCAmelCase_ , features=lowerCAmelCase_ , cache_dir=lowerCAmelCase_ , working_dir=lowerCAmelCase_ , **lowerCAmelCase_ , ) def lowercase ( self : Union[str, Any] ) -> Any: if self.streaming: return self.builder.as_streaming_dataset(split=self.split ) __lowerCAmelCase = None if self._load_from_cache_file else DownloadMode.FORCE_REDOWNLOAD self.builder.download_and_prepare( download_mode=lowerCAmelCase_ , file_format=self._file_format , ) return self.builder.as_dataset(split=self.split )
357
import argparse import json from typing import List from ltp import LTP from transformers.models.bert.tokenization_bert import BertTokenizer def a_ ( lowerCAmelCase_ : Optional[Any] ): # This defines a "chinese character" as anything in the CJK Unicode block: # https://en.wikipedia.org/wiki/CJK_Unified_Ideographs_(Unicode_block) # # Note that the CJK Unicode block is NOT all Japanese and Korean characters, # despite its name. The modern Korean Hangul alphabet is a different block, # as is Japanese Hiragana and Katakana. Those alphabets are used to write # space-separated words, so they are not treated specially and handled # like the all of the other languages. if ( (cp >= 0X4e00 and cp <= 0X9fff) or (cp >= 0X3400 and cp <= 0X4dbf) # or (cp >= 0X20000 and cp <= 0X2a6df) # or (cp >= 0X2a700 and cp <= 0X2b73f) # or (cp >= 0X2b740 and cp <= 0X2b81f) # or (cp >= 0X2b820 and cp <= 0X2ceaf) # or (cp >= 0Xf900 and cp <= 0Xfaff) or (cp >= 0X2f800 and cp <= 0X2fa1f) # ): # return True return False def a_ ( lowerCAmelCase_ : str ): # word like '180' or '身高' or '神' for char in word: __lowerCAmelCase = ord(lowerCAmelCase_ ) if not _is_chinese_char(lowerCAmelCase_ ): return 0 return 1 def a_ ( lowerCAmelCase_ : List[str] ): __lowerCAmelCase = set() for token in tokens: __lowerCAmelCase = len(lowerCAmelCase_ ) > 1 and is_chinese(lowerCAmelCase_ ) if chinese_word: word_set.add(lowerCAmelCase_ ) __lowerCAmelCase = list(lowerCAmelCase_ ) return word_list def a_ ( lowerCAmelCase_ : List[str], lowerCAmelCase_ : set() ): if not chinese_word_set: return bert_tokens __lowerCAmelCase = max([len(lowerCAmelCase_ ) for w in chinese_word_set] ) __lowerCAmelCase = bert_tokens __lowerCAmelCase , __lowerCAmelCase = 0, len(lowerCAmelCase_ ) while start < end: __lowerCAmelCase = True if is_chinese(bert_word[start] ): __lowerCAmelCase = min(end - start, lowerCAmelCase_ ) for i in range(lowerCAmelCase_, 1, -1 ): __lowerCAmelCase = ''.join(bert_word[start : start + i] ) if whole_word in chinese_word_set: for j in range(start + 1, start + i ): __lowerCAmelCase = '##' + bert_word[j] __lowerCAmelCase = start + i __lowerCAmelCase = False break if single_word: start += 1 return bert_word def a_ ( lowerCAmelCase_ : List[str], lowerCAmelCase_ : LTP, lowerCAmelCase_ : BertTokenizer ): __lowerCAmelCase = [] for i in range(0, len(lowerCAmelCase_ ), 100 ): __lowerCAmelCase = ltp_tokenizer.pipeline(lines[i : i + 100], tasks=['cws'] ).cws __lowerCAmelCase = [get_chinese_word(lowerCAmelCase_ ) for r in res] ltp_res.extend(lowerCAmelCase_ ) assert len(lowerCAmelCase_ ) == len(lowerCAmelCase_ ) __lowerCAmelCase = [] for i in range(0, len(lowerCAmelCase_ ), 100 ): __lowerCAmelCase = bert_tokenizer(lines[i : i + 100], add_special_tokens=lowerCAmelCase_, truncation=lowerCAmelCase_, max_length=512 ) bert_res.extend(res['input_ids'] ) assert len(lowerCAmelCase_ ) == len(lowerCAmelCase_ ) __lowerCAmelCase = [] for input_ids, chinese_word in zip(lowerCAmelCase_, lowerCAmelCase_ ): __lowerCAmelCase = [] for id in input_ids: __lowerCAmelCase = bert_tokenizer._convert_id_to_token(lowerCAmelCase_ ) input_tokens.append(lowerCAmelCase_ ) __lowerCAmelCase = add_sub_symbol(lowerCAmelCase_, lowerCAmelCase_ ) __lowerCAmelCase = [] # We only save pos of chinese subwords start with ##, which mean is part of a whole word. for i, token in enumerate(lowerCAmelCase_ ): if token[:2] == "##": __lowerCAmelCase = token[2:] # save chinese tokens' pos if len(lowerCAmelCase_ ) == 1 and _is_chinese_char(ord(lowerCAmelCase_ ) ): ref_id.append(lowerCAmelCase_ ) ref_ids.append(lowerCAmelCase_ ) assert len(lowerCAmelCase_ ) == len(lowerCAmelCase_ ) return ref_ids def a_ ( lowerCAmelCase_ : int ): # For Chinese (Ro)Bert, the best result is from : RoBERTa-wwm-ext (https://github.com/ymcui/Chinese-BERT-wwm) # If we want to fine-tune these model, we have to use same tokenizer : LTP (https://github.com/HIT-SCIR/ltp) with open(args.file_name, 'r', encoding='utf-8' ) as f: __lowerCAmelCase = f.readlines() __lowerCAmelCase = [line.strip() for line in data if len(lowerCAmelCase_ ) > 0 and not line.isspace()] # avoid delimiter like '\u2029' __lowerCAmelCase = LTP(args.ltp ) # faster in GPU device __lowerCAmelCase = BertTokenizer.from_pretrained(args.bert ) __lowerCAmelCase = prepare_ref(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) with open(args.save_path, 'w', encoding='utf-8' ) as f: __lowerCAmelCase = [json.dumps(lowerCAmelCase_ ) + '\n' for ref in ref_ids] f.writelines(lowerCAmelCase_ ) if __name__ == "__main__": _snake_case : int = argparse.ArgumentParser(description='prepare_chinese_ref') parser.add_argument( '--file_name', required=False, type=str, default='./resources/chinese-demo.txt', help='file need process, same as training data in lm', ) parser.add_argument( '--ltp', required=False, type=str, default='./resources/ltp', help='resources for LTP tokenizer, usually a path', ) parser.add_argument( '--bert', required=False, type=str, default='./resources/robert', help='resources for Bert tokenizer', ) parser.add_argument( '--save_path', required=False, type=str, default='./resources/ref.txt', help='path to save res', ) _snake_case : List[str] = parser.parse_args() main(args)
207
0
'''simple docstring''' import json import os import shutil import warnings from argparse import ArgumentParser, Namespace from pathlib import Path from typing import List from ..utils import logging from . import BaseTransformersCLICommand try: from cookiecutter.main import cookiecutter a__ : str =True except ImportError: a__ : Tuple =False a__ : int =logging.get_logger(__name__) # pylint: disable=invalid-name def lowercase__ ( __lowercase : Namespace ) -> List[str]: """simple docstring""" return AddNewModelCommand(args.testing , args.testing_file , path=args.path ) class snake_case ( __lowerCamelCase ): """simple docstring""" @staticmethod def _lowerCamelCase ( __A : ArgumentParser ): __UpperCamelCase = parser.add_parser('add-new-model' ) add_new_model_parser.add_argument('--testing' , action='store_true' , help='If in testing mode.' ) add_new_model_parser.add_argument('--testing_file' , type=__A , help='Configuration file on which to run.' ) add_new_model_parser.add_argument( '--path' , type=__A , help='Path to cookiecutter. Should only be used for testing purposes.' ) add_new_model_parser.set_defaults(func=__A ) def __init__( self : Dict , __A : bool , __A : str , __A : Any=None , *__A : Optional[int] ): __UpperCamelCase = testing __UpperCamelCase = testing_file __UpperCamelCase = path def _lowerCamelCase ( self : str ): warnings.warn( 'The command `transformers-cli add-new-model` is deprecated and will be removed in v5 of Transformers. ' 'It is not actively maintained anymore, so might give a result that won\'t pass all tests and quality ' 'checks, you should use `transformers-cli add-new-model-like` instead.' ) if not _has_cookiecutter: raise ImportError( 'Model creation dependencies are required to use the `add_new_model` command. Install them by running ' 'the following at the root of your `transformers` clone:\n\n\t$ pip install -e .[modelcreation]\n' ) # Ensure that there is no other `cookiecutter-template-xxx` directory in the current working directory __UpperCamelCase = [directory for directory in os.listdir() if 'cookiecutter-template-' == directory[:2_2]] if len(__A ) > 0: raise ValueError( 'Several directories starting with `cookiecutter-template-` in current working directory. ' 'Please clean your directory by removing all folders starting with `cookiecutter-template-` or ' 'change your working directory.' ) __UpperCamelCase = ( Path(__A ).parent.parent.parent.parent if self._path is None else Path(self._path ).parent.parent ) __UpperCamelCase = path_to_transformer_root / 'templates' / 'adding_a_new_model' # Execute cookiecutter if not self._testing: cookiecutter(str(__A ) ) else: with open(self._testing_file , 'r' ) as configuration_file: __UpperCamelCase = json.load(__A ) cookiecutter( str(path_to_cookiecutter if self._path is None else self._path ) , no_input=__A , extra_context=__A , ) __UpperCamelCase = [directory for directory in os.listdir() if 'cookiecutter-template-' in directory[:2_2]][0] # Retrieve configuration with open(directory + '/configuration.json' , 'r' ) as configuration_file: __UpperCamelCase = json.load(__A ) __UpperCamelCase = configuration['lowercase_modelname'] __UpperCamelCase = configuration['generate_tensorflow_pytorch_and_flax'] os.remove(f'''{directory}/configuration.json''' ) __UpperCamelCase = 'PyTorch' in generate_tensorflow_pytorch_and_flax __UpperCamelCase = 'TensorFlow' in generate_tensorflow_pytorch_and_flax __UpperCamelCase = 'Flax' in generate_tensorflow_pytorch_and_flax __UpperCamelCase = f'''{path_to_transformer_root}/src/transformers/models/{lowercase_model_name}''' os.makedirs(__A , exist_ok=__A ) os.makedirs(f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}''' , exist_ok=__A ) # Tests require submodules as they have parent imports with open(f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/__init__.py''' , 'w' ): pass shutil.move( f'''{directory}/__init__.py''' , f'''{model_dir}/__init__.py''' , ) shutil.move( f'''{directory}/configuration_{lowercase_model_name}.py''' , f'''{model_dir}/configuration_{lowercase_model_name}.py''' , ) def remove_copy_lines(__A : Any ): with open(__A , 'r' ) as f: __UpperCamelCase = f.readlines() with open(__A , 'w' ) as f: for line in lines: if "# Copied from transformers." not in line: f.write(__A ) if output_pytorch: if not self._testing: remove_copy_lines(f'''{directory}/modeling_{lowercase_model_name}.py''' ) shutil.move( f'''{directory}/modeling_{lowercase_model_name}.py''' , f'''{model_dir}/modeling_{lowercase_model_name}.py''' , ) shutil.move( f'''{directory}/test_modeling_{lowercase_model_name}.py''' , f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_{lowercase_model_name}.py''' , ) else: os.remove(f'''{directory}/modeling_{lowercase_model_name}.py''' ) os.remove(f'''{directory}/test_modeling_{lowercase_model_name}.py''' ) if output_tensorflow: if not self._testing: remove_copy_lines(f'''{directory}/modeling_tf_{lowercase_model_name}.py''' ) shutil.move( f'''{directory}/modeling_tf_{lowercase_model_name}.py''' , f'''{model_dir}/modeling_tf_{lowercase_model_name}.py''' , ) shutil.move( f'''{directory}/test_modeling_tf_{lowercase_model_name}.py''' , f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_tf_{lowercase_model_name}.py''' , ) else: os.remove(f'''{directory}/modeling_tf_{lowercase_model_name}.py''' ) os.remove(f'''{directory}/test_modeling_tf_{lowercase_model_name}.py''' ) if output_flax: if not self._testing: remove_copy_lines(f'''{directory}/modeling_flax_{lowercase_model_name}.py''' ) shutil.move( f'''{directory}/modeling_flax_{lowercase_model_name}.py''' , f'''{model_dir}/modeling_flax_{lowercase_model_name}.py''' , ) shutil.move( f'''{directory}/test_modeling_flax_{lowercase_model_name}.py''' , f'''{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_flax_{lowercase_model_name}.py''' , ) else: os.remove(f'''{directory}/modeling_flax_{lowercase_model_name}.py''' ) os.remove(f'''{directory}/test_modeling_flax_{lowercase_model_name}.py''' ) shutil.move( f'''{directory}/{lowercase_model_name}.md''' , f'''{path_to_transformer_root}/docs/source/en/model_doc/{lowercase_model_name}.md''' , ) shutil.move( f'''{directory}/tokenization_{lowercase_model_name}.py''' , f'''{model_dir}/tokenization_{lowercase_model_name}.py''' , ) shutil.move( f'''{directory}/tokenization_fast_{lowercase_model_name}.py''' , f'''{model_dir}/tokenization_{lowercase_model_name}_fast.py''' , ) from os import fdopen, remove from shutil import copymode, move from tempfile import mkstemp def replace(__A : str , __A : str , __A : List[str] ): # Create temp file __UpperCamelCase , __UpperCamelCase = mkstemp() __UpperCamelCase = False with fdopen(__A , 'w' ) as new_file: with open(__A ) as old_file: for line in old_file: new_file.write(__A ) if line_to_copy_below in line: __UpperCamelCase = True for line_to_copy in lines_to_copy: new_file.write(__A ) if not line_found: raise ValueError(f'''Line {line_to_copy_below} was not found in file.''' ) # Copy the file permissions from the old file to the new file copymode(__A , __A ) # Remove original file remove(__A ) # Move new file move(__A , __A ) def skip_units(__A : Dict ): return ( ("generating PyTorch" in line and not output_pytorch) or ("generating TensorFlow" in line and not output_tensorflow) or ("generating Flax" in line and not output_flax) ) def replace_in_files(__A : Optional[Any] ): with open(__A ) as datafile: __UpperCamelCase = [] __UpperCamelCase = False __UpperCamelCase = False for line in datafile: if "# To replace in: " in line and "##" not in line: __UpperCamelCase = line.split('"' )[1] __UpperCamelCase = skip_units(__A ) elif "# Below: " in line and "##" not in line: __UpperCamelCase = line.split('"' )[1] __UpperCamelCase = skip_units(__A ) elif "# End." in line and "##" not in line: if not skip_file and not skip_snippet: replace(__A , __A , __A ) __UpperCamelCase = [] elif "# Replace with" in line and "##" not in line: __UpperCamelCase = [] elif "##" not in line: lines_to_copy.append(__A ) remove(__A ) replace_in_files(f'''{directory}/to_replace_{lowercase_model_name}.py''' ) os.rmdir(__A )
53
_a = { '''A''': ['''B''', '''C''', '''E'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F''', '''G'''], '''D''': ['''B'''], '''E''': ['''A''', '''B''', '''D'''], '''F''': ['''C'''], '''G''': ['''C'''], } def _a ( SCREAMING_SNAKE_CASE : dict , SCREAMING_SNAKE_CASE : int , SCREAMING_SNAKE_CASE : Any ) -> list[str]: """simple docstring""" __lowerCAmelCase: int = set() # keep track of all the paths to be checked __lowerCAmelCase: str = [[start]] # return path if start is goal if start == goal: return [start] # keeps looping until all possible paths have been checked while queue: # pop the first path from the queue __lowerCAmelCase: str = queue.pop(0 ) # get the last node from the path __lowerCAmelCase: Union[str, Any] = path[-1] if node not in explored: __lowerCAmelCase: Dict = graph[node] # go through all neighbour nodes, construct a new path and # push it into the queue for neighbour in neighbours: __lowerCAmelCase: Dict = list(SCREAMING_SNAKE_CASE ) new_path.append(SCREAMING_SNAKE_CASE ) queue.append(SCREAMING_SNAKE_CASE ) # return path if neighbour is goal if neighbour == goal: return new_path # mark node as explored explored.add(SCREAMING_SNAKE_CASE ) # in case there's no path between the 2 nodes return [] def _a ( SCREAMING_SNAKE_CASE : dict , SCREAMING_SNAKE_CASE : Union[str, Any] , SCREAMING_SNAKE_CASE : Any ) -> int: """simple docstring""" if not graph or start not in graph or target not in graph: return -1 if start == target: return 0 __lowerCAmelCase: Optional[int] = [start] __lowerCAmelCase: Dict = set(SCREAMING_SNAKE_CASE ) # Keep tab on distances from `start` node. __lowerCAmelCase: Optional[int] = {start: 0, target: -1} while queue: __lowerCAmelCase: Any = queue.pop(0 ) if node == target: __lowerCAmelCase: Optional[int] = ( dist[node] if dist[target] == -1 else min(dist[target] , dist[node] ) ) for adjacent in graph[node]: if adjacent not in visited: visited.add(SCREAMING_SNAKE_CASE ) queue.append(SCREAMING_SNAKE_CASE ) __lowerCAmelCase: Union[str, Any] = dist[node] + 1 return dist[target] if __name__ == "__main__": print(bfs_shortest_path(demo_graph, '''G''', '''D''')) # returns ['G', 'C', 'A', 'B', 'D'] print(bfs_shortest_path_distance(demo_graph, '''G''', '''D''')) # returns 4
322
0
"""simple docstring""" from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. lowercase__ = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be smaller than N_POPULATION. lowercase__ = 50 # Probability that an element of a generation can mutate, changing one of its genes. # This will guarantee that all genes will be used during evolution. lowercase__ = 0.4 # Just a seed to improve randomness required by the algorithm. random.seed(random.randint(0, 1000)) def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : str = len([g for position, g in enumerate(lowercase__ ) if g == main_target[position]] ) return (item, float(lowercase__ )) def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : Any = random.randint(0 , len(lowercase__ ) - 1 ) _lowerCamelCase : Union[str, Any] = parent_a[:random_slice] + parent_a[random_slice:] _lowerCamelCase : Tuple = parent_a[:random_slice] + parent_a[random_slice:] return (child_a, child_a) def _snake_case ( lowercase__ , lowercase__ ): _lowerCamelCase : List[str] = list(lowercase__ ) if random.uniform(0 , 1 ) < MUTATION_PROBABILITY: _lowerCamelCase : Tuple = random.choice(lowercase__ ) return "".join(lowercase__ ) def _snake_case ( lowercase__ , lowercase__ , lowercase__ , ): _lowerCamelCase : Tuple = [] # Generate more children proportionally to the fitness score. _lowerCamelCase : str = int(parent_a[1] * 100 ) + 1 _lowerCamelCase : Dict = 10 if child_n >= 10 else child_n for _ in range(lowercase__ ): _lowerCamelCase : str = population_score[random.randint(0 , lowercase__ )][0] _lowerCamelCase, _lowerCamelCase : Dict = crossover(parent_a[0] , lowercase__ ) # Append new string to the population list. pop.append(mutate(lowercase__ , lowercase__ ) ) pop.append(mutate(lowercase__ , lowercase__ ) ) return pop def _snake_case ( lowercase__ , lowercase__ , lowercase__ = True ): # Verify if N_POPULATION is bigger than N_SELECTED if N_POPULATION < N_SELECTED: _lowerCamelCase : str = f'''{N_POPULATION} must be bigger than {N_SELECTED}''' raise ValueError(lowercase__ ) # Verify that the target contains no genes besides the ones inside genes variable. _lowerCamelCase : Tuple = sorted({c for c in target if c not in genes} ) if not_in_genes_list: _lowerCamelCase : Tuple = f'''{not_in_genes_list} is not in genes list, evolution cannot converge''' raise ValueError(lowercase__ ) # Generate random starting population. _lowerCamelCase : Tuple = [] for _ in range(lowercase__ ): population.append(''.join([random.choice(lowercase__ ) for i in range(len(lowercase__ ) )] ) ) # Just some logs to know what the algorithms is doing. _lowerCamelCase, _lowerCamelCase : Optional[Any] = 0, 0 # This loop will end when we find a perfect match for our target. while True: generation += 1 total_population += len(lowercase__ ) # Random population created. Now it's time to evaluate. # Adding a bit of concurrency can make everything faster, # # import concurrent.futures # population_score: list[tuple[str, float]] = [] # with concurrent.futures.ThreadPoolExecutor( # max_workers=NUM_WORKERS) as executor: # futures = {executor.submit(evaluate, item) for item in population} # concurrent.futures.wait(futures) # population_score = [item.result() for item in futures] # # but with a simple algorithm like this, it will probably be slower. # We just need to call evaluate for every item inside the population. _lowerCamelCase : Union[str, Any] = [evaluate(lowercase__ , lowercase__ ) for item in population] # Check if there is a matching evolution. _lowerCamelCase : List[str] = sorted(lowercase__ , key=lambda lowercase__ : x[1] , reverse=lowercase__ ) if population_score[0][0] == target: return (generation, total_population, population_score[0][0]) # Print the best result every 10 generation. # Just to know that the algorithm is working. if debug and generation % 10 == 0: print( f'''\nGeneration: {generation}''' f'''\nTotal Population:{total_population}''' f'''\nBest score: {population_score[0][1]}''' f'''\nBest string: {population_score[0][0]}''' ) # Flush the old population, keeping some of the best evolutions. # Keeping this avoid regression of evolution. _lowerCamelCase : Tuple = population[: int(N_POPULATION / 3 )] population.clear() population.extend(lowercase__ ) # Normalize population score to be between 0 and 1. _lowerCamelCase : str = [ (item, score / len(lowercase__ )) for item, score in population_score ] # This is selection for i in range(lowercase__ ): population.extend(select(population_score[int(lowercase__ )] , lowercase__ , lowercase__ ) ) # Check if the population has already reached the maximum value and if so, # break the cycle. If this check is disabled, the algorithm will take # forever to compute large strings, but will also calculate small strings in # a far fewer generations. if len(lowercase__ ) > N_POPULATION: break if __name__ == "__main__": lowercase__ = ( """This is a genetic algorithm to evaluate, combine, evolve, and mutate a string!""" ) lowercase__ = list( """ ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklm""" """nopqrstuvwxyz.,;!?+-*#@^'èéòà€ù=)(&%$£/\\""" ) lowercase__ , lowercase__ , lowercase__ = basic(target_str, genes_list) print( F"\nGeneration: {generation}\nTotal Population: {population}\nTarget: {target}" )
12
"""simple docstring""" import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' lowerCamelCase__ = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def A_ ( self , lowercase , lowercase , lowercase ): _lowerCamelCase : Optional[int] = hf_hub_download( repo_id='nateraw/video-demo' , filename='archery.mp4' , repo_type='dataset' ) _lowerCamelCase : Tuple = VideoClassificationPipeline(model=lowercase , image_processor=lowercase , top_k=2 ) _lowerCamelCase : List[str] = [ example_video_filepath, 'https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4', ] return video_classifier, examples def A_ ( self , lowercase , lowercase ): for example in examples: _lowerCamelCase : Tuple = video_classifier(lowercase ) self.assertEqual( lowercase , [ {'score': ANY(lowercase ), 'label': ANY(lowercase )}, {'score': ANY(lowercase ), 'label': ANY(lowercase )}, ] , ) @require_torch def A_ ( self ): _lowerCamelCase : Optional[Any] = 'hf-internal-testing/tiny-random-VideoMAEForVideoClassification' _lowerCamelCase : Tuple = VideoMAEFeatureExtractor( size={'shortest_edge': 10} , crop_size={'height': 10, 'width': 10} ) _lowerCamelCase : Dict = pipeline( 'video-classification' , model=lowercase , feature_extractor=lowercase , frame_sampling_rate=4 ) _lowerCamelCase : Any = hf_hub_download(repo_id='nateraw/video-demo' , filename='archery.mp4' , repo_type='dataset' ) _lowerCamelCase : Dict = video_classifier(lowercase , top_k=2 ) self.assertEqual( nested_simplify(lowercase , decimals=4 ) , [{'score': 0.51_99, 'label': 'LABEL_0'}, {'score': 0.48_01, 'label': 'LABEL_1'}] , ) _lowerCamelCase : str = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(lowercase , decimals=4 ) , [ [{'score': 0.51_99, 'label': 'LABEL_0'}, {'score': 0.48_01, 'label': 'LABEL_1'}], [{'score': 0.51_99, 'label': 'LABEL_0'}, {'score': 0.48_01, 'label': 'LABEL_1'}], ] , ) @require_tf def A_ ( self ): pass
12
1
"""simple docstring""" import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class _UpperCAmelCase : def __init__( self : Union[str, Any] , lowercase_ : List[Any] , lowercase_ : int=13 , lowercase_ : Optional[int]=7 , lowercase_ : Any=True , lowercase_ : Dict=True , lowercase_ : Dict=True , lowercase_ : Optional[Any]=99 , lowercase_ : Union[str, Any]=32 , lowercase_ : str=5 , lowercase_ : Union[str, Any]=4 , lowercase_ : Any=37 , lowercase_ : Tuple="gelu" , lowercase_ : Dict=0.1 , lowercase_ : Tuple=0.1 , lowercase_ : Optional[int]=512 , lowercase_ : Optional[Any]=16 , lowercase_ : Optional[Any]=2 , lowercase_ : Optional[Any]=0.02 , lowercase_ : List[Any]=3 , lowercase_ : Union[str, Any]=4 , lowercase_ : List[Any]=None , ): snake_case_ : Any = parent snake_case_ : List[str] = batch_size snake_case_ : List[Any] = seq_length snake_case_ : Optional[int] = is_training snake_case_ : Union[str, Any] = use_token_type_ids snake_case_ : Optional[Any] = use_labels snake_case_ : Union[str, Any] = vocab_size snake_case_ : Any = hidden_size snake_case_ : List[Any] = num_hidden_layers snake_case_ : Any = num_attention_heads snake_case_ : Dict = intermediate_size snake_case_ : Union[str, Any] = hidden_act snake_case_ : Optional[int] = hidden_dropout_prob snake_case_ : Optional[Any] = attention_probs_dropout_prob snake_case_ : Tuple = max_position_embeddings snake_case_ : int = type_vocab_size snake_case_ : Tuple = type_sequence_label_size snake_case_ : str = initializer_range snake_case_ : Tuple = num_labels snake_case_ : str = num_choices snake_case_ : Any = scope snake_case_ : Dict = self.vocab_size - 1 def _snake_case ( self : int ): snake_case_ : Optional[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case_ : Optional[Any] = None if self.use_token_type_ids: snake_case_ : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) snake_case_ : str = None snake_case_ : Dict = None snake_case_ : str = None if self.use_labels: snake_case_ : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) snake_case_ : Tuple = ids_tensor([self.batch_size] , self.num_choices ) snake_case_ : int = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) snake_case_ : Any = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def _snake_case ( self : Tuple , lowercase_ : Any , lowercase_ : Union[str, Any] , lowercase_ : str , lowercase_ : Dict , *lowercase_ : Dict ): snake_case_ : List[Any] = OpenAIGPTModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() snake_case_ : Any = model(lowercase_ , token_type_ids=lowercase_ , head_mask=lowercase_ ) snake_case_ : Optional[Any] = model(lowercase_ , token_type_ids=lowercase_ ) snake_case_ : Optional[Any] = model(lowercase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _snake_case ( self : Tuple , lowercase_ : Dict , lowercase_ : str , lowercase_ : Optional[Any] , lowercase_ : List[Any] , *lowercase_ : Optional[Any] ): snake_case_ : Union[str, Any] = OpenAIGPTLMHeadModel(lowercase_ ) model.to(lowercase_ ) model.eval() snake_case_ : Union[str, Any] = model(lowercase_ , token_type_ids=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _snake_case ( self : List[str] , lowercase_ : Dict , lowercase_ : List[str] , lowercase_ : Any , lowercase_ : Dict , *lowercase_ : Union[str, Any] ): snake_case_ : Tuple = OpenAIGPTDoubleHeadsModel(lowercase_ ) model.to(lowercase_ ) model.eval() snake_case_ : Dict = model(lowercase_ , token_type_ids=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _snake_case ( self : Any , lowercase_ : str , lowercase_ : List[str] , lowercase_ : Optional[Any] , lowercase_ : Optional[Any] , *lowercase_ : Any ): snake_case_ : int = self.num_labels snake_case_ : Any = OpenAIGPTForSequenceClassification(lowercase_ ) model.to(lowercase_ ) model.eval() snake_case_ : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) snake_case_ : Optional[Any] = model(lowercase_ , token_type_ids=lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _snake_case ( self : int ): snake_case_ : Dict = self.prepare_config_and_inputs() ( ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ( snake_case_ ), ) : str = config_and_inputs snake_case_ : str = { '''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''head_mask''': head_mask, } return config, inputs_dict @require_torch class _UpperCAmelCase ( lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , unittest.TestCase): _lowerCAmelCase : Dict = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) _lowerCAmelCase : int = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly _lowerCAmelCase : Union[str, Any] = ( { """feature-extraction""": OpenAIGPTModel, """text-classification""": OpenAIGPTForSequenceClassification, """text-generation""": OpenAIGPTLMHeadModel, """zero-shot""": OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def _snake_case ( self : Tuple , lowercase_ : Optional[int] , lowercase_ : int , lowercase_ : List[Any] , lowercase_ : List[Any] , lowercase_ : Union[str, Any] ): if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def _snake_case ( self : Optional[int] , lowercase_ : List[Any] , lowercase_ : Optional[int] , lowercase_ : List[str]=False ): snake_case_ : Dict = super()._prepare_for_class(lowercase_ , lowercase_ , return_labels=lowercase_ ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": snake_case_ : List[str] = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=lowercase_ , ) snake_case_ : int = inputs_dict['''labels'''] snake_case_ : Optional[Any] = inputs_dict['''labels'''] snake_case_ : int = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=lowercase_ , ) snake_case_ : Tuple = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowercase_ ) return inputs_dict def _snake_case ( self : Any ): snake_case_ : List[str] = OpenAIGPTModelTester(self ) snake_case_ : Dict = ConfigTester(self , config_class=lowercase_ , n_embd=37 ) def _snake_case ( self : List[str] ): self.config_tester.run_common_tests() def _snake_case ( self : Optional[Any] ): snake_case_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*lowercase_ ) def _snake_case ( self : List[str] ): snake_case_ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*lowercase_ ) def _snake_case ( self : int ): snake_case_ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*lowercase_ ) def _snake_case ( self : List[str] ): snake_case_ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*lowercase_ ) @slow def _snake_case ( self : Dict ): for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case_ : Optional[Any] = OpenAIGPTModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) @require_torch class _UpperCAmelCase ( unittest.TestCase): @slow def _snake_case ( self : Optional[int] ): snake_case_ : Optional[Any] = OpenAIGPTLMHeadModel.from_pretrained('''openai-gpt''' ) model.to(lowercase_ ) snake_case_ : List[str] = torch.tensor([[481, 4735, 544]] , dtype=torch.long , device=lowercase_ ) # the president is snake_case_ : List[Any] = [ 481, 4735, 544, 246, 963, 870, 762, 239, 244, 40477, 244, 249, 719, 881, 487, 544, 240, 244, 603, 481, ] # the president is a very good man. " \n " i\'m sure he is, " said the snake_case_ : Optional[Any] = model.generate(lowercase_ , do_sample=lowercase_ ) self.assertListEqual(output_ids[0].tolist() , lowercase_ )
264
"""simple docstring""" import shutil import tempfile import unittest from unittest.mock import patch from transformers import ( DefaultFlowCallback, IntervalStrategy, PrinterCallback, ProgressCallback, Trainer, TrainerCallback, TrainingArguments, is_torch_available, ) from transformers.testing_utils import require_torch if is_torch_available(): from transformers.trainer import DEFAULT_CALLBACKS from .test_trainer import RegressionDataset, RegressionModelConfig, RegressionPreTrainedModel class _UpperCAmelCase ( lowerCAmelCase__): def __init__( self : Optional[int] ): snake_case_ : str = [] def _snake_case ( self : List[Any] , lowercase_ : Any , lowercase_ : Union[str, Any] , lowercase_ : List[str] , **lowercase_ : Tuple ): self.events.append('''on_init_end''' ) def _snake_case ( self : List[Any] , lowercase_ : str , lowercase_ : Optional[int] , lowercase_ : List[str] , **lowercase_ : List[str] ): self.events.append('''on_train_begin''' ) def _snake_case ( self : Any , lowercase_ : List[str] , lowercase_ : Tuple , lowercase_ : List[Any] , **lowercase_ : Optional[int] ): self.events.append('''on_train_end''' ) def _snake_case ( self : str , lowercase_ : Optional[int] , lowercase_ : int , lowercase_ : Optional[Any] , **lowercase_ : List[Any] ): self.events.append('''on_epoch_begin''' ) def _snake_case ( self : Tuple , lowercase_ : List[str] , lowercase_ : Dict , lowercase_ : Union[str, Any] , **lowercase_ : Optional[Any] ): self.events.append('''on_epoch_end''' ) def _snake_case ( self : List[str] , lowercase_ : Optional[Any] , lowercase_ : Optional[Any] , lowercase_ : int , **lowercase_ : Optional[Any] ): self.events.append('''on_step_begin''' ) def _snake_case ( self : int , lowercase_ : int , lowercase_ : Union[str, Any] , lowercase_ : List[Any] , **lowercase_ : List[str] ): self.events.append('''on_step_end''' ) def _snake_case ( self : str , lowercase_ : int , lowercase_ : Dict , lowercase_ : List[str] , **lowercase_ : List[str] ): self.events.append('''on_evaluate''' ) def _snake_case ( self : Dict , lowercase_ : Union[str, Any] , lowercase_ : Any , lowercase_ : List[Any] , **lowercase_ : str ): self.events.append('''on_predict''' ) def _snake_case ( self : List[Any] , lowercase_ : Union[str, Any] , lowercase_ : List[Any] , lowercase_ : int , **lowercase_ : Union[str, Any] ): self.events.append('''on_save''' ) def _snake_case ( self : str , lowercase_ : Tuple , lowercase_ : Optional[int] , lowercase_ : List[str] , **lowercase_ : Any ): self.events.append('''on_log''' ) def _snake_case ( self : Dict , lowercase_ : Optional[int] , lowercase_ : List[str] , lowercase_ : Union[str, Any] , **lowercase_ : Optional[int] ): self.events.append('''on_prediction_step''' ) @require_torch class _UpperCAmelCase ( unittest.TestCase): def _snake_case ( self : List[str] ): snake_case_ : Tuple = tempfile.mkdtemp() def _snake_case ( self : Tuple ): shutil.rmtree(self.output_dir ) def _snake_case ( self : int , lowercase_ : Union[str, Any]=0 , lowercase_ : Dict=0 , lowercase_ : List[str]=64 , lowercase_ : Union[str, Any]=64 , lowercase_ : Union[str, Any]=None , lowercase_ : Any=False , **lowercase_ : List[Any] ): # disable_tqdm in TrainingArguments has a flaky default since it depends on the level of logging. We make sure # its set to False since the tests later on depend on its value. snake_case_ : int = RegressionDataset(length=lowercase_ ) snake_case_ : Any = RegressionDataset(length=lowercase_ ) snake_case_ : int = RegressionModelConfig(a=lowercase_ , b=lowercase_ ) snake_case_ : Tuple = RegressionPreTrainedModel(lowercase_ ) snake_case_ : Any = TrainingArguments(self.output_dir , disable_tqdm=lowercase_ , report_to=[] , **lowercase_ ) return Trainer( lowercase_ , lowercase_ , train_dataset=lowercase_ , eval_dataset=lowercase_ , callbacks=lowercase_ , ) def _snake_case ( self : Optional[int] , lowercase_ : Any , lowercase_ : List[Any] ): self.assertEqual(len(lowercase_ ) , len(lowercase_ ) ) # Order doesn't matter snake_case_ : Any = sorted(lowercase_ , key=lambda lowercase_ : cb.__name__ if isinstance(lowercase_ , lowercase_ ) else cb.__class__.__name__ ) snake_case_ : List[str] = sorted(lowercase_ , key=lambda lowercase_ : cb.__name__ if isinstance(lowercase_ , lowercase_ ) else cb.__class__.__name__ ) for cba, cba in zip(lowercase_ , lowercase_ ): if isinstance(lowercase_ , lowercase_ ) and isinstance(lowercase_ , lowercase_ ): self.assertEqual(lowercase_ , lowercase_ ) elif isinstance(lowercase_ , lowercase_ ) and not isinstance(lowercase_ , lowercase_ ): self.assertEqual(lowercase_ , cba.__class__ ) elif not isinstance(lowercase_ , lowercase_ ) and isinstance(lowercase_ , lowercase_ ): self.assertEqual(cba.__class__ , lowercase_ ) else: self.assertEqual(lowercase_ , lowercase_ ) def _snake_case ( self : Optional[Any] , lowercase_ : Tuple ): snake_case_ : Tuple = ['''on_init_end''', '''on_train_begin'''] snake_case_ : List[Any] = 0 snake_case_ : Union[str, Any] = len(trainer.get_eval_dataloader() ) snake_case_ : List[Any] = ['''on_prediction_step'''] * len(trainer.get_eval_dataloader() ) + ['''on_log''', '''on_evaluate'''] for _ in range(trainer.state.num_train_epochs ): expected_events.append('''on_epoch_begin''' ) for _ in range(lowercase_ ): step += 1 expected_events += ["on_step_begin", "on_step_end"] if step % trainer.args.logging_steps == 0: expected_events.append('''on_log''' ) if trainer.args.evaluation_strategy == IntervalStrategy.STEPS and step % trainer.args.eval_steps == 0: expected_events += evaluation_events.copy() if step % trainer.args.save_steps == 0: expected_events.append('''on_save''' ) expected_events.append('''on_epoch_end''' ) if trainer.args.evaluation_strategy == IntervalStrategy.EPOCH: expected_events += evaluation_events.copy() expected_events += ["on_log", "on_train_end"] return expected_events def _snake_case ( self : List[str] ): snake_case_ : Union[str, Any] = self.get_trainer() snake_case_ : Dict = DEFAULT_CALLBACKS.copy() + [ProgressCallback] self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) # Callbacks passed at init are added to the default callbacks snake_case_ : Optional[Any] = self.get_trainer(callbacks=[MyTestTrainerCallback] ) expected_callbacks.append(lowercase_ ) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) # TrainingArguments.disable_tqdm controls if use ProgressCallback or PrinterCallback snake_case_ : Optional[int] = self.get_trainer(disable_tqdm=lowercase_ ) snake_case_ : List[Any] = DEFAULT_CALLBACKS.copy() + [PrinterCallback] self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) def _snake_case ( self : int ): snake_case_ : int = DEFAULT_CALLBACKS.copy() + [ProgressCallback] snake_case_ : List[Any] = self.get_trainer() # We can add, pop, or remove by class name trainer.remove_callback(lowercase_ ) expected_callbacks.remove(lowercase_ ) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) snake_case_ : Dict = self.get_trainer() snake_case_ : Optional[int] = trainer.pop_callback(lowercase_ ) self.assertEqual(cb.__class__ , lowercase_ ) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) trainer.add_callback(lowercase_ ) expected_callbacks.insert(0 , lowercase_ ) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) # We can also add, pop, or remove by instance snake_case_ : Optional[int] = self.get_trainer() snake_case_ : List[Any] = trainer.callback_handler.callbacks[0] trainer.remove_callback(lowercase_ ) expected_callbacks.remove(lowercase_ ) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) snake_case_ : List[Any] = self.get_trainer() snake_case_ : Optional[int] = trainer.callback_handler.callbacks[0] snake_case_ : Optional[Any] = trainer.pop_callback(lowercase_ ) self.assertEqual(lowercase_ , lowercase_ ) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) trainer.add_callback(lowercase_ ) expected_callbacks.insert(0 , lowercase_ ) self.check_callbacks_equality(trainer.callback_handler.callbacks , lowercase_ ) def _snake_case ( self : List[Any] ): import warnings # XXX: for now ignore scatter_gather warnings in this test since it's not relevant to what's being tested warnings.simplefilter(action='''ignore''' , category=lowercase_ ) snake_case_ : int = self.get_trainer(callbacks=[MyTestTrainerCallback] ) trainer.train() snake_case_ : Union[str, Any] = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowercase_ , self.get_expected_events(lowercase_ ) ) # Independent log/save/eval snake_case_ : int = self.get_trainer(callbacks=[MyTestTrainerCallback] , logging_steps=5 ) trainer.train() snake_case_ : str = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowercase_ , self.get_expected_events(lowercase_ ) ) snake_case_ : List[Any] = self.get_trainer(callbacks=[MyTestTrainerCallback] , save_steps=5 ) trainer.train() snake_case_ : int = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowercase_ , self.get_expected_events(lowercase_ ) ) snake_case_ : List[Any] = self.get_trainer(callbacks=[MyTestTrainerCallback] , eval_steps=5 , evaluation_strategy='''steps''' ) trainer.train() snake_case_ : Union[str, Any] = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowercase_ , self.get_expected_events(lowercase_ ) ) snake_case_ : Union[str, Any] = self.get_trainer(callbacks=[MyTestTrainerCallback] , evaluation_strategy='''epoch''' ) trainer.train() snake_case_ : Dict = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowercase_ , self.get_expected_events(lowercase_ ) ) # A bit of everything snake_case_ : str = self.get_trainer( callbacks=[MyTestTrainerCallback] , logging_steps=3 , save_steps=10 , eval_steps=5 , evaluation_strategy='''steps''' , ) trainer.train() snake_case_ : str = trainer.callback_handler.callbacks[-2].events self.assertEqual(lowercase_ , self.get_expected_events(lowercase_ ) ) # warning should be emitted for duplicated callbacks with patch('''transformers.trainer_callback.logger.warning''' ) as warn_mock: snake_case_ : Dict = self.get_trainer( callbacks=[MyTestTrainerCallback, MyTestTrainerCallback] , ) assert str(lowercase_ ) in warn_mock.call_args[0][0]
264
1
"""simple docstring""" def _SCREAMING_SNAKE_CASE (__lowerCAmelCase ) -> bool: '''simple docstring''' return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
362
"""simple docstring""" from __future__ import annotations from typing import Any class SCREAMING_SNAKE_CASE__ : def __init__( self : Any , lowerCAmelCase_ : int = 6): """simple docstring""" lowercase_ = None lowercase_ = None self.create_linked_list(lowerCAmelCase_) def _UpperCAmelCase ( self : List[str] , lowerCAmelCase_ : int): """simple docstring""" lowercase_ = Node() lowercase_ = current_node lowercase_ = current_node lowercase_ = current_node for _ in range(1 , lowerCAmelCase_): lowercase_ = Node() lowercase_ = current_node lowercase_ = previous_node lowercase_ = current_node lowercase_ = self.front lowercase_ = previous_node def _UpperCAmelCase ( self : Union[str, Any]): """simple docstring""" return ( self.front == self.rear and self.front is not None and self.front.data is None ) def _UpperCAmelCase ( self : Optional[Any]): """simple docstring""" self.check_can_perform_operation() return self.front.data if self.front else None def _UpperCAmelCase ( self : int , lowerCAmelCase_ : Any): """simple docstring""" if self.rear is None: return self.check_is_full() if not self.is_empty(): lowercase_ = self.rear.next if self.rear: lowercase_ = data def _UpperCAmelCase ( self : str): """simple docstring""" self.check_can_perform_operation() if self.rear is None or self.front is None: return None if self.front == self.rear: lowercase_ = self.front.data lowercase_ = None return data lowercase_ = self.front lowercase_ = old_front.next lowercase_ = old_front.data lowercase_ = None return data def _UpperCAmelCase ( self : Any): """simple docstring""" if self.is_empty(): raise Exception("""Empty Queue""") def _UpperCAmelCase ( self : Tuple): """simple docstring""" if self.rear and self.rear.next == self.front: raise Exception("""Full Queue""") class SCREAMING_SNAKE_CASE__ : def __init__( self : List[str]): """simple docstring""" lowercase_ = None lowercase_ = None lowercase_ = None if __name__ == "__main__": import doctest doctest.testmod()
313
0
'''simple docstring''' class _lowercase : def __init__( self: Dict , UpperCamelCase__: int , UpperCamelCase__: Dict , UpperCamelCase__: List[Any] ): lowerCamelCase__ : Dict = name lowerCamelCase__ : Union[str, Any] = value lowerCamelCase__ : str = weight def __repr__( self: Dict ): return F'''{self.__class__.__name__}({self.name}, {self.value}, {self.weight})''' def lowerCamelCase_ ( self: int ): return self.value def lowerCamelCase_ ( self: Tuple ): return self.name def lowerCamelCase_ ( self: str ): return self.weight def lowerCamelCase_ ( self: Any ): return self.value / self.weight def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Tuple: lowerCamelCase__ : List[str] = [] for i in range(len(UpperCamelCase ) ): menu.append(Things(name[i] , value[i] , weight[i] ) ) return menu def SCREAMING_SNAKE_CASE_ (UpperCamelCase , UpperCamelCase , UpperCamelCase ) -> Union[str, Any]: lowerCamelCase__ : List[Any] = sorted(UpperCamelCase , key=UpperCamelCase , reverse=UpperCamelCase ) lowerCamelCase__ : Optional[int] = [] lowerCamelCase__ , lowerCamelCase__ : Dict = 0.0, 0.0 for i in range(len(UpperCamelCase ) ): if (total_cost + items_copy[i].get_weight()) <= max_cost: result.append(items_copy[i] ) total_cost += items_copy[i].get_weight() total_value += items_copy[i].get_value() return (result, total_value) def SCREAMING_SNAKE_CASE_ () -> Tuple: pass if __name__ == "__main__": import doctest doctest.testmod()
41
'''simple docstring''' import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class _lowercase ( _lowercase ): def lowerCamelCase_ ( self: Any ): lowerCamelCase__ : str = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(UpperCamelCase__ , """tf_padding""" ) ) self.parent.assertTrue(hasattr(UpperCamelCase__ , """depth_multiplier""" ) ) class _lowercase : def __init__( self: str , UpperCamelCase__: Dict , UpperCamelCase__: Tuple=13 , UpperCamelCase__: Optional[int]=3 , UpperCamelCase__: List[Any]=32 , UpperCamelCase__: Optional[Any]=0.25 , UpperCamelCase__: int=8 , UpperCamelCase__: Any=True , UpperCamelCase__: Dict=1_024 , UpperCamelCase__: Optional[int]=32 , UpperCamelCase__: Tuple="relu6" , UpperCamelCase__: int=0.1 , UpperCamelCase__: List[Any]=0.02 , UpperCamelCase__: Optional[Any]=True , UpperCamelCase__: Union[str, Any]=True , UpperCamelCase__: Union[str, Any]=10 , UpperCamelCase__: str=None , ): lowerCamelCase__ : Optional[Any] = parent lowerCamelCase__ : List[str] = batch_size lowerCamelCase__ : Optional[int] = num_channels lowerCamelCase__ : Optional[int] = image_size lowerCamelCase__ : Optional[Any] = depth_multiplier lowerCamelCase__ : Union[str, Any] = min_depth lowerCamelCase__ : Optional[Any] = tf_padding lowerCamelCase__ : str = int(last_hidden_size * depth_multiplier ) lowerCamelCase__ : Any = output_stride lowerCamelCase__ : int = hidden_act lowerCamelCase__ : Tuple = classifier_dropout_prob lowerCamelCase__ : Dict = use_labels lowerCamelCase__ : Tuple = is_training lowerCamelCase__ : Optional[Any] = num_labels lowerCamelCase__ : Union[str, Any] = initializer_range lowerCamelCase__ : Optional[Any] = scope def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowerCamelCase__ : Optional[Any] = None lowerCamelCase__ : Dict = None if self.use_labels: lowerCamelCase__ : Union[str, Any] = ids_tensor([self.batch_size] , self.num_labels ) lowerCamelCase__ : Dict = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) lowerCamelCase__ : Dict = self.get_config() return config, pixel_values, labels, pixel_labels def lowerCamelCase_ ( self: str ): return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , min_depth=self.min_depth , tf_padding=self.tf_padding , hidden_act=self.hidden_act , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def lowerCamelCase_ ( self: Tuple , UpperCamelCase__: int , UpperCamelCase__: str , UpperCamelCase__: Any , UpperCamelCase__: Optional[int] ): lowerCamelCase__ : List[str] = MobileNetVaModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : List[str] = model(UpperCamelCase__ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def lowerCamelCase_ ( self: str , UpperCamelCase__: Tuple , UpperCamelCase__: Optional[int] , UpperCamelCase__: List[Any] , UpperCamelCase__: Union[str, Any] ): lowerCamelCase__ : List[str] = self.num_labels lowerCamelCase__ : Optional[Any] = MobileNetVaForImageClassification(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() lowerCamelCase__ : List[Any] = model(UpperCamelCase__ , labels=UpperCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase_ ( self: Optional[Any] ): lowerCamelCase__ : str = self.prepare_config_and_inputs() lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ : int = config_and_inputs lowerCamelCase__ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _lowercase ( _lowercase , _lowercase , unittest.TestCase ): a = (MobileNetVaModel, MobileNetVaForImageClassification) if is_torch_available() else () a = ( {"""feature-extraction""": MobileNetVaModel, """image-classification""": MobileNetVaForImageClassification} if is_torch_available() else {} ) a = False a = False a = False a = False def lowerCamelCase_ ( self: List[str] ): lowerCamelCase__ : Optional[int] = MobileNetVaModelTester(self ) lowerCamelCase__ : List[str] = MobileNetVaConfigTester(self , config_class=UpperCamelCase__ , has_text_modality=UpperCamelCase__ ) def lowerCamelCase_ ( self: Union[str, Any] ): self.config_tester.run_common_tests() @unittest.skip(reason="""MobileNetV1 does not use inputs_embeds""" ) def lowerCamelCase_ ( self: Optional[int] ): pass @unittest.skip(reason="""MobileNetV1 does not support input and output embeddings""" ) def lowerCamelCase_ ( self: Optional[Any] ): pass @unittest.skip(reason="""MobileNetV1 does not output attentions""" ) def lowerCamelCase_ ( self: Any ): pass def lowerCamelCase_ ( self: Any ): lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : Dict = model_class(UpperCamelCase__ ) lowerCamelCase__ : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowerCamelCase__ : List[Any] = [*signature.parameters.keys()] lowerCamelCase__ : Dict = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): lowerCamelCase__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) def lowerCamelCase_ ( self: str ): def check_hidden_states_output(UpperCamelCase__: List[Any] , UpperCamelCase__: Dict , UpperCamelCase__: List[Any] ): lowerCamelCase__ : str = model_class(UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() with torch.no_grad(): lowerCamelCase__ : Union[str, Any] = model(**self._prepare_for_class(UpperCamelCase__ , UpperCamelCase__ ) ) lowerCamelCase__ : List[Any] = outputs.hidden_states lowerCamelCase__ : Tuple = 26 self.assertEqual(len(UpperCamelCase__ ) , UpperCamelCase__ ) lowerCamelCase__ , lowerCamelCase__ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowerCamelCase__ : List[Any] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] lowerCamelCase__ : Optional[Any] = True check_hidden_states_output(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*UpperCamelCase__ ) @slow def lowerCamelCase_ ( self: List[str] ): for model_name in MOBILENET_V1_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowerCamelCase__ : Dict = MobileNetVaModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE_ () -> Union[str, Any]: lowerCamelCase__ : Optional[int] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_torch @require_vision class _lowercase ( unittest.TestCase ): @cached_property def lowerCamelCase_ ( self: Optional[int] ): return ( MobileNetVaImageProcessor.from_pretrained("""google/mobilenet_v1_1.0_224""" ) if is_vision_available() else None ) @slow def lowerCamelCase_ ( self: Dict ): lowerCamelCase__ : List[Any] = MobileNetVaForImageClassification.from_pretrained("""google/mobilenet_v1_1.0_224""" ).to(UpperCamelCase__ ) lowerCamelCase__ : Dict = self.default_image_processor lowerCamelCase__ : int = prepare_img() lowerCamelCase__ : List[Any] = image_processor(images=UpperCamelCase__ , return_tensors="""pt""" ).to(UpperCamelCase__ ) # forward pass with torch.no_grad(): lowerCamelCase__ : str = model(**UpperCamelCase__ ) # verify the logits lowerCamelCase__ : List[str] = torch.Size((1, 1_001) ) self.assertEqual(outputs.logits.shape , UpperCamelCase__ ) lowerCamelCase__ : List[str] = torch.tensor([-4.1_739, -1.1_233, 3.1_205] ).to(UpperCamelCase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , UpperCamelCase__ , atol=1e-4 ) )
41
1
from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .multicontrolnet import MultiControlNetModel from .pipeline_controlnet import StableDiffusionControlNetPipeline from .pipeline_controlnet_imgaimg import StableDiffusionControlNetImgaImgPipeline from .pipeline_controlnet_inpaint import StableDiffusionControlNetInpaintPipeline if is_transformers_available() and is_flax_available(): from .pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline
304
from typing import Any import numpy as np def _lowerCamelCase( lowercase__ ) -> bool: '''simple docstring''' return np.array_equal(lowercase__ , matrix.conjugate().T ) def _lowerCamelCase( lowercase__ , lowercase__ ) -> Any: '''simple docstring''' __lowercase= v.conjugate().T __lowercase= v_star.dot(lowercase__ ) assert isinstance(lowercase__ , np.ndarray ) return (v_star_dot.dot(lowercase__ )) / (v_star.dot(lowercase__ )) def _lowerCamelCase( ) -> None: '''simple docstring''' __lowercase= np.array([[2, 2 + 1j, 4], [2 - 1j, 3, 1j], [4, -1j, 1]] ) __lowercase= np.array([[1], [2], [3]] ) assert is_hermitian(lowercase__ ), F'{a} is not hermitian.' print(rayleigh_quotient(lowercase__ , lowercase__ ) ) __lowercase= np.array([[1, 2, 4], [2, 3, -1], [4, -1, 1]] ) assert is_hermitian(lowercase__ ), F'{a} is not hermitian.' assert rayleigh_quotient(lowercase__ , lowercase__ ) == float(3 ) if __name__ == "__main__": import doctest doctest.testmod() tests()
304
1
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class A( unittest.TestCase ): '''simple docstring''' def a__ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" lowerCamelCase_ = 0 def a__ ( self : Optional[Any] ) -> str: """simple docstring""" lowerCamelCase_ = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32' ) self.assertIsInstance(snake_case_ , snake_case_ ) def a__ ( self : Dict ) -> Any: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase_ = Path(snake_case_ ) / """preprocessor_config.json""" lowerCamelCase_ = Path(snake_case_ ) / """config.json""" json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(snake_case_ , 'w' ) , ) json.dump({'model_type': 'clip'} , open(snake_case_ , 'w' ) ) lowerCamelCase_ = AutoImageProcessor.from_pretrained(snake_case_ ) self.assertIsInstance(snake_case_ , snake_case_ ) def a__ ( self : Optional[Any] ) -> List[Any]: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase_ = Path(snake_case_ ) / """preprocessor_config.json""" lowerCamelCase_ = Path(snake_case_ ) / """config.json""" json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(snake_case_ , 'w' ) , ) json.dump({'model_type': 'clip'} , open(snake_case_ , 'w' ) ) lowerCamelCase_ = AutoImageProcessor.from_pretrained(snake_case_ ) self.assertIsInstance(snake_case_ , snake_case_ ) def a__ ( self : Optional[int] ) -> List[str]: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase_ = CLIPConfig() # Create a dummy config file with image_proceesor_type lowerCamelCase_ = Path(snake_case_ ) / """preprocessor_config.json""" lowerCamelCase_ = Path(snake_case_ ) / """config.json""" json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(snake_case_ , 'w' ) , ) json.dump({'model_type': 'clip'} , open(snake_case_ , 'w' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally lowerCamelCase_ = AutoImageProcessor.from_pretrained(snake_case_ ).to_dict() config_dict.pop('image_processor_type' ) lowerCamelCase_ = CLIPImageProcessor(**snake_case_ ) # save in new folder model_config.save_pretrained(snake_case_ ) config.save_pretrained(snake_case_ ) lowerCamelCase_ = AutoImageProcessor.from_pretrained(snake_case_ ) # make sure private variable is not incorrectly saved lowerCamelCase_ = json.loads(config.to_json_string() ) self.assertTrue('_processor_class' not in dict_as_saved ) self.assertIsInstance(snake_case_ , snake_case_ ) def a__ ( self : Any ) -> str: """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase_ = Path(snake_case_ ) / """preprocessor_config.json""" json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(snake_case_ , 'w' ) , ) lowerCamelCase_ = AutoImageProcessor.from_pretrained(snake_case_ ) self.assertIsInstance(snake_case_ , snake_case_ ) def a__ ( self : Optional[int] ) -> List[Any]: """simple docstring""" with self.assertRaisesRegex( snake_case_ , 'clip-base is not a local folder and is not a valid model identifier' ): lowerCamelCase_ = AutoImageProcessor.from_pretrained('clip-base' ) def a__ ( self : Any ) -> int: """simple docstring""" with self.assertRaisesRegex( snake_case_ , r'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): lowerCamelCase_ = AutoImageProcessor.from_pretrained(snake_case_ , revision='aaaaaa' ) def a__ ( self : Any ) -> str: """simple docstring""" with self.assertRaisesRegex( snake_case_ , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): lowerCamelCase_ = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model' ) def a__ ( self : List[str] ) -> List[Any]: """simple docstring""" with self.assertRaises(snake_case_ ): lowerCamelCase_ = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(snake_case_ ): lowerCamelCase_ = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=snake_case_ ) lowerCamelCase_ = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=snake_case_ ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(snake_case_ ) lowerCamelCase_ = AutoImageProcessor.from_pretrained(snake_case_ , trust_remote_code=snake_case_ ) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor' ) def a__ ( self : str ) -> Optional[Any]: """simple docstring""" try: AutoConfig.register('custom' , snake_case_ ) AutoImageProcessor.register(snake_case_ , snake_case_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(snake_case_ ): AutoImageProcessor.register(snake_case_ , snake_case_ ) with tempfile.TemporaryDirectory() as tmpdirname: lowerCamelCase_ = Path(snake_case_ ) / """preprocessor_config.json""" lowerCamelCase_ = Path(snake_case_ ) / """config.json""" json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(snake_case_ , 'w' ) , ) json.dump({'model_type': 'clip'} , open(snake_case_ , 'w' ) ) lowerCamelCase_ = CustomImageProcessor.from_pretrained(snake_case_ ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(snake_case_ ) lowerCamelCase_ = AutoImageProcessor.from_pretrained(snake_case_ ) self.assertIsInstance(snake_case_ , snake_case_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def a__ ( self : Tuple ) -> Dict: """simple docstring""" class A( _a ): '''simple docstring''' UpperCamelCase = True try: AutoConfig.register('custom' , snake_case_ ) AutoImageProcessor.register(snake_case_ , snake_case_ ) # If remote code is not set, the default is to use local lowerCamelCase_ = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. lowerCamelCase_ = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=snake_case_ ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub lowerCamelCase_ = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=snake_case_ ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(not hasattr(snake_case_ , 'is_local' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
204
'''simple docstring''' import string from math import logaa def __snake_case( _lowerCAmelCase , _lowerCAmelCase ) -> int: snake_case__ : List[str] = document.translate( str.maketrans("""""" , """""" , string.punctuation ) ).replace("""\n""" , """""" ) snake_case__ : List[str] = document_without_punctuation.split(""" """ ) # word tokenization return len([word for word in tokenize_document if word.lower() == term.lower()] ) def __snake_case( _lowerCAmelCase , _lowerCAmelCase ) -> tuple[int, int]: snake_case__ : Dict = corpus.lower().translate( str.maketrans("""""" , """""" , string.punctuation ) ) # strip all punctuation and replace it with '' snake_case__ : Any = corpus_without_punctuation.split("""\n""" ) snake_case__ : int = term.lower() return (len([doc for doc in docs if term in doc] ), len(_lowerCAmelCase )) def __snake_case( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=False ) -> float: if smoothing: if n == 0: raise ValueError("""log10(0) is undefined.""" ) return round(1 + logaa(n / (1 + df) ) , 3 ) if df == 0: raise ZeroDivisionError("""df must be > 0""" ) elif n == 0: raise ValueError("""log10(0) is undefined.""" ) return round(logaa(n / df ) , 3 ) def __snake_case( _lowerCAmelCase , _lowerCAmelCase ) -> float: return round(tf * idf , 3 )
35
0
from argparse import ArgumentParser, Namespace from ..utils import logging from . import BaseTransformersCLICommand def snake_case (UpperCAmelCase__ ) -> Optional[int]: return ConvertCommand( args.model_type , args.tf_checkpoint , args.pytorch_dump_output , args.config , args.finetuning_task_name ) A_ : Optional[int] = '\ntransformers can only be used from the commandline to convert TensorFlow models in PyTorch, In that case, it requires\nTensorFlow to be installed. Please see https://www.tensorflow.org/install/ for installation instructions.\n' class _lowerCAmelCase( UpperCAmelCase_ ): """simple docstring""" @staticmethod def _a ( _lowerCamelCase ): UpperCamelCase_: Union[str, Any] = parser.add_parser( 'convert' , help='CLI tool to run convert model from original author checkpoints to Transformers PyTorch checkpoints.' , ) train_parser.add_argument('--model_type' , type=_lowerCamelCase , required=_lowerCamelCase , help='Model\'s type.' ) train_parser.add_argument( '--tf_checkpoint' , type=_lowerCamelCase , required=_lowerCamelCase , help='TensorFlow checkpoint path or folder.' ) train_parser.add_argument( '--pytorch_dump_output' , type=_lowerCamelCase , required=_lowerCamelCase , help='Path to the PyTorch saved model output.' ) train_parser.add_argument('--config' , type=_lowerCamelCase , default='' , help='Configuration file path or folder.' ) train_parser.add_argument( '--finetuning_task_name' , type=_lowerCamelCase , default=_lowerCamelCase , help='Optional fine-tuning task name if the TF model was a finetuned model.' , ) train_parser.set_defaults(func=_lowerCamelCase ) def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , *_lowerCamelCase , ): UpperCamelCase_: int = logging.get_logger('transformers-cli/converting' ) self._logger.info(f'''Loading model {model_type}''' ) UpperCamelCase_: int = model_type UpperCamelCase_: List[Any] = tf_checkpoint UpperCamelCase_: List[str] = pytorch_dump_output UpperCamelCase_: Optional[Any] = config UpperCamelCase_: str = finetuning_task_name def _a ( self ): if self._model_type == "albert": try: from ..models.albert.convert_albert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_lowerCamelCase ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "bert": try: from ..models.bert.convert_bert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_lowerCamelCase ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "funnel": try: from ..models.funnel.convert_funnel_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_lowerCamelCase ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "t5": try: from ..models.ta.convert_ta_original_tf_checkpoint_to_pytorch import convert_tf_checkpoint_to_pytorch except ImportError: raise ImportError(_lowerCamelCase ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "gpt": from ..models.openai.convert_openai_original_tf_checkpoint_to_pytorch import ( convert_openai_checkpoint_to_pytorch, ) convert_openai_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "transfo_xl": try: from ..models.transfo_xl.convert_transfo_xl_original_tf_checkpoint_to_pytorch import ( convert_transfo_xl_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_lowerCamelCase ) if "ckpt" in self._tf_checkpoint.lower(): UpperCamelCase_: Optional[int] = self._tf_checkpoint UpperCamelCase_: str = '' else: UpperCamelCase_: int = self._tf_checkpoint UpperCamelCase_: List[str] = '' convert_transfo_xl_checkpoint_to_pytorch( _lowerCamelCase , self._config , self._pytorch_dump_output , _lowerCamelCase ) elif self._model_type == "gpt2": try: from ..models.gpta.convert_gpta_original_tf_checkpoint_to_pytorch import ( convert_gpta_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_lowerCamelCase ) convert_gpta_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "xlnet": try: from ..models.xlnet.convert_xlnet_original_tf_checkpoint_to_pytorch import ( convert_xlnet_checkpoint_to_pytorch, ) except ImportError: raise ImportError(_lowerCamelCase ) convert_xlnet_checkpoint_to_pytorch( self._tf_checkpoint , self._config , self._pytorch_dump_output , self._finetuning_task_name ) elif self._model_type == "xlm": from ..models.xlm.convert_xlm_original_pytorch_checkpoint_to_pytorch import ( convert_xlm_checkpoint_to_pytorch, ) convert_xlm_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "lxmert": from ..models.lxmert.convert_lxmert_original_tf_checkpoint_to_pytorch import ( convert_lxmert_checkpoint_to_pytorch, ) convert_lxmert_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "rembert": from ..models.rembert.convert_rembert_tf_checkpoint_to_pytorch import ( convert_rembert_tf_checkpoint_to_pytorch, ) convert_rembert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) else: raise ValueError( '--model_type should be selected in the list [bert, gpt, gpt2, t5, transfo_xl, xlnet, xlm, lxmert]' )
292
import numpy as np from sklearn.datasets import fetch_california_housing from sklearn.metrics import mean_absolute_error, mean_squared_error from sklearn.model_selection import train_test_split from xgboost import XGBRegressor def snake_case (UpperCAmelCase__ ) -> tuple: return (data["data"], data["target"]) def snake_case (UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) -> np.ndarray: UpperCamelCase_: Dict = XGBRegressor(verbosity=0 , random_state=4_2 ) xgb.fit(UpperCAmelCase__ , UpperCAmelCase__ ) # Predict target for test data UpperCamelCase_: int = xgb.predict(UpperCAmelCase__ ) UpperCamelCase_: Any = predictions.reshape(len(UpperCAmelCase__ ) , 1 ) return predictions def snake_case () -> None: UpperCamelCase_: Union[str, Any] = fetch_california_housing() UpperCamelCase_ ,UpperCamelCase_: Tuple = data_handling(UpperCAmelCase__ ) UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_ ,UpperCamelCase_: Optional[Any] = train_test_split( UpperCAmelCase__ , UpperCAmelCase__ , test_size=0.25 , random_state=1 ) UpperCamelCase_: Union[str, Any] = xgboost(UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ ) # Error printing print(F'''Mean Absolute Error : {mean_absolute_error(UpperCAmelCase__ , UpperCAmelCase__ )}''' ) print(F'''Mean Square Error : {mean_squared_error(UpperCAmelCase__ , UpperCAmelCase__ )}''' ) if __name__ == "__main__": import doctest doctest.testmod(verbose=True) main()
292
1
import argparse import re import torch from CLAP import create_model from transformers import AutoFeatureExtractor, ClapConfig, ClapModel snake_case_ = { 'text_branch': 'text_model', 'audio_branch': 'audio_model.audio_encoder', 'attn': 'attention.self', 'self.proj': 'output.dense', 'attention.self_mask': 'attn_mask', 'mlp.fc1': 'intermediate.dense', 'mlp.fc2': 'output.dense', 'norm1': 'layernorm_before', 'norm2': 'layernorm_after', 'bn0': 'batch_norm', } snake_case_ = AutoFeatureExtractor.from_pretrained('laion/clap-htsat-unfused', truncation='rand_trunc') def lowerCamelCase__ ( snake_case_ : str , snake_case_ : Any=False ) -> Optional[Any]: __snake_case , __snake_case = create_model( '''HTSAT-tiny''' , '''roberta''' , snake_case_ , precision='''fp32''' , device='''cuda:0''' if torch.cuda.is_available() else '''cpu''' , enable_fusion=snake_case_ , fusion_type='''aff_2d''' if enable_fusion else None , ) return model, model_cfg def lowerCamelCase__ ( snake_case_ : Dict ) -> Any: __snake_case = {} __snake_case = R'''.*sequential.(\d+).*''' __snake_case = R'''.*_projection.(\d+).*''' for key, value in state_dict.items(): # check if any key needs to be modified for key_to_modify, new_key in KEYS_TO_MODIFY_MAPPING.items(): if key_to_modify in key: __snake_case = key.replace(snake_case_ , snake_case_ ) if re.match(snake_case_ , snake_case_ ): # replace sequential layers with list __snake_case = re.match(snake_case_ , snake_case_ ).group(1 ) __snake_case = key.replace(f"""sequential.{sequential_layer}.""" , f"""layers.{int(snake_case_ )//3}.linear.""" ) elif re.match(snake_case_ , snake_case_ ): __snake_case = int(re.match(snake_case_ , snake_case_ ).group(1 ) ) # Because in CLAP they use `nn.Sequential`... __snake_case = 1 if projecton_layer == 0 else 2 __snake_case = key.replace(f"""_projection.{projecton_layer}.""" , f"""_projection.linear{transformers_projection_layer}.""" ) if "audio" and "qkv" in key: # split qkv into query key and value __snake_case = value __snake_case = mixed_qkv.size(0 ) // 3 __snake_case = mixed_qkv[:qkv_dim] __snake_case = mixed_qkv[qkv_dim : qkv_dim * 2] __snake_case = mixed_qkv[qkv_dim * 2 :] __snake_case = query_layer __snake_case = key_layer __snake_case = value_layer else: __snake_case = value return model_state_dict def lowerCamelCase__ ( snake_case_ : str , snake_case_ : Any , snake_case_ : Optional[int] , snake_case_ : Union[str, Any]=False ) -> List[str]: __snake_case , __snake_case = init_clap(snake_case_ , enable_fusion=snake_case_ ) clap_model.eval() __snake_case = clap_model.state_dict() __snake_case = rename_state_dict(snake_case_ ) __snake_case = ClapConfig() __snake_case = enable_fusion __snake_case = ClapModel(snake_case_ ) # ignore the spectrogram embedding layer model.load_state_dict(snake_case_ , strict=snake_case_ ) model.save_pretrained(snake_case_ ) transformers_config.save_pretrained(snake_case_ ) if __name__ == "__main__": snake_case_ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to fairseq checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') parser.add_argument('--enable_fusion', action='store_true', help='Whether to enable fusion or not') snake_case_ = parser.parse_args() convert_clap_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.enable_fusion)
24
def lowerCamelCase__ ( snake_case_ : int ) -> int: if not isinstance(snake_case_ , snake_case_ ) or number < 0: raise ValueError('''Input must be a non-negative integer''' ) __snake_case = 0 while number: # This way we arrive at next set bit (next 1) instead of looping # through each bit and checking for 1s hence the # loop won't run 32 times it will only run the number of `1` times number &= number - 1 count += 1 return count if __name__ == "__main__": import doctest doctest.testmod()
24
1
# This code is adapted from OpenAI's release # https://github.com/openai/human-eval/blob/master/human_eval/execution.py import contextlib import faulthandler import io import multiprocessing import os import platform import signal import tempfile def lowerCAmelCase_ (lowerCAmelCase__: Dict , lowerCAmelCase__: int , lowerCAmelCase__: Optional[int] , lowerCAmelCase__: Tuple ): """simple docstring""" UpperCAmelCase_: str = multiprocessing.Manager() UpperCAmelCase_: Tuple = manager.list() UpperCAmelCase_: str = multiprocessing.Process(target=snake_case__ , args=(check_program, result, timeout) ) p.start() p.join(timeout=timeout + 1 ) if p.is_alive(): p.kill() if not result: result.append("""timed out""" ) return { "task_id": task_id, "passed": result[0] == "passed", "result": result[0], "completion_id": completion_id, } def lowerCAmelCase_ (lowerCAmelCase__: Any , lowerCAmelCase__: Optional[int] , lowerCAmelCase__: Union[str, Any] ): """simple docstring""" with create_tempdir(): # These system calls are needed when cleaning up tempdir. import os import shutil UpperCAmelCase_: Tuple = shutil.rmtree UpperCAmelCase_: str = os.rmdir UpperCAmelCase_: str = os.chdir # Disable functionalities that can make destructive changes to the test. reliability_guard() # Run program. try: UpperCAmelCase_: List[str] = {} with swallow_io(): with time_limit(snake_case__ ): exec(snake_case__ , snake_case__ ) result.append("""passed""" ) except TimeoutException: result.append("""timed out""" ) except BaseException as e: result.append(F'failed: {e}' ) # Needed for cleaning up. UpperCAmelCase_: str = rmtree UpperCAmelCase_: Any = rmdir UpperCAmelCase_: Union[str, Any] = chdir @contextlib.contextmanager def lowerCAmelCase_ (lowerCAmelCase__: Optional[Any] ): """simple docstring""" def signal_handler(lowerCAmelCase__: Tuple , lowerCAmelCase__: Optional[int] ): raise TimeoutException("""Timed out!""" ) signal.setitimer(signal.ITIMER_REAL , snake_case__ ) signal.signal(signal.SIGALRM , snake_case__ ) try: yield finally: signal.setitimer(signal.ITIMER_REAL , 0 ) @contextlib.contextmanager def lowerCAmelCase_ (): """simple docstring""" UpperCAmelCase_: Dict = WriteOnlyStringIO() with contextlib.redirect_stdout(snake_case__ ): with contextlib.redirect_stderr(snake_case__ ): with redirect_stdin(snake_case__ ): yield @contextlib.contextmanager def lowerCAmelCase_ (): """simple docstring""" with tempfile.TemporaryDirectory() as dirname: with chdir(snake_case__ ): yield dirname class _a ( __lowerCAmelCase ): pass class _a ( io.StringIO ): def __snake_case (self, *SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) -> List[str]: raise OSError def __snake_case (self, *SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) -> Tuple: raise OSError def __snake_case (self, *SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) -> str: raise OSError def __snake_case (self, *SCREAMING_SNAKE_CASE_, **SCREAMING_SNAKE_CASE_ ) -> str: return False class _a ( contextlib._RedirectStream ): # type: ignore A = '''stdin''' @contextlib.contextmanager def lowerCAmelCase_ (lowerCAmelCase__: str ): """simple docstring""" if root == ".": yield return UpperCAmelCase_: str = os.getcwd() os.chdir(snake_case__ ) try: yield except BaseException as exc: raise exc finally: os.chdir(snake_case__ ) def lowerCAmelCase_ (lowerCAmelCase__: List[str]=None ): """simple docstring""" if maximum_memory_bytes is not None: import resource resource.setrlimit(resource.RLIMIT_AS , (maximum_memory_bytes, maximum_memory_bytes) ) resource.setrlimit(resource.RLIMIT_DATA , (maximum_memory_bytes, maximum_memory_bytes) ) if not platform.uname().system == "Darwin": resource.setrlimit(resource.RLIMIT_STACK , (maximum_memory_bytes, maximum_memory_bytes) ) faulthandler.disable() import builtins UpperCAmelCase_: Any = None UpperCAmelCase_: Optional[int] = None import os UpperCAmelCase_: Any = """1""" UpperCAmelCase_: Any = None UpperCAmelCase_: List[str] = None UpperCAmelCase_: str = None UpperCAmelCase_: Tuple = None UpperCAmelCase_: List[Any] = None UpperCAmelCase_: List[str] = None UpperCAmelCase_: Optional[Any] = None UpperCAmelCase_: Optional[int] = None UpperCAmelCase_: str = None UpperCAmelCase_: Dict = None UpperCAmelCase_: Dict = None UpperCAmelCase_: Optional[int] = None UpperCAmelCase_: int = None UpperCAmelCase_: Tuple = None UpperCAmelCase_: int = None UpperCAmelCase_: Tuple = None UpperCAmelCase_: Optional[Any] = None UpperCAmelCase_: Optional[int] = None UpperCAmelCase_: Dict = None UpperCAmelCase_: List[Any] = None UpperCAmelCase_: List[Any] = None UpperCAmelCase_: List[Any] = None UpperCAmelCase_: List[str] = None UpperCAmelCase_: Dict = None UpperCAmelCase_: str = None UpperCAmelCase_: Dict = None UpperCAmelCase_: str = None import shutil UpperCAmelCase_: Union[str, Any] = None UpperCAmelCase_: List[str] = None UpperCAmelCase_: Optional[int] = None import subprocess UpperCAmelCase_: Dict = None # type: ignore UpperCAmelCase_: Union[str, Any] = None import sys UpperCAmelCase_: List[str] = None UpperCAmelCase_: Any = None UpperCAmelCase_: List[Any] = None UpperCAmelCase_: int = None UpperCAmelCase_: Any = None
365
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, Pipeline, ZeroShotClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipelines_common import ANY # These 2 model types require different inputs than those of the usual text models. a : Tuple = {'LayoutLMv2Config', 'LayoutLMv3Config'} @is_pipeline_test class _a ( unittest.TestCase ): A = MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING A = TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if model_mapping is not None: A = {config: model for config, model in model_mapping.items() if config.__name__ not in _TO_SKIP} if tf_model_mapping is not None: A = { config: model for config, model in tf_model_mapping.items() if config.__name__ not in _TO_SKIP } def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Dict: UpperCAmelCase_: Dict = ZeroShotClassificationPipeline( model=SCREAMING_SNAKE_CASE_, tokenizer=SCREAMING_SNAKE_CASE_, candidate_labels=["""polics""", """health"""] ) return classifier, ["Who are you voting for in 2020?", "My stomach hurts."] def __snake_case (self, SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) -> Dict: UpperCAmelCase_: Dict = classifier("""Who are you voting for in 2020?""", candidate_labels="""politics""" ) self.assertEqual(SCREAMING_SNAKE_CASE_, {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ )]} ) # No kwarg UpperCAmelCase_: Optional[int] = classifier("""Who are you voting for in 2020?""", ["""politics"""] ) self.assertEqual(SCREAMING_SNAKE_CASE_, {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ )]} ) UpperCAmelCase_: Optional[int] = classifier("""Who are you voting for in 2020?""", candidate_labels=["""politics"""] ) self.assertEqual(SCREAMING_SNAKE_CASE_, {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ )]} ) UpperCAmelCase_: List[Any] = classifier("""Who are you voting for in 2020?""", candidate_labels="""politics, public health""" ) self.assertEqual( SCREAMING_SNAKE_CASE_, {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["""scores"""] ) ), 1.0 ) UpperCAmelCase_: Tuple = classifier("""Who are you voting for in 2020?""", candidate_labels=["""politics""", """public health"""] ) self.assertEqual( SCREAMING_SNAKE_CASE_, {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )]} ) self.assertAlmostEqual(sum(nested_simplify(outputs["""scores"""] ) ), 1.0 ) UpperCAmelCase_: str = classifier( """Who are you voting for in 2020?""", candidate_labels="""politics""", hypothesis_template="""This text is about {}""" ) self.assertEqual(SCREAMING_SNAKE_CASE_, {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ )]} ) # https://github.com/huggingface/transformers/issues/13846 UpperCAmelCase_: Union[str, Any] = classifier(["""I am happy"""], ["""positive""", """negative"""] ) self.assertEqual( SCREAMING_SNAKE_CASE_, [ {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )]} for i in range(1 ) ], ) UpperCAmelCase_: Dict = classifier(["""I am happy""", """I am sad"""], ["""positive""", """negative"""] ) self.assertEqual( SCREAMING_SNAKE_CASE_, [ {"""sequence""": ANY(SCREAMING_SNAKE_CASE_ ), """labels""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )], """scores""": [ANY(SCREAMING_SNAKE_CASE_ ), ANY(SCREAMING_SNAKE_CASE_ )]} for i in range(2 ) ], ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): classifier("""""", candidate_labels="""politics""" ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): classifier(SCREAMING_SNAKE_CASE_, candidate_labels="""politics""" ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): classifier("""Who are you voting for in 2020?""", candidate_labels="""""" ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): classifier("""Who are you voting for in 2020?""", candidate_labels=SCREAMING_SNAKE_CASE_ ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): classifier( """Who are you voting for in 2020?""", candidate_labels="""politics""", hypothesis_template="""Not formatting template""", ) with self.assertRaises(SCREAMING_SNAKE_CASE_ ): classifier( """Who are you voting for in 2020?""", candidate_labels="""politics""", hypothesis_template=SCREAMING_SNAKE_CASE_, ) self.run_entailment_id(SCREAMING_SNAKE_CASE_ ) def __snake_case (self, SCREAMING_SNAKE_CASE_ ) -> Union[str, Any]: UpperCAmelCase_: int = zero_shot_classifier.model.config UpperCAmelCase_: Optional[int] = config.labelaid UpperCAmelCase_: str = zero_shot_classifier.entailment_id UpperCAmelCase_: Union[str, Any] = {"""LABEL_0""": 0, """LABEL_1""": 1, """LABEL_2""": 2} self.assertEqual(zero_shot_classifier.entailment_id, -1 ) UpperCAmelCase_: int = {"""entailment""": 0, """neutral""": 1, """contradiction""": 2} self.assertEqual(zero_shot_classifier.entailment_id, 0 ) UpperCAmelCase_: Dict = {"""ENTAIL""": 0, """NON-ENTAIL""": 1} self.assertEqual(zero_shot_classifier.entailment_id, 0 ) UpperCAmelCase_: Tuple = {"""ENTAIL""": 2, """NEUTRAL""": 1, """CONTR""": 0} self.assertEqual(zero_shot_classifier.entailment_id, 2 ) UpperCAmelCase_: Any = original_labelaid self.assertEqual(SCREAMING_SNAKE_CASE_, zero_shot_classifier.entailment_id ) @require_torch def __snake_case (self ) -> str: UpperCAmelCase_: Any = pipeline( """zero-shot-classification""", model="""sshleifer/tiny-distilbert-base-cased-distilled-squad""", framework="""pt""", ) # There was a regression in 4.10 for this # Adding a test so we don't make the mistake again. # https://github.com/huggingface/transformers/issues/13381#issuecomment-912343499 zero_shot_classifier( """Who are you voting for in 2020?""" * 100, candidate_labels=["""politics""", """public health""", """science"""] ) @require_torch def __snake_case (self ) -> Union[str, Any]: UpperCAmelCase_: str = pipeline( """zero-shot-classification""", model="""sshleifer/tiny-distilbert-base-cased-distilled-squad""", framework="""pt""", ) UpperCAmelCase_: Tuple = zero_shot_classifier( """Who are you voting for in 2020?""", candidate_labels=["""politics""", """public health""", """science"""] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE_ ), { """sequence""": """Who are you voting for in 2020?""", """labels""": ["""science""", """public health""", """politics"""], """scores""": [0.3_3_3, 0.3_3_3, 0.3_3_3], }, ) @require_tf def __snake_case (self ) -> int: UpperCAmelCase_: List[Any] = pipeline( """zero-shot-classification""", model="""sshleifer/tiny-distilbert-base-cased-distilled-squad""", framework="""tf""", ) UpperCAmelCase_: Optional[Any] = zero_shot_classifier( """Who are you voting for in 2020?""", candidate_labels=["""politics""", """public health""", """science"""] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE_ ), { """sequence""": """Who are you voting for in 2020?""", """labels""": ["""science""", """public health""", """politics"""], """scores""": [0.3_3_3, 0.3_3_3, 0.3_3_3], }, ) @slow @require_torch def __snake_case (self ) -> Optional[int]: UpperCAmelCase_: List[Any] = pipeline("""zero-shot-classification""", model="""roberta-large-mnli""", framework="""pt""" ) UpperCAmelCase_: Optional[int] = zero_shot_classifier( """Who are you voting for in 2020?""", candidate_labels=["""politics""", """public health""", """science"""] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE_ ), { """sequence""": """Who are you voting for in 2020?""", """labels""": ["""politics""", """public health""", """science"""], """scores""": [0.9_7_6, 0.0_1_5, 0.0_0_9], }, ) UpperCAmelCase_: Optional[Any] = zero_shot_classifier( """The dominant sequence transduction models are based on complex recurrent or convolutional neural networks""" """ in an encoder-decoder configuration. The best performing models also connect the encoder and decoder""" """ through an attention mechanism. We propose a new simple network architecture, the Transformer, based""" """ solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two""" """ machine translation tasks show these models to be superior in quality while being more parallelizable""" """ and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014""" """ English-to-German translation task, improving over the existing best results, including ensembles by""" """ over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new""" """ single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small""" """ fraction of the training costs of the best models from the literature. We show that the Transformer""" """ generalizes well to other tasks by applying it successfully to English constituency parsing both with""" """ large and limited training data.""", candidate_labels=["""machine learning""", """statistics""", """translation""", """vision"""], multi_label=SCREAMING_SNAKE_CASE_, ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE_ ), { """sequence""": ( """The dominant sequence transduction models are based on complex recurrent or convolutional neural""" """ networks in an encoder-decoder configuration. The best performing models also connect the""" """ encoder and decoder through an attention mechanism. We propose a new simple network""" """ architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence""" """ and convolutions entirely. Experiments on two machine translation tasks show these models to be""" """ superior in quality while being more parallelizable and requiring significantly less time to""" """ train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,""" """ improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014""" """ English-to-French translation task, our model establishes a new single-model state-of-the-art""" """ BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training""" """ costs of the best models from the literature. We show that the Transformer generalizes well to""" """ other tasks by applying it successfully to English constituency parsing both with large and""" """ limited training data.""" ), """labels""": ["""translation""", """machine learning""", """vision""", """statistics"""], """scores""": [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], }, ) @slow @require_tf def __snake_case (self ) -> Optional[int]: UpperCAmelCase_: List[str] = pipeline("""zero-shot-classification""", model="""roberta-large-mnli""", framework="""tf""" ) UpperCAmelCase_: Optional[Any] = zero_shot_classifier( """Who are you voting for in 2020?""", candidate_labels=["""politics""", """public health""", """science"""] ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE_ ), { """sequence""": """Who are you voting for in 2020?""", """labels""": ["""politics""", """public health""", """science"""], """scores""": [0.9_7_6, 0.0_1_5, 0.0_0_9], }, ) UpperCAmelCase_: Any = zero_shot_classifier( """The dominant sequence transduction models are based on complex recurrent or convolutional neural networks""" """ in an encoder-decoder configuration. The best performing models also connect the encoder and decoder""" """ through an attention mechanism. We propose a new simple network architecture, the Transformer, based""" """ solely on attention mechanisms, dispensing with recurrence and convolutions entirely. Experiments on two""" """ machine translation tasks show these models to be superior in quality while being more parallelizable""" """ and requiring significantly less time to train. Our model achieves 28.4 BLEU on the WMT 2014""" """ English-to-German translation task, improving over the existing best results, including ensembles by""" """ over 2 BLEU. On the WMT 2014 English-to-French translation task, our model establishes a new""" """ single-model state-of-the-art BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small""" """ fraction of the training costs of the best models from the literature. We show that the Transformer""" """ generalizes well to other tasks by applying it successfully to English constituency parsing both with""" """ large and limited training data.""", candidate_labels=["""machine learning""", """statistics""", """translation""", """vision"""], multi_label=SCREAMING_SNAKE_CASE_, ) self.assertEqual( nested_simplify(SCREAMING_SNAKE_CASE_ ), { """sequence""": ( """The dominant sequence transduction models are based on complex recurrent or convolutional neural""" """ networks in an encoder-decoder configuration. The best performing models also connect the""" """ encoder and decoder through an attention mechanism. We propose a new simple network""" """ architecture, the Transformer, based solely on attention mechanisms, dispensing with recurrence""" """ and convolutions entirely. Experiments on two machine translation tasks show these models to be""" """ superior in quality while being more parallelizable and requiring significantly less time to""" """ train. Our model achieves 28.4 BLEU on the WMT 2014 English-to-German translation task,""" """ improving over the existing best results, including ensembles by over 2 BLEU. On the WMT 2014""" """ English-to-French translation task, our model establishes a new single-model state-of-the-art""" """ BLEU score of 41.8 after training for 3.5 days on eight GPUs, a small fraction of the training""" """ costs of the best models from the literature. We show that the Transformer generalizes well to""" """ other tasks by applying it successfully to English constituency parsing both with large and""" """ limited training data.""" ), """labels""": ["""translation""", """machine learning""", """vision""", """statistics"""], """scores""": [0.8_1_7, 0.7_1_3, 0.0_1_8, 0.0_1_8], }, )
82
0
import shutil import tempfile import unittest from transformers import ( SPIECE_UNDERLINE, AddedToken, BatchEncoding, NllbTokenizer, NllbTokenizerFast, is_torch_available, ) from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, ) from ...test_tokenization_common import TokenizerTesterMixin snake_case_ = get_tests_dir('fixtures/test_sentencepiece.model') if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right snake_case_ = 256047 snake_case_ = 256145 @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( _UpperCAmelCase , unittest.TestCase ): A_ : Any = NllbTokenizer A_ : Tuple = NllbTokenizerFast A_ : Union[str, Any] = True A_ : List[str] = True A_ : int = {} def a (self : Any ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __snake_case = NllbTokenizer(a__ , keep_accents=a__ ) tokenizer.save_pretrained(self.tmpdirname ) def a (self : List[Any] ): """simple docstring""" __snake_case = NllbTokenizer(a__ , keep_accents=a__ ) __snake_case = tokenizer.tokenize('''This is a test''' ) self.assertListEqual(a__ , ['''▁This''', '''▁is''', '''▁a''', '''▁t''', '''est'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(a__ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __snake_case = tokenizer.tokenize('''I was born in 92000, and this is falsé.''' ) self.assertListEqual( a__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''9''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''é''', '''.''', ] , ) __snake_case = tokenizer.convert_tokens_to_ids(a__ ) self.assertListEqual( a__ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) __snake_case = tokenizer.convert_ids_to_tokens(a__ ) self.assertListEqual( a__ , [ SPIECE_UNDERLINE + '''I''', SPIECE_UNDERLINE + '''was''', SPIECE_UNDERLINE + '''b''', '''or''', '''n''', SPIECE_UNDERLINE + '''in''', SPIECE_UNDERLINE + '''''', '''<unk>''', '''2''', '''0''', '''0''', '''0''', ''',''', SPIECE_UNDERLINE + '''and''', SPIECE_UNDERLINE + '''this''', SPIECE_UNDERLINE + '''is''', SPIECE_UNDERLINE + '''f''', '''al''', '''s''', '''<unk>''', '''.''', ] , ) def a (self : Optional[Any] ): """simple docstring""" __snake_case = (self.rust_tokenizer_class, '''hf-internal-testing/tiny-random-nllb''', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __snake_case = self.rust_tokenizer_class.from_pretrained(a__ , **a__ ) __snake_case = self.tokenizer_class.from_pretrained(a__ , **a__ ) __snake_case = tempfile.mkdtemp() __snake_case = tokenizer_r.save_pretrained(a__ ) __snake_case = tokenizer_p.save_pretrained(a__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) __snake_case = tuple(f for f in tokenizer_r_files if '''tokenizer.json''' not in f ) self.assertSequenceEqual(a__ , a__ ) # Checks everything loads correctly in the same way __snake_case = tokenizer_r.from_pretrained(a__ ) __snake_case = tokenizer_p.from_pretrained(a__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a__ , a__ ) ) shutil.rmtree(a__ ) # Save tokenizer rust, legacy_format=True __snake_case = tempfile.mkdtemp() __snake_case = tokenizer_r.save_pretrained(a__ , legacy_format=a__ ) __snake_case = tokenizer_p.save_pretrained(a__ ) # Checks it save with the same files self.assertSequenceEqual(a__ , a__ ) # Checks everything loads correctly in the same way __snake_case = tokenizer_r.from_pretrained(a__ ) __snake_case = tokenizer_p.from_pretrained(a__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a__ , a__ ) ) shutil.rmtree(a__ ) # Save tokenizer rust, legacy_format=False __snake_case = tempfile.mkdtemp() __snake_case = tokenizer_r.save_pretrained(a__ , legacy_format=a__ ) __snake_case = tokenizer_p.save_pretrained(a__ ) # Checks it saved the tokenizer.json file self.assertTrue(any('''tokenizer.json''' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way __snake_case = tokenizer_r.from_pretrained(a__ ) __snake_case = tokenizer_p.from_pretrained(a__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a__ , a__ ) ) shutil.rmtree(a__ ) @require_torch def a (self : Optional[Any] ): """simple docstring""" if not self.test_seqaseq: return __snake_case = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(f"""{tokenizer.__class__.__name__}""" ): # Longer text that will definitely require truncation. __snake_case = [ ''' UN Chief Says There Is No Military Solution in Syria''', ''' Secretary-General Ban Ki-moon says his response to Russia\'s stepped up military support for''' ''' Syria is that \'there is no military solution\' to the nearly five-year conflict and more weapons''' ''' will only worsen the violence and misery for millions of people.''', ] __snake_case = [ '''Şeful ONU declară că nu există o soluţie militară în Siria''', '''Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al''' ''' Rusiei pentru Siria este că "nu există o soluţie militară" la conflictul de aproape cinci ani şi''' ''' că noi arme nu vor face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.''', ] try: __snake_case = tokenizer.prepare_seqaseq_batch( src_texts=a__ , tgt_texts=a__ , max_length=3 , max_target_length=10 , return_tensors='''pt''' , src_lang='''eng_Latn''' , tgt_lang='''ron_Latn''' , ) except NotImplementedError: return self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.labels.shape[1] , 10 ) # max_target_length will default to max_length if not specified __snake_case = tokenizer.prepare_seqaseq_batch( a__ , tgt_texts=a__ , max_length=3 , return_tensors='''pt''' ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.labels.shape[1] , 3 ) __snake_case = tokenizer.prepare_seqaseq_batch( src_texts=a__ , max_length=3 , max_target_length=10 , return_tensors='''pt''' ) self.assertEqual(batch_encoder_only.input_ids.shape[1] , 3 ) self.assertEqual(batch_encoder_only.attention_mask.shape[1] , 3 ) self.assertNotIn('''decoder_input_ids''' , a__ ) @unittest.skip('''Unfortunately way too slow to build a BPE with SentencePiece.''' ) def a (self : List[str] ): """simple docstring""" pass def a (self : int ): """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): __snake_case = [AddedToken('''<special>''' , lstrip=a__ )] __snake_case = self.rust_tokenizer_class.from_pretrained( a__ , additional_special_tokens=a__ , **a__ ) __snake_case = tokenizer_r.encode('''Hey this is a <special> token''' ) __snake_case = tokenizer_r.encode('''<special>''' , add_special_tokens=a__ )[0] self.assertTrue(special_token_id in r_output ) if self.test_slow_tokenizer: __snake_case = self.rust_tokenizer_class.from_pretrained( a__ , additional_special_tokens=a__ , **a__ , ) __snake_case = self.tokenizer_class.from_pretrained( a__ , additional_special_tokens=a__ , **a__ ) __snake_case = tokenizer_p.encode('''Hey this is a <special> token''' ) __snake_case = tokenizer_cr.encode('''Hey this is a <special> token''' ) self.assertEqual(a__ , a__ ) self.assertEqual(a__ , a__ ) self.assertTrue(special_token_id in p_output ) self.assertTrue(special_token_id in cr_output ) @require_torch @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): A_ : List[Any] = 'facebook/nllb-200-distilled-600M' A_ : List[str] = [ ' UN Chief Says There Is No Military Solution in Syria', ' Secretary-General Ban Ki-moon says his response to Russia\'s stepped up military support for Syria is that "there is no military solution" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.', ] A_ : List[Any] = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', 'Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei' ' pentru Siria este că "nu există o soluţie militară" la conflictul de aproape cinci ani şi că noi arme nu vor' ' face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.', ] A_ : List[Any] = [ 256_047, 16_297, 134_408, 8_165, 248_066, 14_734, 950, 1_135, 105_721, 3_573, 83, 27_352, 108, 49_486, 2, ] @classmethod def a (cls : Tuple ): """simple docstring""" __snake_case = NllbTokenizer.from_pretrained( cls.checkpoint_name , src_lang='''eng_Latn''' , tgt_lang='''ron_Latn''' ) __snake_case = 1 return cls def a (self : Optional[int] ): """simple docstring""" self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ace_Arab'''] , 25_6001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''ace_Latn'''] , 25_6002 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids['''fra_Latn'''] , 25_6057 ) def a (self : Union[str, Any] ): """simple docstring""" __snake_case = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , a__ ) def a (self : Optional[Any] ): """simple docstring""" self.assertIn(a__ , self.tokenizer.all_special_ids ) # fmt: off __snake_case = [RO_CODE, 4254, 9_8068, 11_2923, 3_9072, 3909, 713, 10_2767, 26, 1_7314, 3_5642, 1_4683, 3_3118, 2022, 6_6987, 2, 25_6047] # fmt: on __snake_case = self.tokenizer.decode(a__ , skip_special_tokens=a__ ) __snake_case = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=a__ ) self.assertEqual(a__ , a__ ) self.assertNotIn(self.tokenizer.eos_token , a__ ) def a (self : Optional[int] ): """simple docstring""" __snake_case = ['''this is gunna be a long sentence ''' * 20] assert isinstance(src_text[0] , a__ ) __snake_case = 10 __snake_case = self.tokenizer(a__ , max_length=a__ , truncation=a__ ).input_ids[0] self.assertEqual(ids[-1] , 2 ) self.assertEqual(ids[0] , a__ ) self.assertEqual(len(a__ ) , a__ ) def a (self : Tuple ): """simple docstring""" self.assertListEqual(self.tokenizer.convert_tokens_to_ids(['''<mask>''', '''ar_AR'''] ) , [25_6203, 3] ) def a (self : Any ): """simple docstring""" __snake_case = tempfile.mkdtemp() __snake_case = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(a__ ) __snake_case = NllbTokenizer.from_pretrained(a__ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , a__ ) @require_torch def a (self : str ): """simple docstring""" __snake_case = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=a__ , truncation=a__ , max_length=len(self.expected_src_tokens ) , return_tensors='''pt''' , ) __snake_case = shift_tokens_right( batch['''labels'''] , self.tokenizer.pad_token_id , self.tokenizer.lang_code_to_id['''ron_Latn'''] ) self.assertIsInstance(a__ , a__ ) self.assertEqual((2, 15) , batch.input_ids.shape ) self.assertEqual((2, 15) , batch.attention_mask.shape ) __snake_case = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , a__ ) self.assertEqual(a__ , batch.decoder_input_ids[0, 0] ) # EOS # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) def a (self : Union[str, Any] ): """simple docstring""" __snake_case = self.tokenizer(self.src_text , padding=a__ , truncation=a__ , max_length=3 , return_tensors='''pt''' ) __snake_case = self.tokenizer( text_target=self.tgt_text , padding=a__ , truncation=a__ , max_length=10 , return_tensors='''pt''' ) __snake_case = targets['''input_ids'''] __snake_case = shift_tokens_right( a__ , self.tokenizer.pad_token_id , decoder_start_token_id=self.tokenizer.lang_code_to_id[self.tokenizer.tgt_lang] , ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def a (self : str ): """simple docstring""" __snake_case = self.tokenizer._build_translation_inputs( '''A test''' , return_tensors='''pt''' , src_lang='''eng_Latn''' , tgt_lang='''fra_Latn''' ) self.assertEqual( nested_simplify(a__ ) , { # A, test, EOS, en_XX '''input_ids''': [[25_6047, 70, 7356, 2]], '''attention_mask''': [[1, 1, 1, 1]], # ar_AR '''forced_bos_token_id''': 25_6057, } , ) @require_torch def a (self : Optional[Any] ): """simple docstring""" __snake_case = True __snake_case = self.tokenizer( '''UN Chief says there is no military solution in Syria''' , src_lang='''eng_Latn''' , tgt_lang='''fra_Latn''' ) self.assertEqual( inputs.input_ids , [1_6297, 13_4408, 2_5653, 6370, 248, 254, 10_3929, 9_4995, 108, 4_9486, 2, 25_6047] ) __snake_case = False __snake_case = self.tokenizer( '''UN Chief says there is no military solution in Syria''' , src_lang='''eng_Latn''' , tgt_lang='''fra_Latn''' ) self.assertEqual( inputs.input_ids , [25_6047, 1_6297, 13_4408, 2_5653, 6370, 248, 254, 10_3929, 9_4995, 108, 4_9486, 2] )
24
from math import pi def lowerCamelCase__ ( snake_case_ : int , snake_case_ : int ) -> float: return 2 * pi * radius * (angle / 360) if __name__ == "__main__": print(arc_length(90, 10))
24
1
__snake_case : List[Any] = [0, 2, 4, 6, 8] __snake_case : Dict = [1, 3, 5, 7, 9] def _UpperCamelCase ( UpperCamelCase_ : int , UpperCamelCase_ : int , UpperCamelCase_ : list[int] , UpperCamelCase_ : int ) -> int: """simple docstring""" if remaining_length == 0: if digits[0] == 0 or digits[-1] == 0: return 0 for i in range(length // 2 - 1 , -1 , -1 ): remainder += digits[i] + digits[length - i - 1] if remainder % 2 == 0: return 0 remainder //= 10 return 1 if remaining_length == 1: if remainder % 2 == 0: return 0 lowerCAmelCase__ = 0 for digit in range(10 ): lowerCAmelCase__ = digit result += reversible_numbers( 0 , (remainder + 2 * digit) // 10 , UpperCamelCase_ , UpperCamelCase_ ) return result lowerCAmelCase__ = 0 for digita in range(10 ): lowerCAmelCase__ = digita if (remainder + digita) % 2 == 0: lowerCAmelCase__ = ODD_DIGITS else: lowerCAmelCase__ = EVEN_DIGITS for digita in other_parity_digits: lowerCAmelCase__ = digita result += reversible_numbers( remaining_length - 2 , (remainder + digita + digita) // 10 , UpperCamelCase_ , UpperCamelCase_ , ) return result def _UpperCamelCase ( UpperCamelCase_ : int = 9 ) -> int: """simple docstring""" lowerCAmelCase__ = 0 for length in range(1 , max_power + 1 ): result += reversible_numbers(UpperCamelCase_ , 0 , [0] * length , UpperCamelCase_ ) return result if __name__ == "__main__": print(f'{solution() = }')
122
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging __snake_case : Any = logging.get_logger(__name__) __snake_case : Union[str, Any] = {"""vocab_file""": """vocab.json""", """merges_file""": """merges.txt"""} # See all LED models at https://huggingface.co/models?filter=LED __snake_case : Optional[Any] = { """vocab_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json""", }, """merges_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt""", }, """tokenizer_file""": { """allenai/led-base-16384""": """https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json""", }, } __snake_case : List[str] = { """allenai/led-base-16384""": 1_63_84, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def _UpperCamelCase ( ) -> int: """simple docstring""" lowerCAmelCase__ = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) lowerCAmelCase__ = bs[:] lowerCAmelCase__ = 0 for b in range(2**8 ): if b not in bs: bs.append(UpperCamelCase_ ) cs.append(2**8 + n ) n += 1 lowerCAmelCase__ = [chr(UpperCamelCase_ ) for n in cs] return dict(zip(UpperCamelCase_ , UpperCamelCase_ ) ) def _UpperCamelCase ( UpperCamelCase_ : Optional[Any] ) -> Dict: """simple docstring""" lowerCAmelCase__ = set() lowerCAmelCase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowerCAmelCase__ = char return pairs class __SCREAMING_SNAKE_CASE ( __lowercase): _SCREAMING_SNAKE_CASE : Optional[int] = VOCAB_FILES_NAMES _SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_VOCAB_FILES_MAP _SCREAMING_SNAKE_CASE : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _SCREAMING_SNAKE_CASE : Any = ['''input_ids''', '''attention_mask'''] def __init__( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase="replace" , _UpperCamelCase="<s>" , _UpperCamelCase="</s>" , _UpperCamelCase="</s>" , _UpperCamelCase="<s>" , _UpperCamelCase="<unk>" , _UpperCamelCase="<pad>" , _UpperCamelCase="<mask>" , _UpperCamelCase=False , **_UpperCamelCase , ): """simple docstring""" lowerCAmelCase__ = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else bos_token lowerCAmelCase__ = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else eos_token lowerCAmelCase__ = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else sep_token lowerCAmelCase__ = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else cls_token lowerCAmelCase__ = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else unk_token lowerCAmelCase__ = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowerCAmelCase__ = AddedToken(_UpperCamelCase , lstrip=_UpperCamelCase , rstrip=_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else mask_token super().__init__( errors=_UpperCamelCase , bos_token=_UpperCamelCase , eos_token=_UpperCamelCase , unk_token=_UpperCamelCase , sep_token=_UpperCamelCase , cls_token=_UpperCamelCase , pad_token=_UpperCamelCase , mask_token=_UpperCamelCase , add_prefix_space=_UpperCamelCase , **_UpperCamelCase , ) with open(_UpperCamelCase , encoding='utf-8' ) as vocab_handle: lowerCAmelCase__ = json.load(_UpperCamelCase ) lowerCAmelCase__ = {v: k for k, v in self.encoder.items()} lowerCAmelCase__ = errors # how to handle errors in decoding lowerCAmelCase__ = bytes_to_unicode() lowerCAmelCase__ = {v: k for k, v in self.byte_encoder.items()} with open(_UpperCamelCase , encoding='utf-8' ) as merges_handle: lowerCAmelCase__ = merges_handle.read().split('\n' )[1:-1] lowerCAmelCase__ = [tuple(merge.split() ) for merge in bpe_merges] lowerCAmelCase__ = dict(zip(_UpperCamelCase , range(len(_UpperCamelCase ) ) ) ) lowerCAmelCase__ = {} lowerCAmelCase__ = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowerCAmelCase__ = re.compile(r'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def UpperCamelCase__ ( self ): """simple docstring""" return len(self.encoder ) def UpperCamelCase__ ( self ): """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def UpperCamelCase__ ( self , _UpperCamelCase ): """simple docstring""" if token in self.cache: return self.cache[token] lowerCAmelCase__ = tuple(_UpperCamelCase ) lowerCAmelCase__ = get_pairs(_UpperCamelCase ) if not pairs: return token while True: lowerCAmelCase__ = min(_UpperCamelCase , key=lambda _UpperCamelCase : self.bpe_ranks.get(_UpperCamelCase , float('inf' ) ) ) if bigram not in self.bpe_ranks: break lowerCAmelCase__ , lowerCAmelCase__ = bigram lowerCAmelCase__ = [] lowerCAmelCase__ = 0 while i < len(_UpperCamelCase ): try: lowerCAmelCase__ = word.index(_UpperCamelCase , _UpperCamelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowerCAmelCase__ = j if word[i] == first and i < len(_UpperCamelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowerCAmelCase__ = tuple(_UpperCamelCase ) lowerCAmelCase__ = new_word if len(_UpperCamelCase ) == 1: break else: lowerCAmelCase__ = get_pairs(_UpperCamelCase ) lowerCAmelCase__ = ' '.join(_UpperCamelCase ) lowerCAmelCase__ = word return word def UpperCamelCase__ ( self , _UpperCamelCase ): """simple docstring""" lowerCAmelCase__ = [] for token in re.findall(self.pat , _UpperCamelCase ): lowerCAmelCase__ = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(_UpperCamelCase ).split(' ' ) ) return bpe_tokens def UpperCamelCase__ ( self , _UpperCamelCase ): """simple docstring""" return self.encoder.get(_UpperCamelCase , self.encoder.get(self.unk_token ) ) def UpperCamelCase__ ( self , _UpperCamelCase ): """simple docstring""" return self.decoder.get(_UpperCamelCase ) def UpperCamelCase__ ( self , _UpperCamelCase ): """simple docstring""" lowerCAmelCase__ = ''.join(_UpperCamelCase ) lowerCAmelCase__ = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' , errors=self.errors ) return text def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase = None ): """simple docstring""" if not os.path.isdir(_UpperCamelCase ): logger.error(F"Vocabulary path ({save_directory}) should be a directory" ) return lowerCAmelCase__ = os.path.join( _UpperCamelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowerCAmelCase__ = os.path.join( _UpperCamelCase , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(_UpperCamelCase , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_UpperCamelCase , ensure_ascii=_UpperCamelCase ) + '\n' ) lowerCAmelCase__ = 0 with open(_UpperCamelCase , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _UpperCamelCase : kv[1] ): if index != token_index: logger.warning( F"Saving vocabulary to {merge_file}: BPE merge indices are not consecutive." ' Please check that the tokenizer is not corrupted!' ) lowerCAmelCase__ = token_index writer.write(' '.join(_UpperCamelCase ) + '\n' ) index += 1 return vocab_file, merge_file def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowerCAmelCase__ = [self.cls_token_id] lowerCAmelCase__ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase = None , _UpperCamelCase = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_UpperCamelCase , token_ids_a=_UpperCamelCase , already_has_special_tokens=_UpperCamelCase ) if token_ids_a is None: return [1] + ([0] * len(_UpperCamelCase )) + [1] return [1] + ([0] * len(_UpperCamelCase )) + [1, 1] + ([0] * len(_UpperCamelCase )) + [1] def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase = None ): """simple docstring""" lowerCAmelCase__ = [self.sep_token_id] lowerCAmelCase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase=False , **_UpperCamelCase ): """simple docstring""" lowerCAmelCase__ = kwargs.pop('add_prefix_space' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(_UpperCamelCase ) > 0 and not text[0].isspace()): lowerCAmelCase__ = ' ' + text return (text, kwargs) def UpperCamelCase__ ( self , _UpperCamelCase , _UpperCamelCase = None , _UpperCamelCase = PaddingStrategy.DO_NOT_PAD , _UpperCamelCase = None , _UpperCamelCase = None , ): """simple docstring""" lowerCAmelCase__ = super()._pad( encoded_inputs=_UpperCamelCase , max_length=_UpperCamelCase , padding_strategy=_UpperCamelCase , pad_to_multiple_of=_UpperCamelCase , return_attention_mask=_UpperCamelCase , ) # Load from model defaults if return_attention_mask is None: lowerCAmelCase__ = 'attention_mask' in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowerCAmelCase__ = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowerCAmelCase__ = len(encoded_inputs['global_attention_mask'] ) != len(_UpperCamelCase ) if needs_to_be_padded: lowerCAmelCase__ = len(_UpperCamelCase ) - len(encoded_inputs['global_attention_mask'] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowerCAmelCase__ = ( encoded_inputs['global_attention_mask'] + [-1] * difference ) elif self.padding_side == "left": lowerCAmelCase__ = [-1] * difference + encoded_inputs[ 'global_attention_mask' ] else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return encoded_inputs
122
1
'''simple docstring''' from __future__ import annotations class __magic_name__ : def __init__( self : List[Any] , lowercase_ : int ): lowercase_ : str = data lowercase_ : Node | None = None lowercase_ : Node | None = None def lowerCamelCase ( UpperCAmelCase__ : Node | None ) -> None: # In Order traversal of the tree if tree: display(tree.left ) print(tree.data ) display(tree.right ) def lowerCamelCase ( UpperCAmelCase__ : Node | None ) -> int: return 1 + max(depth_of_tree(tree.left ) , depth_of_tree(tree.right ) ) if tree else 0 def lowerCamelCase ( UpperCAmelCase__ : Node ) -> bool: if not tree: return True if tree.left and tree.right: return is_full_binary_tree(tree.left ) and is_full_binary_tree(tree.right ) else: return not tree.left and not tree.right def lowerCamelCase ( ) -> None: # Main function for testing. lowercase_ : Optional[int] = Node(1 ) lowercase_ : Any = Node(2 ) lowercase_ : List[Any] = Node(3 ) lowercase_ : int = Node(4 ) lowercase_ : str = Node(5 ) lowercase_ : Any = Node(6 ) lowercase_ : Tuple = Node(7 ) lowercase_ : Optional[Any] = Node(8 ) lowercase_ : Tuple = Node(9 ) print(is_full_binary_tree(UpperCAmelCase__ ) ) print(depth_of_tree(UpperCAmelCase__ ) ) print("""Tree is: """ ) display(UpperCAmelCase__ ) if __name__ == "__main__": main()
239
'''simple docstring''' import argparse import json import os import sys import tempfile import unittest from argparse import Namespace from dataclasses import dataclass, field from enum import Enum from pathlib import Path from typing import List, Literal, Optional import yaml from transformers import HfArgumentParser, TrainingArguments from transformers.hf_argparser import make_choice_type_function, string_to_bool # Since Python 3.10, we can use the builtin `|` operator for Union types # See PEP 604: https://peps.python.org/pep-0604 _lowercase : int = sys.version_info >= (3, 10) def lowerCamelCase ( UpperCAmelCase__ : Tuple=None , UpperCAmelCase__ : Union[str, Any]=None ) -> Tuple: return field(default_factory=lambda: default , metadata=UpperCAmelCase__ ) @dataclass class __magic_name__ : UpperCamelCase__ = 42 UpperCamelCase__ = 42 UpperCamelCase__ = 42 UpperCamelCase__ = 42 @dataclass class __magic_name__ : UpperCamelCase__ = 42 UpperCamelCase__ = field(default='''toto''', metadata={'''help''': '''help message'''}) @dataclass class __magic_name__ : UpperCamelCase__ = False UpperCamelCase__ = True UpperCamelCase__ = None class __magic_name__ ( _UpperCAmelCase): UpperCamelCase__ = '''titi''' UpperCamelCase__ = '''toto''' class __magic_name__ ( _UpperCAmelCase): UpperCamelCase__ = '''titi''' UpperCamelCase__ = '''toto''' UpperCamelCase__ = 42 @dataclass class __magic_name__ : UpperCamelCase__ = "toto" def SCREAMING_SNAKE_CASE_ ( self : int ): lowercase_ : Optional[int] = BasicEnum(self.foo ) @dataclass class __magic_name__ : UpperCamelCase__ = "toto" def SCREAMING_SNAKE_CASE_ ( self : int ): lowercase_ : Optional[int] = MixedTypeEnum(self.foo ) @dataclass class __magic_name__ : UpperCamelCase__ = None UpperCamelCase__ = field(default=_UpperCAmelCase, metadata={'''help''': '''help message'''}) UpperCamelCase__ = None UpperCamelCase__ = list_field(default=[]) UpperCamelCase__ = list_field(default=[]) @dataclass class __magic_name__ : UpperCamelCase__ = list_field(default=[]) UpperCamelCase__ = list_field(default=[1, 2, 3]) UpperCamelCase__ = list_field(default=['''Hallo''', '''Bonjour''', '''Hello''']) UpperCamelCase__ = list_field(default=[0.1, 0.2, 0.3]) @dataclass class __magic_name__ : UpperCamelCase__ = field() UpperCamelCase__ = field() UpperCamelCase__ = field() def SCREAMING_SNAKE_CASE_ ( self : List[str] ): lowercase_ : List[Any] = BasicEnum(self.required_enum ) @dataclass class __magic_name__ : UpperCamelCase__ = 42 UpperCamelCase__ = field() UpperCamelCase__ = None UpperCamelCase__ = field(default='''toto''', metadata={'''help''': '''help message'''}) UpperCamelCase__ = list_field(default=['''Hallo''', '''Bonjour''', '''Hello''']) if is_python_no_less_than_3_10: @dataclass class __magic_name__ : UpperCamelCase__ = False UpperCamelCase__ = True UpperCamelCase__ = None @dataclass class __magic_name__ : UpperCamelCase__ = None UpperCamelCase__ = field(default=_UpperCAmelCase, metadata={'''help''': '''help message'''}) UpperCamelCase__ = None UpperCamelCase__ = list_field(default=[]) UpperCamelCase__ = list_field(default=[]) class __magic_name__ ( unittest.TestCase): def SCREAMING_SNAKE_CASE_ ( self : Dict , lowercase_ : argparse.ArgumentParser , lowercase_ : argparse.ArgumentParser ): self.assertEqual(len(a._actions ) , len(b._actions ) ) for x, y in zip(a._actions , b._actions ): lowercase_ : List[str] = {k: v for k, v in vars(lowercase_ ).items() if k != """container"""} lowercase_ : Any = {k: v for k, v in vars(lowercase_ ).items() if k != """container"""} # Choices with mixed type have custom function as "type" # So we need to compare results directly for equality if xx.get("""choices""" , lowercase_ ) and yy.get("""choices""" , lowercase_ ): for expected_choice in yy["choices"] + xx["choices"]: self.assertEqual(xx["""type"""](lowercase_ ) , yy["""type"""](lowercase_ ) ) del xx["type"], yy["type"] self.assertEqual(lowercase_ , lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ): lowercase_ : List[str] = HfArgumentParser(lowercase_ ) lowercase_ : str = argparse.ArgumentParser() expected.add_argument("""--foo""" , type=lowercase_ , required=lowercase_ ) expected.add_argument("""--bar""" , type=lowercase_ , required=lowercase_ ) expected.add_argument("""--baz""" , type=lowercase_ , required=lowercase_ ) expected.add_argument("""--flag""" , type=lowercase_ , default=lowercase_ , const=lowercase_ , nargs="""?""" ) self.argparsersEqual(lowercase_ , lowercase_ ) lowercase_ : List[str] = ["""--foo""", """1""", """--baz""", """quux""", """--bar""", """0.5"""] ((lowercase_) , ) : Optional[Any] = parser.parse_args_into_dataclasses(lowercase_ , look_for_args_file=lowercase_ ) self.assertFalse(example.flag ) def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): lowercase_ : Any = HfArgumentParser(lowercase_ ) lowercase_ : str = argparse.ArgumentParser() expected.add_argument("""--foo""" , default=42 , type=lowercase_ ) expected.add_argument("""--baz""" , default="""toto""" , type=lowercase_ , help="""help message""" ) self.argparsersEqual(lowercase_ , lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : int ): lowercase_ : Optional[int] = argparse.ArgumentParser() expected.add_argument("""--foo""" , type=lowercase_ , default=lowercase_ , const=lowercase_ , nargs="""?""" ) expected.add_argument("""--baz""" , type=lowercase_ , default=lowercase_ , const=lowercase_ , nargs="""?""" ) # A boolean no_* argument always has to come after its "default: True" regular counter-part # and its default must be set to False expected.add_argument("""--no_baz""" , action="""store_false""" , default=lowercase_ , dest="""baz""" ) expected.add_argument("""--opt""" , type=lowercase_ , default=lowercase_ ) lowercase_ : List[Any] = [WithDefaultBoolExample] if is_python_no_less_than_3_10: dataclass_types.append(lowercase_ ) for dataclass_type in dataclass_types: lowercase_ : List[str] = HfArgumentParser(lowercase_ ) self.argparsersEqual(lowercase_ , lowercase_ ) lowercase_ : str = parser.parse_args([] ) self.assertEqual(lowercase_ , Namespace(foo=lowercase_ , baz=lowercase_ , opt=lowercase_ ) ) lowercase_ : Optional[Any] = parser.parse_args(["""--foo""", """--no_baz"""] ) self.assertEqual(lowercase_ , Namespace(foo=lowercase_ , baz=lowercase_ , opt=lowercase_ ) ) lowercase_ : Optional[Any] = parser.parse_args(["""--foo""", """--baz"""] ) self.assertEqual(lowercase_ , Namespace(foo=lowercase_ , baz=lowercase_ , opt=lowercase_ ) ) lowercase_ : Dict = parser.parse_args(["""--foo""", """True""", """--baz""", """True""", """--opt""", """True"""] ) self.assertEqual(lowercase_ , Namespace(foo=lowercase_ , baz=lowercase_ , opt=lowercase_ ) ) lowercase_ : int = parser.parse_args(["""--foo""", """False""", """--baz""", """False""", """--opt""", """False"""] ) self.assertEqual(lowercase_ , Namespace(foo=lowercase_ , baz=lowercase_ , opt=lowercase_ ) ) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): lowercase_ : str = HfArgumentParser(lowercase_ ) lowercase_ : Tuple = argparse.ArgumentParser() expected.add_argument( """--foo""" , default="""toto""" , choices=["""titi""", """toto""", 42] , type=make_choice_type_function(["""titi""", """toto""", 42] ) , ) self.argparsersEqual(lowercase_ , lowercase_ ) lowercase_ : Optional[int] = parser.parse_args([] ) self.assertEqual(args.foo , """toto""" ) lowercase_ : Union[str, Any] = parser.parse_args_into_dataclasses([] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.toto ) lowercase_ : int = parser.parse_args(["""--foo""", """titi"""] ) self.assertEqual(args.foo , """titi""" ) lowercase_ : int = parser.parse_args_into_dataclasses(["""--foo""", """titi"""] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.titi ) lowercase_ : List[str] = parser.parse_args(["""--foo""", """42"""] ) self.assertEqual(args.foo , 42 ) lowercase_ : List[str] = parser.parse_args_into_dataclasses(["""--foo""", """42"""] )[0] self.assertEqual(enum_ex.foo , MixedTypeEnum.fourtytwo ) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): @dataclass class __magic_name__ : UpperCamelCase__ = "toto" lowercase_ : Optional[int] = HfArgumentParser(lowercase_ ) lowercase_ : List[str] = argparse.ArgumentParser() expected.add_argument( """--foo""" , default="""toto""" , choices=("""titi""", """toto""", 42) , type=make_choice_type_function(["""titi""", """toto""", 42] ) , ) self.argparsersEqual(lowercase_ , lowercase_ ) lowercase_ : List[str] = parser.parse_args([] ) self.assertEqual(args.foo , """toto""" ) lowercase_ : Optional[int] = parser.parse_args(["""--foo""", """titi"""] ) self.assertEqual(args.foo , """titi""" ) lowercase_ : List[Any] = parser.parse_args(["""--foo""", """42"""] ) self.assertEqual(args.foo , 42 ) def SCREAMING_SNAKE_CASE_ ( self : Dict ): lowercase_ : int = HfArgumentParser(lowercase_ ) lowercase_ : Dict = argparse.ArgumentParser() expected.add_argument("""--foo_int""" , nargs="""+""" , default=[] , type=lowercase_ ) expected.add_argument("""--bar_int""" , nargs="""+""" , default=[1, 2, 3] , type=lowercase_ ) expected.add_argument("""--foo_str""" , nargs="""+""" , default=["""Hallo""", """Bonjour""", """Hello"""] , type=lowercase_ ) expected.add_argument("""--foo_float""" , nargs="""+""" , default=[0.1, 0.2, 0.3] , type=lowercase_ ) self.argparsersEqual(lowercase_ , lowercase_ ) lowercase_ : int = parser.parse_args([] ) self.assertEqual( lowercase_ , Namespace(foo_int=[] , bar_int=[1, 2, 3] , foo_str=["""Hallo""", """Bonjour""", """Hello"""] , foo_float=[0.1, 0.2, 0.3] ) , ) lowercase_ : Optional[int] = parser.parse_args("""--foo_int 1 --bar_int 2 3 --foo_str a b c --foo_float 0.1 0.7""".split() ) self.assertEqual(lowercase_ , Namespace(foo_int=[1] , bar_int=[2, 3] , foo_str=["""a""", """b""", """c"""] , foo_float=[0.1, 0.7] ) ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ): lowercase_ : Tuple = argparse.ArgumentParser() expected.add_argument("""--foo""" , default=lowercase_ , type=lowercase_ ) expected.add_argument("""--bar""" , default=lowercase_ , type=lowercase_ , help="""help message""" ) expected.add_argument("""--baz""" , default=lowercase_ , type=lowercase_ ) expected.add_argument("""--ces""" , nargs="""+""" , default=[] , type=lowercase_ ) expected.add_argument("""--des""" , nargs="""+""" , default=[] , type=lowercase_ ) lowercase_ : Optional[int] = [OptionalExample] if is_python_no_less_than_3_10: dataclass_types.append(lowercase_ ) for dataclass_type in dataclass_types: lowercase_ : Tuple = HfArgumentParser(lowercase_ ) self.argparsersEqual(lowercase_ , lowercase_ ) lowercase_ : Union[str, Any] = parser.parse_args([] ) self.assertEqual(lowercase_ , Namespace(foo=lowercase_ , bar=lowercase_ , baz=lowercase_ , ces=[] , des=[] ) ) lowercase_ : List[Any] = parser.parse_args("""--foo 12 --bar 3.14 --baz 42 --ces a b c --des 1 2 3""".split() ) self.assertEqual(lowercase_ , Namespace(foo=12 , bar=3.14 , baz="""42""" , ces=["""a""", """b""", """c"""] , des=[1, 2, 3] ) ) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): lowercase_ : Dict = HfArgumentParser(lowercase_ ) lowercase_ : int = argparse.ArgumentParser() expected.add_argument("""--required_list""" , nargs="""+""" , type=lowercase_ , required=lowercase_ ) expected.add_argument("""--required_str""" , type=lowercase_ , required=lowercase_ ) expected.add_argument( """--required_enum""" , type=make_choice_type_function(["""titi""", """toto"""] ) , choices=["""titi""", """toto"""] , required=lowercase_ , ) self.argparsersEqual(lowercase_ , lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] ): lowercase_ : Dict = HfArgumentParser(lowercase_ ) lowercase_ : List[str] = argparse.ArgumentParser() expected.add_argument("""--foo""" , type=lowercase_ , required=lowercase_ ) expected.add_argument( """--required_enum""" , type=make_choice_type_function(["""titi""", """toto"""] ) , choices=["""titi""", """toto"""] , required=lowercase_ , ) expected.add_argument("""--opt""" , type=lowercase_ , default=lowercase_ ) expected.add_argument("""--baz""" , default="""toto""" , type=lowercase_ , help="""help message""" ) expected.add_argument("""--foo_str""" , nargs="""+""" , default=["""Hallo""", """Bonjour""", """Hello"""] , type=lowercase_ ) self.argparsersEqual(lowercase_ , lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): lowercase_ : str = HfArgumentParser(lowercase_ ) lowercase_ : Optional[int] = { """foo""": 12, """bar""": 3.14, """baz""": """42""", """flag""": True, } lowercase_ : Optional[Any] = parser.parse_dict(lowercase_ )[0] lowercase_ : Dict = BasicExample(**lowercase_ ) self.assertEqual(lowercase_ , lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] ): lowercase_ : Dict = HfArgumentParser(lowercase_ ) lowercase_ : Optional[int] = { """foo""": 12, """bar""": 3.14, """baz""": """42""", """flag""": True, """extra""": 42, } self.assertRaises(lowercase_ , parser.parse_dict , lowercase_ , allow_extra_keys=lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : Any ): lowercase_ : List[Any] = HfArgumentParser(lowercase_ ) lowercase_ : int = { """foo""": 12, """bar""": 3.14, """baz""": """42""", """flag""": True, } with tempfile.TemporaryDirectory() as tmp_dir: lowercase_ : Optional[int] = os.path.join(lowercase_ , """temp_json""" ) os.mkdir(lowercase_ ) with open(temp_local_path + """.json""" , """w+""" ) as f: json.dump(lowercase_ , lowercase_ ) lowercase_ : str = parser.parse_yaml_file(Path(temp_local_path + """.json""" ) )[0] lowercase_ : Tuple = BasicExample(**lowercase_ ) self.assertEqual(lowercase_ , lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): lowercase_ : str = HfArgumentParser(lowercase_ ) lowercase_ : List[Any] = { """foo""": 12, """bar""": 3.14, """baz""": """42""", """flag""": True, } with tempfile.TemporaryDirectory() as tmp_dir: lowercase_ : Tuple = os.path.join(lowercase_ , """temp_yaml""" ) os.mkdir(lowercase_ ) with open(temp_local_path + """.yaml""" , """w+""" ) as f: yaml.dump(lowercase_ , lowercase_ ) lowercase_ : List[str] = parser.parse_yaml_file(Path(temp_local_path + """.yaml""" ) )[0] lowercase_ : Any = BasicExample(**lowercase_ ) self.assertEqual(lowercase_ , lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): lowercase_ : str = HfArgumentParser(lowercase_ ) self.assertIsNotNone(lowercase_ )
239
1
"""simple docstring""" import json import os import sys import tempfile import unittest from pathlib import Path from shutil import copyfile from huggingface_hub import HfFolder, Repository, create_repo, delete_repo from requests.exceptions import HTTPError import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, PROCESSOR_MAPPING, TOKENIZER_MAPPING, AutoConfig, AutoFeatureExtractor, AutoProcessor, AutoTokenizer, BertTokenizer, ProcessorMixin, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaProcessor, ) from transformers.testing_utils import TOKEN, USER, get_tests_dir, is_staging_test from transformers.tokenization_utils import TOKENIZER_CONFIG_FILE from transformers.utils import FEATURE_EXTRACTOR_NAME, is_tokenizers_available sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 from test_module.custom_processing import CustomProcessor # noqa E402 from test_module.custom_tokenization import CustomTokenizer # noqa E402 snake_case__ : int = get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') snake_case__ : Optional[int] = get_tests_dir('''fixtures/vocab.json''') snake_case__ : Optional[Any] = get_tests_dir('''fixtures''') class snake_case_( unittest.TestCase ): __UpperCamelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''bla''', '''blou'''] def lowerCamelCase__ ( self : Dict ): lowerCAmelCase : Tuple = 0 def lowerCamelCase__ ( self : List[str] ): lowerCAmelCase : Union[str, Any] = AutoProcessor.from_pretrained('''facebook/wav2vec2-base-960h''' ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) def lowerCamelCase__ ( self : Union[str, Any] ): with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase : Dict = WavaVecaConfig() lowerCAmelCase : Tuple = AutoProcessor.from_pretrained('''facebook/wav2vec2-base-960h''' ) # save in new folder model_config.save_pretrained(UpperCamelCase_ ) processor.save_pretrained(UpperCamelCase_ ) lowerCAmelCase : Optional[Any] = AutoProcessor.from_pretrained(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) def lowerCamelCase__ ( self : Tuple ): with tempfile.TemporaryDirectory() as tmpdirname: # copy relevant files copyfile(UpperCamelCase_ , os.path.join(UpperCamelCase_ , UpperCamelCase_ ) ) copyfile(UpperCamelCase_ , os.path.join(UpperCamelCase_ , '''vocab.json''' ) ) lowerCAmelCase : List[str] = AutoProcessor.from_pretrained(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) def lowerCamelCase__ ( self : Tuple ): with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase : Union[str, Any] = WavaVecaFeatureExtractor() lowerCAmelCase : str = AutoTokenizer.from_pretrained('''facebook/wav2vec2-base-960h''' ) lowerCAmelCase : Optional[Any] = WavaVecaProcessor(UpperCamelCase_ , UpperCamelCase_ ) # save in new folder processor.save_pretrained(UpperCamelCase_ ) # drop `processor_class` in tokenizer with open(os.path.join(UpperCamelCase_ , UpperCamelCase_ ) , '''r''' ) as f: lowerCAmelCase : Tuple = json.load(UpperCamelCase_ ) config_dict.pop('''processor_class''' ) with open(os.path.join(UpperCamelCase_ , UpperCamelCase_ ) , '''w''' ) as f: f.write(json.dumps(UpperCamelCase_ ) ) lowerCAmelCase : Optional[int] = AutoProcessor.from_pretrained(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) def lowerCamelCase__ ( self : Any ): with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase : Any = WavaVecaFeatureExtractor() lowerCAmelCase : Optional[Any] = AutoTokenizer.from_pretrained('''facebook/wav2vec2-base-960h''' ) lowerCAmelCase : Any = WavaVecaProcessor(UpperCamelCase_ , UpperCamelCase_ ) # save in new folder processor.save_pretrained(UpperCamelCase_ ) # drop `processor_class` in feature extractor with open(os.path.join(UpperCamelCase_ , UpperCamelCase_ ) , '''r''' ) as f: lowerCAmelCase : Any = json.load(UpperCamelCase_ ) config_dict.pop('''processor_class''' ) with open(os.path.join(UpperCamelCase_ , UpperCamelCase_ ) , '''w''' ) as f: f.write(json.dumps(UpperCamelCase_ ) ) lowerCAmelCase : Tuple = AutoProcessor.from_pretrained(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) def lowerCamelCase__ ( self : List[Any] ): with tempfile.TemporaryDirectory() as tmpdirname: lowerCAmelCase : List[str] = WavaVecaConfig(processor_class='''Wav2Vec2Processor''' ) model_config.save_pretrained(UpperCamelCase_ ) # copy relevant files copyfile(UpperCamelCase_ , os.path.join(UpperCamelCase_ , '''vocab.json''' ) ) # create emtpy sample processor with open(os.path.join(UpperCamelCase_ , UpperCamelCase_ ) , '''w''' ) as f: f.write('''{}''' ) lowerCAmelCase : Union[str, Any] = AutoProcessor.from_pretrained(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) def lowerCamelCase__ ( self : int ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(UpperCamelCase_ ): lowerCAmelCase : Tuple = AutoProcessor.from_pretrained('''hf-internal-testing/test_dynamic_processor''' ) # If remote code is disabled, we can't load this config. with self.assertRaises(UpperCamelCase_ ): lowerCAmelCase : List[Any] = AutoProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=UpperCamelCase_ ) lowerCAmelCase : Union[str, Any] = AutoProcessor.from_pretrained('''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=UpperCamelCase_ ) self.assertTrue(processor.special_attribute_present ) self.assertEqual(processor.__class__.__name__ , '''NewProcessor''' ) lowerCAmelCase : str = processor.feature_extractor self.assertTrue(feature_extractor.special_attribute_present ) self.assertEqual(feature_extractor.__class__.__name__ , '''NewFeatureExtractor''' ) lowerCAmelCase : Optional[Any] = processor.tokenizer self.assertTrue(tokenizer.special_attribute_present ) if is_tokenizers_available(): self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizerFast''' ) # Test we can also load the slow version lowerCAmelCase : Optional[Any] = AutoProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=UpperCamelCase_ , use_fast=UpperCamelCase_ ) lowerCAmelCase : int = new_processor.tokenizer self.assertTrue(new_tokenizer.special_attribute_present ) self.assertEqual(new_tokenizer.__class__.__name__ , '''NewTokenizer''' ) else: self.assertEqual(tokenizer.__class__.__name__ , '''NewTokenizer''' ) def lowerCamelCase__ ( self : str ): try: AutoConfig.register('''custom''' , UpperCamelCase_ ) AutoFeatureExtractor.register(UpperCamelCase_ , UpperCamelCase_ ) AutoTokenizer.register(UpperCamelCase_ , slow_tokenizer_class=UpperCamelCase_ ) AutoProcessor.register(UpperCamelCase_ , UpperCamelCase_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(UpperCamelCase_ ): AutoProcessor.register(UpperCamelCase_ , UpperCamelCase_ ) # Now that the config is registered, it can be used as any other config with the auto-API lowerCAmelCase : Dict = CustomFeatureExtractor.from_pretrained(UpperCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase : List[Any] = os.path.join(UpperCamelCase_ , '''vocab.txt''' ) with open(UpperCamelCase_ , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in self.vocab_tokens] ) ) lowerCAmelCase : List[str] = CustomTokenizer(UpperCamelCase_ ) lowerCAmelCase : Any = CustomProcessor(UpperCamelCase_ , UpperCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained(UpperCamelCase_ ) lowerCAmelCase : Optional[Any] = AutoProcessor.from_pretrained(UpperCamelCase_ ) self.assertIsInstance(UpperCamelCase_ , UpperCamelCase_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def lowerCamelCase__ ( self : Optional[int] ): class snake_case_( a__ ): __UpperCamelCase = False class snake_case_( a__ ): __UpperCamelCase = False class snake_case_( a__ ): __UpperCamelCase = '''AutoFeatureExtractor''' __UpperCamelCase = '''AutoTokenizer''' __UpperCamelCase = False try: AutoConfig.register('''custom''' , UpperCamelCase_ ) AutoFeatureExtractor.register(UpperCamelCase_ , UpperCamelCase_ ) AutoTokenizer.register(UpperCamelCase_ , slow_tokenizer_class=UpperCamelCase_ ) AutoProcessor.register(UpperCamelCase_ , UpperCamelCase_ ) # If remote code is not set, the default is to use local classes. lowerCAmelCase : int = AutoProcessor.from_pretrained('''hf-internal-testing/test_dynamic_processor''' ) self.assertEqual(processor.__class__.__name__ , '''NewProcessor''' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote code is disabled, we load the local ones. lowerCAmelCase : Tuple = AutoProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=UpperCamelCase_ ) self.assertEqual(processor.__class__.__name__ , '''NewProcessor''' ) self.assertFalse(processor.special_attribute_present ) self.assertFalse(processor.feature_extractor.special_attribute_present ) self.assertFalse(processor.tokenizer.special_attribute_present ) # If remote is enabled, we load from the Hub. lowerCAmelCase : List[Any] = AutoProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_processor''' , trust_remote_code=UpperCamelCase_ ) self.assertEqual(processor.__class__.__name__ , '''NewProcessor''' ) self.assertTrue(processor.special_attribute_present ) self.assertTrue(processor.feature_extractor.special_attribute_present ) self.assertTrue(processor.tokenizer.special_attribute_present ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] if CustomConfig in TOKENIZER_MAPPING._extra_content: del TOKENIZER_MAPPING._extra_content[CustomConfig] if CustomConfig in PROCESSOR_MAPPING._extra_content: del PROCESSOR_MAPPING._extra_content[CustomConfig] def lowerCamelCase__ ( self : List[Any] ): lowerCAmelCase : Any = AutoProcessor.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) self.assertEqual(processor.__class__.__name__ , '''BertTokenizerFast''' ) def lowerCamelCase__ ( self : List[str] ): lowerCAmelCase : Tuple = AutoProcessor.from_pretrained('''hf-internal-testing/tiny-random-convnext''' ) self.assertEqual(processor.__class__.__name__ , '''ConvNextImageProcessor''' ) @is_staging_test class snake_case_( unittest.TestCase ): __UpperCamelCase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''bla''', '''blou'''] @classmethod def lowerCamelCase__ ( cls : Optional[Any] ): lowerCAmelCase : Any = TOKEN HfFolder.save_token(UpperCamelCase_ ) @classmethod def lowerCamelCase__ ( cls : str ): try: delete_repo(token=cls._token , repo_id='''test-processor''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-processor-org''' ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''test-dynamic-processor''' ) except HTTPError: pass def lowerCamelCase__ ( self : int ): lowerCAmelCase : Union[str, Any] = WavaVecaProcessor.from_pretrained(UpperCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(UpperCamelCase_ , '''test-processor''' ) , push_to_hub=UpperCamelCase_ , use_auth_token=self._token ) lowerCAmelCase : Tuple = WavaVecaProcessor.from_pretrained(F'''{USER}/test-processor''' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(UpperCamelCase_ , getattr(new_processor.feature_extractor , UpperCamelCase_ ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def lowerCamelCase__ ( self : Union[str, Any] ): lowerCAmelCase : Tuple = WavaVecaProcessor.from_pretrained(UpperCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: processor.save_pretrained( os.path.join(UpperCamelCase_ , '''test-processor-org''' ) , push_to_hub=UpperCamelCase_ , use_auth_token=self._token , organization='''valid_org''' , ) lowerCAmelCase : Union[str, Any] = WavaVecaProcessor.from_pretrained('''valid_org/test-processor-org''' ) for k, v in processor.feature_extractor.__dict__.items(): self.assertEqual(UpperCamelCase_ , getattr(new_processor.feature_extractor , UpperCamelCase_ ) ) self.assertDictEqual(new_processor.tokenizer.get_vocab() , processor.tokenizer.get_vocab() ) def lowerCamelCase__ ( self : Any ): CustomFeatureExtractor.register_for_auto_class() CustomTokenizer.register_for_auto_class() CustomProcessor.register_for_auto_class() lowerCAmelCase : Any = CustomFeatureExtractor.from_pretrained(UpperCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: lowerCAmelCase : Any = os.path.join(UpperCamelCase_ , '''vocab.txt''' ) with open(UpperCamelCase_ , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in self.vocab_tokens] ) ) lowerCAmelCase : int = CustomTokenizer(UpperCamelCase_ ) lowerCAmelCase : Dict = CustomProcessor(UpperCamelCase_ , UpperCamelCase_ ) with tempfile.TemporaryDirectory() as tmp_dir: create_repo(F'''{USER}/test-dynamic-processor''' , token=self._token ) lowerCAmelCase : Dict = Repository(UpperCamelCase_ , clone_from=F'''{USER}/test-dynamic-processor''' , token=self._token ) processor.save_pretrained(UpperCamelCase_ ) # This has added the proper auto_map field to the feature extractor config self.assertDictEqual( processor.feature_extractor.auto_map , { '''AutoFeatureExtractor''': '''custom_feature_extraction.CustomFeatureExtractor''', '''AutoProcessor''': '''custom_processing.CustomProcessor''', } , ) # This has added the proper auto_map field to the tokenizer config with open(os.path.join(UpperCamelCase_ , '''tokenizer_config.json''' ) ) as f: lowerCAmelCase : Tuple = json.load(UpperCamelCase_ ) self.assertDictEqual( tokenizer_config['''auto_map'''] , { '''AutoTokenizer''': ['''custom_tokenization.CustomTokenizer''', None], '''AutoProcessor''': '''custom_processing.CustomProcessor''', } , ) # The code has been copied from fixtures self.assertTrue(os.path.isfile(os.path.join(UpperCamelCase_ , '''custom_feature_extraction.py''' ) ) ) self.assertTrue(os.path.isfile(os.path.join(UpperCamelCase_ , '''custom_tokenization.py''' ) ) ) self.assertTrue(os.path.isfile(os.path.join(UpperCamelCase_ , '''custom_processing.py''' ) ) ) repo.push_to_hub() lowerCAmelCase : int = AutoProcessor.from_pretrained(F'''{USER}/test-dynamic-processor''' , trust_remote_code=UpperCamelCase_ ) # Can't make an isinstance check because the new_processor is from the CustomProcessor class of a dynamic module self.assertEqual(new_processor.__class__.__name__ , '''CustomProcessor''' )
314
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from ...schedulers import DDIMScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class snake_case_( a__ ): def __init__( self : Dict , UpperCamelCase_ : Any , UpperCamelCase_ : List[str] ): super().__init__() # make sure scheduler can always be converted to DDIM lowerCAmelCase : str = DDIMScheduler.from_config(scheduler.config ) self.register_modules(unet=UpperCamelCase_ , scheduler=UpperCamelCase_ ) @torch.no_grad() def __call__( self : str , UpperCamelCase_ : int = 1 , UpperCamelCase_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , UpperCamelCase_ : float = 0.0 , UpperCamelCase_ : int = 5_0 , UpperCamelCase_ : Optional[bool] = None , UpperCamelCase_ : Optional[str] = "pil" , UpperCamelCase_ : bool = True , ): # Sample gaussian noise to begin loop if isinstance(self.unet.config.sample_size , UpperCamelCase_ ): lowerCAmelCase : Dict = ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size, ) else: lowerCAmelCase : str = (batch_size, self.unet.config.in_channels, *self.unet.config.sample_size) if isinstance(UpperCamelCase_ , UpperCamelCase_ ) and len(UpperCamelCase_ ) != batch_size: raise ValueError( F'''You have passed a list of generators of length {len(UpperCamelCase_ )}, but requested an effective batch''' F''' size of {batch_size}. Make sure the batch size matches the length of the generators.''' ) lowerCAmelCase : int = randn_tensor(UpperCamelCase_ , generator=UpperCamelCase_ , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(UpperCamelCase_ ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output lowerCAmelCase : Optional[Any] = self.unet(UpperCamelCase_ , UpperCamelCase_ ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 lowerCAmelCase : Dict = self.scheduler.step( UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , eta=UpperCamelCase_ , use_clipped_model_output=UpperCamelCase_ , generator=UpperCamelCase_ ).prev_sample lowerCAmelCase : Tuple = (image / 2 + 0.5).clamp(0 , 1 ) lowerCAmelCase : str = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCAmelCase : Any = self.numpy_to_pil(UpperCamelCase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=UpperCamelCase_ )
314
1
'''simple docstring''' import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def lowerCAmelCase_ ( snake_case_ : List[Any] , snake_case_ : int ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = checkpoint UpperCAmelCase_ = {} UpperCAmelCase_ = vae_state_dict["encoder.conv_in.weight"] UpperCAmelCase_ = vae_state_dict["encoder.conv_in.bias"] UpperCAmelCase_ = vae_state_dict["encoder.conv_out.weight"] UpperCAmelCase_ = vae_state_dict["encoder.conv_out.bias"] UpperCAmelCase_ = vae_state_dict["encoder.norm_out.weight"] UpperCAmelCase_ = vae_state_dict["encoder.norm_out.bias"] UpperCAmelCase_ = vae_state_dict["decoder.conv_in.weight"] UpperCAmelCase_ = vae_state_dict["decoder.conv_in.bias"] UpperCAmelCase_ = vae_state_dict["decoder.conv_out.weight"] UpperCAmelCase_ = vae_state_dict["decoder.conv_out.bias"] UpperCAmelCase_ = vae_state_dict["decoder.norm_out.weight"] UpperCAmelCase_ = vae_state_dict["decoder.norm_out.bias"] UpperCAmelCase_ = vae_state_dict["quant_conv.weight"] UpperCAmelCase_ = vae_state_dict["quant_conv.bias"] UpperCAmelCase_ = vae_state_dict["post_quant_conv.weight"] UpperCAmelCase_ = vae_state_dict["post_quant_conv.bias"] # Retrieves the keys for the encoder down blocks only UpperCAmelCase_ = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "encoder.down" in layer} ) UpperCAmelCase_ = { layer_id: [key for key in vae_state_dict if f"""down.{layer_id}""" in key] for layer_id in range(snake_case_ ) } # Retrieves the keys for the decoder up blocks only UpperCAmelCase_ = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "decoder.up" in layer} ) UpperCAmelCase_ = { layer_id: [key for key in vae_state_dict if f"""up.{layer_id}""" in key] for layer_id in range(snake_case_ ) } for i in range(snake_case_ ): UpperCAmelCase_ = [key for key in down_blocks[i] if f"""down.{i}""" in key and f"""down.{i}.downsample""" not in key] if f"""encoder.down.{i}.downsample.conv.weight""" in vae_state_dict: UpperCAmelCase_ = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.weight""" ) UpperCAmelCase_ = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.bias""" ) UpperCAmelCase_ = renew_vae_resnet_paths(snake_case_ ) UpperCAmelCase_ = {"old": f"""down.{i}.block""", "new": f"""down_blocks.{i}.resnets"""} assign_to_checkpoint(snake_case_ , snake_case_ , snake_case_ , additional_replacements=[meta_path] , config=snake_case_ ) UpperCAmelCase_ = [key for key in vae_state_dict if "encoder.mid.block" in key] UpperCAmelCase_ = 2 for i in range(1 , num_mid_res_blocks + 1 ): UpperCAmelCase_ = [key for key in mid_resnets if f"""encoder.mid.block_{i}""" in key] UpperCAmelCase_ = renew_vae_resnet_paths(snake_case_ ) UpperCAmelCase_ = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(snake_case_ , snake_case_ , snake_case_ , additional_replacements=[meta_path] , config=snake_case_ ) UpperCAmelCase_ = [key for key in vae_state_dict if "encoder.mid.attn" in key] UpperCAmelCase_ = renew_vae_attention_paths(snake_case_ ) UpperCAmelCase_ = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(snake_case_ , snake_case_ , snake_case_ , additional_replacements=[meta_path] , config=snake_case_ ) conv_attn_to_linear(snake_case_ ) for i in range(snake_case_ ): UpperCAmelCase_ = num_up_blocks - 1 - i UpperCAmelCase_ = [ key for key in up_blocks[block_id] if f"""up.{block_id}""" in key and f"""up.{block_id}.upsample""" not in key ] if f"""decoder.up.{block_id}.upsample.conv.weight""" in vae_state_dict: UpperCAmelCase_ = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.weight""" ] UpperCAmelCase_ = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.bias""" ] UpperCAmelCase_ = renew_vae_resnet_paths(snake_case_ ) UpperCAmelCase_ = {"old": f"""up.{block_id}.block""", "new": f"""up_blocks.{i}.resnets"""} assign_to_checkpoint(snake_case_ , snake_case_ , snake_case_ , additional_replacements=[meta_path] , config=snake_case_ ) UpperCAmelCase_ = [key for key in vae_state_dict if "decoder.mid.block" in key] UpperCAmelCase_ = 2 for i in range(1 , num_mid_res_blocks + 1 ): UpperCAmelCase_ = [key for key in mid_resnets if f"""decoder.mid.block_{i}""" in key] UpperCAmelCase_ = renew_vae_resnet_paths(snake_case_ ) UpperCAmelCase_ = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(snake_case_ , snake_case_ , snake_case_ , additional_replacements=[meta_path] , config=snake_case_ ) UpperCAmelCase_ = [key for key in vae_state_dict if "decoder.mid.attn" in key] UpperCAmelCase_ = renew_vae_attention_paths(snake_case_ ) UpperCAmelCase_ = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(snake_case_ , snake_case_ , snake_case_ , additional_replacements=[meta_path] , config=snake_case_ ) conv_attn_to_linear(snake_case_ ) return new_checkpoint def lowerCAmelCase_ ( snake_case_ : str , snake_case_ : str , ) -> Dict: '''simple docstring''' UpperCAmelCase_ = requests.get( " https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml" ) UpperCAmelCase_ = io.BytesIO(r.content ) UpperCAmelCase_ = OmegaConf.load(snake_case_ ) UpperCAmelCase_ = 5_12 UpperCAmelCase_ = "cuda" if torch.cuda.is_available() else "cpu" if checkpoint_path.endswith("safetensors" ): from safetensors import safe_open UpperCAmelCase_ = {} with safe_open(snake_case_ , framework="pt" , device="cpu" ) as f: for key in f.keys(): UpperCAmelCase_ = f.get_tensor(snake_case_ ) else: UpperCAmelCase_ = torch.load(snake_case_ , map_location=snake_case_ )["state_dict"] # Convert the VAE model. UpperCAmelCase_ = create_vae_diffusers_config(snake_case_ , image_size=snake_case_ ) UpperCAmelCase_ = custom_convert_ldm_vae_checkpoint(snake_case_ , snake_case_ ) UpperCAmelCase_ = AutoencoderKL(**snake_case_ ) vae.load_state_dict(snake_case_ ) vae.save_pretrained(snake_case_ ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_: Optional[int] =argparse.ArgumentParser() parser.add_argument('--vae_pt_path', default=None, type=str, required=True, help='Path to the VAE.pt to convert.') parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the VAE.pt to convert.') SCREAMING_SNAKE_CASE_: str =parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
1
'''simple docstring''' from math import log from scipy.constants import Boltzmann, physical_constants SCREAMING_SNAKE_CASE_: Optional[int] =3_00 # TEMPERATURE (unit = K) def lowerCAmelCase_ ( snake_case_ : float , snake_case_ : float , snake_case_ : float , ) -> float: '''simple docstring''' if donor_conc <= 0: raise ValueError("Donor concentration should be positive" ) elif acceptor_conc <= 0: raise ValueError("Acceptor concentration should be positive" ) elif intrinsic_conc <= 0: raise ValueError("Intrinsic concentration should be positive" ) elif donor_conc <= intrinsic_conc: raise ValueError( "Donor concentration should be greater than intrinsic concentration" ) elif acceptor_conc <= intrinsic_conc: raise ValueError( "Acceptor concentration should be greater than intrinsic concentration" ) else: return ( Boltzmann * T * log((donor_conc * acceptor_conc) / intrinsic_conc**2 ) / physical_constants["electron volt"][0] ) if __name__ == "__main__": import doctest doctest.testmod()
1
1
"""simple docstring""" import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def __snake_case ( SCREAMING_SNAKE_CASE__ : bytes , SCREAMING_SNAKE_CASE__ : int ) -> np.array: '''simple docstring''' _UpperCAmelCase : List[str] = f'{sampling_rate}' _UpperCAmelCase : str = "1" _UpperCAmelCase : Optional[Any] = "f32le" _UpperCAmelCase : Optional[int] = [ "ffmpeg", "-i", "pipe:0", "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-hide_banner", "-loglevel", "quiet", "pipe:1", ] try: with subprocess.Popen(SCREAMING_SNAKE_CASE__ , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: _UpperCAmelCase : List[Any] = ffmpeg_process.communicate(SCREAMING_SNAKE_CASE__ ) except FileNotFoundError as error: raise ValueError("ffmpeg was not found but is required to load audio files from filename" ) from error _UpperCAmelCase : Dict = output_stream[0] _UpperCAmelCase : Optional[Any] = np.frombuffer(SCREAMING_SNAKE_CASE__ , np.floataa ) if audio.shape[0] == 0: raise ValueError("Malformed soundfile" ) return audio def __snake_case ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : float , SCREAMING_SNAKE_CASE__ : str = "f32le" , ) -> Optional[int]: '''simple docstring''' _UpperCAmelCase : List[Any] = f'{sampling_rate}' _UpperCAmelCase : int = "1" if format_for_conversion == "s16le": _UpperCAmelCase : Dict = 2 elif format_for_conversion == "f32le": _UpperCAmelCase : Optional[int] = 4 else: raise ValueError(f'Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`' ) _UpperCAmelCase : List[Any] = platform.system() if system == "Linux": _UpperCAmelCase : Dict = "alsa" _UpperCAmelCase : Optional[Any] = "default" elif system == "Darwin": _UpperCAmelCase : Union[str, Any] = "avfoundation" _UpperCAmelCase : str = ":0" elif system == "Windows": _UpperCAmelCase : Optional[int] = "dshow" _UpperCAmelCase : Union[str, Any] = "default" _UpperCAmelCase : str = [ "ffmpeg", "-f", format_, "-i", input_, "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-fflags", "nobuffer", "-hide_banner", "-loglevel", "quiet", "pipe:1", ] _UpperCAmelCase : Tuple = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample _UpperCAmelCase : Optional[Any] = _ffmpeg_stream(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ ) for item in iterator: yield item def __snake_case ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : float , SCREAMING_SNAKE_CASE__ : Optional[int] = None , SCREAMING_SNAKE_CASE__ : Optional[Union[Tuple[float, float], float]] = None , SCREAMING_SNAKE_CASE__ : str = "f32le" , ) -> Tuple: '''simple docstring''' if stream_chunk_s is not None: _UpperCAmelCase : Dict = stream_chunk_s else: _UpperCAmelCase : Any = chunk_length_s _UpperCAmelCase : Any = ffmpeg_microphone(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , format_for_conversion=SCREAMING_SNAKE_CASE__ ) if format_for_conversion == "s16le": _UpperCAmelCase : str = np.intaa _UpperCAmelCase : List[str] = 2 elif format_for_conversion == "f32le": _UpperCAmelCase : Tuple = np.floataa _UpperCAmelCase : Union[str, Any] = 4 else: raise ValueError(f'Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`' ) if stride_length_s is None: _UpperCAmelCase : Tuple = chunk_length_s / 6 _UpperCAmelCase : Dict = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(SCREAMING_SNAKE_CASE__ , (int, float) ): _UpperCAmelCase : Tuple = [stride_length_s, stride_length_s] _UpperCAmelCase : str = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample _UpperCAmelCase : Tuple = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample _UpperCAmelCase : Union[str, Any] = datetime.datetime.now() _UpperCAmelCase : Any = datetime.timedelta(seconds=SCREAMING_SNAKE_CASE__ ) for item in chunk_bytes_iter(SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ , stride=(stride_left, stride_right) , stream=SCREAMING_SNAKE_CASE__ ): # Put everything back in numpy scale _UpperCAmelCase : str = np.frombuffer(item["raw"] , dtype=SCREAMING_SNAKE_CASE__ ) _UpperCAmelCase : Optional[Any] = ( item["stride"][0] // size_of_sample, item["stride"][1] // size_of_sample, ) _UpperCAmelCase : Any = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def __snake_case ( SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Tuple[int, int] , SCREAMING_SNAKE_CASE__ : bool = False ) -> Union[str, Any]: '''simple docstring''' _UpperCAmelCase : int = b"" _UpperCAmelCase : Optional[Any] = stride if stride_left + stride_right >= chunk_len: raise ValueError( f'Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}' ) _UpperCAmelCase : Optional[Any] = 0 for raw in iterator: acc += raw if stream and len(SCREAMING_SNAKE_CASE__ ) < chunk_len: _UpperCAmelCase : Optional[int] = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(SCREAMING_SNAKE_CASE__ ) >= chunk_len: # We are flushing the accumulator _UpperCAmelCase : List[Any] = (_stride_left, stride_right) _UpperCAmelCase : List[str] = {"raw": acc[:chunk_len], "stride": stride} if stream: _UpperCAmelCase : int = False yield item _UpperCAmelCase : Optional[int] = stride_left _UpperCAmelCase : List[Any] = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(SCREAMING_SNAKE_CASE__ ) > stride_left: _UpperCAmelCase : Optional[int] = {"raw": acc, "stride": (_stride_left, 0)} if stream: _UpperCAmelCase : List[Any] = False yield item def __snake_case ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : int ) -> Dict: '''simple docstring''' _UpperCAmelCase : str = 2**24 # 16Mo try: with subprocess.Popen(SCREAMING_SNAKE_CASE__ , stdout=subprocess.PIPE , bufsize=SCREAMING_SNAKE_CASE__ ) as ffmpeg_process: while True: _UpperCAmelCase : Optional[int] = ffmpeg_process.stdout.read(SCREAMING_SNAKE_CASE__ ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError("ffmpeg was not found but is required to stream audio files from filename" ) from error
359
"""simple docstring""" import inspect import unittest from transformers import DPTConfig from transformers.file_utils import is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MODEL_MAPPING, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel from transformers.models.dpt.modeling_dpt import DPT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class UpperCAmelCase_ : def __init__( self : Tuple , A : str , A : Dict=2 , A : List[Any]=3_2 , A : Optional[Any]=1_6 , A : Tuple=3 , A : Optional[Any]=True , A : List[Any]=True , A : Optional[int]=3_2 , A : Optional[int]=4 , A : Tuple=[0, 1, 2, 3] , A : Optional[int]=4 , A : Tuple=3_7 , A : List[Any]="gelu" , A : List[Any]=0.1 , A : List[str]=0.1 , A : Union[str, Any]=0.02 , A : Optional[int]=3 , A : Optional[Any]=[1, 3_8_4, 2_4, 2_4] , A : Union[str, Any]=True , A : Any=None , ): _UpperCAmelCase : str = parent _UpperCAmelCase : int = batch_size _UpperCAmelCase : List[str] = image_size _UpperCAmelCase : Tuple = patch_size _UpperCAmelCase : Any = num_channels _UpperCAmelCase : List[str] = is_training _UpperCAmelCase : Optional[Any] = use_labels _UpperCAmelCase : str = hidden_size _UpperCAmelCase : Dict = num_hidden_layers _UpperCAmelCase : List[Any] = backbone_out_indices _UpperCAmelCase : Optional[Any] = num_attention_heads _UpperCAmelCase : Optional[int] = intermediate_size _UpperCAmelCase : str = hidden_act _UpperCAmelCase : int = hidden_dropout_prob _UpperCAmelCase : Any = attention_probs_dropout_prob _UpperCAmelCase : Optional[Any] = initializer_range _UpperCAmelCase : Optional[Any] = num_labels _UpperCAmelCase : Tuple = backbone_featmap_shape _UpperCAmelCase : Optional[Any] = scope _UpperCAmelCase : Union[str, Any] = is_hybrid # sequence length of DPT = num_patches + 1 (we add 1 for the [CLS] token) _UpperCAmelCase : Optional[int] = (image_size // patch_size) ** 2 _UpperCAmelCase : int = num_patches + 1 def snake_case_ ( self : List[str] ): _UpperCAmelCase : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase : Dict = None if self.use_labels: _UpperCAmelCase : List[str] = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) _UpperCAmelCase : str = self.get_config() return config, pixel_values, labels def snake_case_ ( self : str ): _UpperCAmelCase : Tuple = { "global_padding": "same", "layer_type": "bottleneck", "depths": [3, 4, 9], "out_features": ["stage1", "stage2", "stage3"], "embedding_dynamic_padding": True, "hidden_sizes": [9_6, 1_9_2, 3_8_4, 7_6_8], "num_groups": 2, } return DPTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , backbone_out_indices=self.backbone_out_indices , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=A , initializer_range=self.initializer_range , is_hybrid=self.is_hybrid , backbone_config=A , backbone_featmap_shape=self.backbone_featmap_shape , ) def snake_case_ ( self : Any , A : Optional[Any] , A : str , A : Dict ): _UpperCAmelCase : List[str] = DPTModel(config=A ) model.to(A ) model.eval() _UpperCAmelCase : List[str] = model(A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case_ ( self : List[str] , A : str , A : Any , A : List[Any] ): _UpperCAmelCase : str = self.num_labels _UpperCAmelCase : Any = DPTForDepthEstimation(A ) model.to(A ) model.eval() _UpperCAmelCase : Optional[int] = model(A ) self.parent.assertEqual(result.predicted_depth.shape , (self.batch_size, self.image_size, self.image_size) ) def snake_case_ ( self : List[Any] , A : Any , A : Optional[int] , A : Union[str, Any] ): _UpperCAmelCase : List[Any] = self.num_labels _UpperCAmelCase : Union[str, Any] = DPTForSemanticSegmentation(A ) model.to(A ) model.eval() _UpperCAmelCase : int = model(A , labels=A ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def snake_case_ ( self : Union[str, Any] ): _UpperCAmelCase : Union[str, Any] = self.prepare_config_and_inputs() _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Dict = config_and_inputs _UpperCAmelCase : int = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class UpperCAmelCase_ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE : List[str] = (DPTModel, DPTForDepthEstimation, DPTForSemanticSegmentation) if is_torch_available() else () __SCREAMING_SNAKE_CASE : Any = ( { 'depth-estimation': DPTForDepthEstimation, 'feature-extraction': DPTModel, 'image-segmentation': DPTForSemanticSegmentation, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE : Dict = False __SCREAMING_SNAKE_CASE : Optional[int] = False __SCREAMING_SNAKE_CASE : Optional[Any] = False def snake_case_ ( self : int ): _UpperCAmelCase : List[str] = DPTModelTester(self ) _UpperCAmelCase : Union[str, Any] = ConfigTester(self , config_class=A , has_text_modality=A , hidden_size=3_7 ) def snake_case_ ( self : int ): self.config_tester.run_common_tests() @unittest.skip(reason="DPT does not use inputs_embeds" ) def snake_case_ ( self : Union[str, Any] ): pass def snake_case_ ( self : Tuple ): _UpperCAmelCase , _UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase : str = model_class(A ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _UpperCAmelCase : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(A , nn.Linear ) ) def snake_case_ ( self : Optional[Any] ): _UpperCAmelCase , _UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase : int = model_class(A ) _UpperCAmelCase : int = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase : str = [*signature.parameters.keys()] _UpperCAmelCase : Any = ["pixel_values"] self.assertListEqual(arg_names[:1] , A ) def snake_case_ ( self : List[str] ): _UpperCAmelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def snake_case_ ( self : Dict ): _UpperCAmelCase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_depth_estimation(*A ) def snake_case_ ( self : Dict ): _UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*A ) def snake_case_ ( self : Optional[int] ): for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue _UpperCAmelCase , _UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase : Optional[int] = True if model_class in get_values(A ): continue _UpperCAmelCase : int = model_class(A ) model.to(A ) model.train() _UpperCAmelCase : Optional[Any] = self._prepare_for_class(A , A , return_labels=A ) _UpperCAmelCase : Optional[Any] = model(**A ).loss loss.backward() def snake_case_ ( self : Dict ): for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue _UpperCAmelCase , _UpperCAmelCase : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase : Tuple = False _UpperCAmelCase : Tuple = True if model_class in get_values(A ) or not model_class.supports_gradient_checkpointing: continue _UpperCAmelCase : int = model_class(A ) model.to(A ) model.gradient_checkpointing_enable() model.train() _UpperCAmelCase : str = self._prepare_for_class(A , A , return_labels=A ) _UpperCAmelCase : List[str] = model(**A ).loss loss.backward() def snake_case_ ( self : Union[str, Any] ): _UpperCAmelCase , _UpperCAmelCase : Tuple = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase : int = _config_zero_init(A ) for model_class in self.all_model_classes: _UpperCAmelCase : List[Any] = model_class(config=A ) # Skip the check for the backbone _UpperCAmelCase : Dict = [] for name, module in model.named_modules(): if module.__class__.__name__ == "DPTViTHybridEmbeddings": _UpperCAmelCase : List[str] = [f'{name}.{key}' for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f'Parameter {name} of model {model_class} seems not properly initialized' , ) @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def snake_case_ ( self : int ): pass @slow def snake_case_ ( self : Dict ): for model_name in DPT_PRETRAINED_MODEL_ARCHIVE_LIST[1:]: _UpperCAmelCase : Any = DPTModel.from_pretrained(A ) self.assertIsNotNone(A ) def snake_case_ ( self : Tuple ): # We do this test only for DPTForDepthEstimation since it is the only model that uses readout_type _UpperCAmelCase , _UpperCAmelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() _UpperCAmelCase : Optional[Any] = "add" with self.assertRaises(A ): _UpperCAmelCase : List[Any] = DPTForDepthEstimation(A ) def __snake_case ( ) -> Optional[int]: '''simple docstring''' _UpperCAmelCase : int = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision @slow class UpperCAmelCase_ ( unittest.TestCase ): def snake_case_ ( self : List[str] ): _UpperCAmelCase : Optional[int] = DPTImageProcessor.from_pretrained("Intel/dpt-hybrid-midas" ) _UpperCAmelCase : Any = DPTForDepthEstimation.from_pretrained("Intel/dpt-hybrid-midas" ).to(A ) _UpperCAmelCase : str = prepare_img() _UpperCAmelCase : Tuple = image_processor(images=A , return_tensors="pt" ).to(A ) # forward pass with torch.no_grad(): _UpperCAmelCase : int = model(**A ) _UpperCAmelCase : List[str] = outputs.predicted_depth # verify the predicted depth _UpperCAmelCase : int = torch.Size((1, 3_8_4, 3_8_4) ) self.assertEqual(predicted_depth.shape , A ) _UpperCAmelCase : int = torch.tensor( [[[5.6_437, 5.6_146, 5.6_511], [5.4_371, 5.5_649, 5.5_958], [5.5_215, 5.5_184, 5.5_293]]] ).to(A ) self.assertTrue(torch.allclose(outputs.predicted_depth[:3, :3, :3] / 1_0_0 , A , atol=1e-4 ) )
202
0