code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { 'funnel-transformer/small': 'https://huggingface.co/funnel-transformer/small/resolve/main/config.json', 'funnel-transformer/small-base': 'https://huggingface.co/funnel-transformer/small-base/resolve/main/config.json', 'funnel-transformer/medium': 'https://huggingface.co/funnel-transformer/medium/resolve/main/config.json', 'funnel-transformer/medium-base': 'https://huggingface.co/funnel-transformer/medium-base/resolve/main/config.json', 'funnel-transformer/intermediate': ( 'https://huggingface.co/funnel-transformer/intermediate/resolve/main/config.json' ), 'funnel-transformer/intermediate-base': ( 'https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/config.json' ), 'funnel-transformer/large': 'https://huggingface.co/funnel-transformer/large/resolve/main/config.json', 'funnel-transformer/large-base': 'https://huggingface.co/funnel-transformer/large-base/resolve/main/config.json', 'funnel-transformer/xlarge': 'https://huggingface.co/funnel-transformer/xlarge/resolve/main/config.json', 'funnel-transformer/xlarge-base': 'https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/config.json', } class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = """funnel""" SCREAMING_SNAKE_CASE__ : int = { """hidden_size""": """d_model""", """num_attention_heads""": """n_head""", } def __init__( self , lowercase_=3_0522 , lowercase_=[4, 4, 4] , lowercase_=None , lowercase_=2 , lowercase_=768 , lowercase_=12 , lowercase_=64 , lowercase_=3072 , lowercase_="gelu_new" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=0.0 , lowercase_=0.1 , lowercase_=None , lowercase_=1E-9 , lowercase_="mean" , lowercase_="relative_shift" , lowercase_=True , lowercase_=True , lowercase_=True , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = vocab_size UpperCAmelCase_ : List[Any] = block_sizes UpperCAmelCase_ : Optional[Any] = [1] * len(lowercase_ ) if block_repeats is None else block_repeats assert len(lowercase_ ) == len( self.block_repeats ), "`block_sizes` and `block_repeats` should have the same length." UpperCAmelCase_ : str = num_decoder_layers UpperCAmelCase_ : List[Any] = d_model UpperCAmelCase_ : Any = n_head UpperCAmelCase_ : List[Any] = d_head UpperCAmelCase_ : int = d_inner UpperCAmelCase_ : Dict = hidden_act UpperCAmelCase_ : Union[str, Any] = hidden_dropout UpperCAmelCase_ : Optional[int] = attention_dropout UpperCAmelCase_ : Any = activation_dropout UpperCAmelCase_ : int = initializer_range UpperCAmelCase_ : str = initializer_std UpperCAmelCase_ : Any = layer_norm_eps assert pooling_type in [ "mean", "max", ], F"""Got {pooling_type} for `pooling_type` but only 'mean' and 'max' are supported.""" UpperCAmelCase_ : Any = pooling_type assert attention_type in [ "relative_shift", "factorized", ], F"""Got {attention_type} for `attention_type` but only 'relative_shift' and 'factorized' are supported.""" UpperCAmelCase_ : Union[str, Any] = attention_type UpperCAmelCase_ : Union[str, Any] = separate_cls UpperCAmelCase_ : Dict = truncate_seq UpperCAmelCase_ : Any = pool_q_only super().__init__(**lowercase_ ) @property def UpperCamelCase__ ( self ): """simple docstring""" return sum(self.block_sizes ) @num_hidden_layers.setter def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" raise NotImplementedError( "This model does not support the setting of `num_hidden_layers`. Please set `block_sizes`." ) @property def UpperCamelCase__ ( self ): """simple docstring""" return len(self.block_sizes ) @num_blocks.setter def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" raise NotImplementedError("This model does not support the setting of `num_blocks`. Please set `block_sizes`." )
23
"""simple docstring""" _a = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(100_000)] def __a ( __lowerCamelCase ): UpperCAmelCase_ : Optional[int] = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 10_0000] number //= 10_0000 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution _a = [None] * 10_000_000 _a = True _a = False def __a ( __lowerCamelCase ): if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore UpperCAmelCase_ : Dict = chain(next_number(__lowerCamelCase ) ) UpperCAmelCase_ : List[str] = number_chain while number < 1000_0000: UpperCAmelCase_ : List[Any] = number_chain number *= 10 return number_chain def __a ( __lowerCamelCase = 1000_0000 ): for i in range(1, __lowerCamelCase ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(__lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() print(f"""{solution() = }""")
23
1
"""simple docstring""" import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType ######################################################################## # This is a fully working simple example to use Accelerate # and perform gradient accumulation # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## _a = 16 _a = 32 def __a ( __lowerCamelCase, __lowerCamelCase = 16 ): UpperCAmelCase_ : Optional[Any] = AutoTokenizer.from_pretrained("bert-base-cased" ) UpperCAmelCase_ : int = load_dataset("glue", "mrpc" ) def tokenize_function(__lowerCamelCase ): # max_length=None => use the model max length (it's actually the default) UpperCAmelCase_ : Optional[Any] = tokenizer(examples["sentence1"], examples["sentence2"], truncation=__lowerCamelCase, max_length=__lowerCamelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): UpperCAmelCase_ : List[str] = datasets.map( __lowerCamelCase, batched=__lowerCamelCase, remove_columns=["idx", "sentence1", "sentence2"], ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library UpperCAmelCase_ : Union[str, Any] = tokenized_datasets.rename_column("label", "labels" ) def collate_fn(__lowerCamelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. UpperCAmelCase_ : Optional[Any] = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": UpperCAmelCase_ : int = 16 elif accelerator.mixed_precision != "no": UpperCAmelCase_ : Optional[Any] = 8 else: UpperCAmelCase_ : Optional[Any] = None return tokenizer.pad( __lowerCamelCase, padding="longest", max_length=__lowerCamelCase, pad_to_multiple_of=__lowerCamelCase, return_tensors="pt", ) # Instantiate dataloaders. UpperCAmelCase_ : List[Any] = DataLoader( tokenized_datasets["train"], shuffle=__lowerCamelCase, collate_fn=__lowerCamelCase, batch_size=__lowerCamelCase ) UpperCAmelCase_ : Dict = DataLoader( tokenized_datasets["validation"], shuffle=__lowerCamelCase, collate_fn=__lowerCamelCase, batch_size=__lowerCamelCase ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders _a = mocked_dataloaders # noqa: F811 def __a ( __lowerCamelCase, __lowerCamelCase ): # For testing only if os.environ.get("TESTING_MOCKED_DATALOADERS", __lowerCamelCase ) == "1": UpperCAmelCase_ : str = 2 # New Code # UpperCAmelCase_ : Dict = int(args.gradient_accumulation_steps ) # Initialize accelerator UpperCAmelCase_ : List[Any] = Accelerator( cpu=args.cpu, mixed_precision=args.mixed_precision, gradient_accumulation_steps=__lowerCamelCase ) if accelerator.distributed_type == DistributedType.TPU and gradient_accumulation_steps > 1: raise NotImplementedError( "Gradient accumulation on TPUs is currently not supported. Pass `gradient_accumulation_steps=1`" ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs UpperCAmelCase_ : Optional[int] = config["lr"] UpperCAmelCase_ : Union[str, Any] = int(config["num_epochs"] ) UpperCAmelCase_ : Optional[Any] = int(config["seed"] ) UpperCAmelCase_ : Optional[int] = int(config["batch_size"] ) UpperCAmelCase_ : str = evaluate.load("glue", "mrpc" ) set_seed(__lowerCamelCase ) UpperCAmelCase_ , UpperCAmelCase_ : Tuple = get_dataloaders(__lowerCamelCase, __lowerCamelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) UpperCAmelCase_ : Optional[int] = AutoModelForSequenceClassification.from_pretrained("bert-base-cased", return_dict=__lowerCamelCase ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). UpperCAmelCase_ : Optional[int] = model.to(accelerator.device ) # Instantiate optimizer UpperCAmelCase_ : str = AdamW(params=model.parameters(), lr=__lowerCamelCase ) # Instantiate scheduler UpperCAmelCase_ : Tuple = get_linear_schedule_with_warmup( optimizer=__lowerCamelCase, num_warmup_steps=100, num_training_steps=(len(__lowerCamelCase ) * num_epochs), ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Dict = accelerator.prepare( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) # Now we train the model for epoch in range(__lowerCamelCase ): model.train() for step, batch in enumerate(__lowerCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(__lowerCamelCase ): UpperCAmelCase_ : Tuple = model(**__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = output.loss accelerator.backward(__lowerCamelCase ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__lowerCamelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): UpperCAmelCase_ : List[Any] = model(**__lowerCamelCase ) UpperCAmelCase_ : List[Any] = outputs.logits.argmax(dim=-1 ) UpperCAmelCase_ , UpperCAmelCase_ : List[str] = accelerator.gather_for_metrics((predictions, batch["labels"]) ) metric.add_batch( predictions=__lowerCamelCase, references=__lowerCamelCase, ) UpperCAmelCase_ : Optional[int] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""", __lowerCamelCase ) def __a ( ): UpperCAmelCase_ : Tuple = argparse.ArgumentParser(description="Simple example of training script." ) parser.add_argument( "--mixed_precision", type=__lowerCamelCase, default=__lowerCamelCase, choices=["no", "fp16", "bf16", "fp8"], help="Whether to use mixed precision. Choose" "between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10." "and an Nvidia Ampere GPU.", ) # New Code # parser.add_argument( "--gradient_accumulation_steps", type=__lowerCamelCase, default=1, help="The number of minibatches to be ran before gradients are accumulated.", ) parser.add_argument("--cpu", action="store_true", help="If passed, will train on the CPU." ) UpperCAmelCase_ : int = parser.parse_args() UpperCAmelCase_ : List[Any] = {"lr": 2E-5, "num_epochs": 3, "seed": 42, "batch_size": 16} training_function(__lowerCamelCase, __lowerCamelCase ) if __name__ == "__main__": main()
23
"""simple docstring""" def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # Return True if there is node that has not iterated. UpperCAmelCase_ : List[Any] = [False] * len(__lowerCamelCase ) UpperCAmelCase_ : Any = [] queue.append(__lowerCamelCase ) UpperCAmelCase_ : Tuple = True while queue: UpperCAmelCase_ : str = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(__lowerCamelCase ) UpperCAmelCase_ : Any = True UpperCAmelCase_ : Union[str, Any] = u return visited[t] def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # This array is filled by BFS and to store path UpperCAmelCase_ : List[str] = [-1] * (len(__lowerCamelCase )) UpperCAmelCase_ : Any = 0 while bfs(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : int = float("Inf" ) UpperCAmelCase_ : Tuple = sink while s != source: # Find the minimum value in select path UpperCAmelCase_ : Tuple = min(__lowerCamelCase, graph[parent[s]][s] ) UpperCAmelCase_ : Dict = parent[s] max_flow += path_flow UpperCAmelCase_ : Optional[Any] = sink while v != source: UpperCAmelCase_ : List[str] = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow UpperCAmelCase_ : Optional[int] = parent[v] return max_flow _a = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] _a , _a = 0, 5 print(ford_fulkerson(graph, source, sink))
23
1
"""simple docstring""" import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import BatchEncoding, PreTrainedTokenizer from ...utils import logging _a = logging.get_logger(__name__) _a = '▁' _a = { 'vocab_file': 'vocab.json', 'spm_file': 'sentencepiece.bpe.model', 'tokenizer_config_file': 'tokenizer_config.json', } _a = { 'vocab_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/vocab.json', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/vocab.json', }, 'spm_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/sentencepiece.bpe.model', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/sentencepiece.bpe.model', }, 'tokenizer_config_file': { 'facebook/m2m100_418M': 'https://huggingface.co/facebook/m2m100_418M/resolve/main/tokenizer_config.json', 'facebook/m2m100_1.2B': 'https://huggingface.co/facebook/m2m100_1.2B/resolve/main/tokenizer_config.json', }, } _a = { 'facebook/m2m100_418M': 1_024, } # fmt: off _a = { 'm2m100': ['af', 'am', 'ar', 'ast', 'az', 'ba', 'be', 'bg', 'bn', 'br', 'bs', 'ca', 'ceb', 'cs', 'cy', 'da', 'de', 'el', 'en', 'es', 'et', 'fa', 'ff', 'fi', 'fr', 'fy', 'ga', 'gd', 'gl', 'gu', 'ha', 'he', 'hi', 'hr', 'ht', 'hu', 'hy', 'id', 'ig', 'ilo', 'is', 'it', 'ja', 'jv', 'ka', 'kk', 'km', 'kn', 'ko', 'lb', 'lg', 'ln', 'lo', 'lt', 'lv', 'mg', 'mk', 'ml', 'mn', 'mr', 'ms', 'my', 'ne', 'nl', 'no', 'ns', 'oc', 'or', 'pa', 'pl', 'ps', 'pt', 'ro', 'ru', 'sd', 'si', 'sk', 'sl', 'so', 'sq', 'sr', 'ss', 'su', 'sv', 'sw', 'ta', 'th', 'tl', 'tn', 'tr', 'uk', 'ur', 'uz', 'vi', 'wo', 'xh', 'yi', 'yo', 'zh', 'zu'], 'wmt21': ['en', 'ha', 'is', 'ja', 'cs', 'ru', 'zh', 'de'] } class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : str = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : int = ["""input_ids""", """attention_mask"""] SCREAMING_SNAKE_CASE__ : List[int] = [] SCREAMING_SNAKE_CASE__ : List[int] = [] def __init__( self , lowercase_ , lowercase_ , lowercase_=None , lowercase_=None , lowercase_="<s>" , lowercase_="</s>" , lowercase_="</s>" , lowercase_="<pad>" , lowercase_="<unk>" , lowercase_="m2m100" , lowercase_ = None , lowercase_=8 , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Dict = {} if sp_model_kwargs is None else sp_model_kwargs UpperCAmelCase_ : List[Any] = language_codes UpperCAmelCase_ : Any = FAIRSEQ_LANGUAGE_CODES[language_codes] UpperCAmelCase_ : Optional[int] = {lang_code: F"""__{lang_code}__""" for lang_code in fairseq_language_code} UpperCAmelCase_ : List[str] = kwargs.get("additional_special_tokens" , [] ) kwargs["additional_special_tokens"] += [ self.get_lang_token(lowercase_ ) for lang_code in fairseq_language_code if self.get_lang_token(lowercase_ ) not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=lowercase_ , tgt_lang=lowercase_ , bos_token=lowercase_ , eos_token=lowercase_ , sep_token=lowercase_ , unk_token=lowercase_ , pad_token=lowercase_ , language_codes=lowercase_ , sp_model_kwargs=self.sp_model_kwargs , num_madeup_words=lowercase_ , **lowercase_ , ) UpperCAmelCase_ : Dict = vocab_file UpperCAmelCase_ : str = load_json(lowercase_ ) UpperCAmelCase_ : Dict = {v: k for k, v in self.encoder.items()} UpperCAmelCase_ : str = spm_file UpperCAmelCase_ : Union[str, Any] = load_spm(lowercase_ , self.sp_model_kwargs ) UpperCAmelCase_ : List[str] = len(self.encoder ) UpperCAmelCase_ : Optional[int] = { self.get_lang_token(lowercase_ ): self.encoder_size + i for i, lang_code in enumerate(lowercase_ ) } UpperCAmelCase_ : str = {lang_code: self.encoder_size + i for i, lang_code in enumerate(lowercase_ )} UpperCAmelCase_ : Dict = {v: k for k, v in self.lang_token_to_id.items()} UpperCAmelCase_ : Optional[Any] = src_lang if src_lang is not None else "en" UpperCAmelCase_ : str = tgt_lang UpperCAmelCase_ : Any = self.get_lang_id(self._src_lang ) self.set_src_lang_special_tokens(self._src_lang ) UpperCAmelCase_ : Union[str, Any] = num_madeup_words @property def UpperCamelCase__ ( self ): """simple docstring""" return len(self.encoder ) + len(self.lang_token_to_id ) @property def UpperCamelCase__ ( self ): """simple docstring""" return self._src_lang @src_lang.setter def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Tuple = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.sp_model.encode(lowercase_ , out_type=lowercase_ ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" if token in self.lang_token_to_id: return self.lang_token_to_id[token] return self.encoder.get(lowercase_ , self.encoder[self.unk_token] ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" if index in self.id_to_lang_token: return self.id_to_lang_token[index] return self.decoder.get(lowercase_ , self.unk_token ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[str] = [] UpperCAmelCase_ : int = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(lowercase_ ) + token UpperCAmelCase_ : Optional[Any] = [] else: current_sub_tokens.append(lowercase_ ) out_string += self.sp_model.decode(lowercase_ ) return out_string.strip() def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowercase_ , token_ids_a=lowercase_ , already_has_special_tokens=lowercase_ ) UpperCAmelCase_ : Tuple = [1] * len(self.prefix_tokens ) UpperCAmelCase_ : Union[str, Any] = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(lowercase_ )) + suffix_ones return prefix_ones + ([0] * len(lowercase_ )) + ([0] * len(lowercase_ )) + suffix_ones def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Any = {self.convert_ids_to_tokens(lowercase_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.__dict__.copy() UpperCAmelCase_ : List[str] = None return state def __setstate__( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : int = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): UpperCAmelCase_ : Union[str, Any] = {} UpperCAmelCase_ : Optional[Any] = load_spm(self.spm_file , self.sp_model_kwargs ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" UpperCAmelCase_ : Any = Path(lowercase_ ) if not save_dir.is_dir(): raise OSError(F"""{save_directory} should be a directory""" ) UpperCAmelCase_ : List[str] = save_dir / ( (filename_prefix + "-" if filename_prefix else "") + self.vocab_files_names["vocab_file"] ) UpperCAmelCase_ : List[str] = save_dir / ( (filename_prefix + "-" if filename_prefix else "") + self.vocab_files_names["spm_file"] ) save_json(self.encoder , lowercase_ ) if os.path.abspath(self.spm_file ) != os.path.abspath(lowercase_ ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , lowercase_ ) elif not os.path.isfile(self.spm_file ): with open(lowercase_ , "wb" ) as fi: UpperCAmelCase_ : int = self.sp_model.serialized_model_proto() fi.write(lowercase_ ) return (str(lowercase_ ), str(lowercase_ )) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = "en" , lowercase_ = None , lowercase_ = "ro" , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Optional[int] = src_lang UpperCAmelCase_ : Union[str, Any] = tgt_lang self.set_src_lang_special_tokens(self.src_lang ) return super().prepare_seqaseq_batch(lowercase_ , lowercase_ , **lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) UpperCAmelCase_ : Union[str, Any] = src_lang UpperCAmelCase_ : Union[str, Any] = self(lowercase_ , add_special_tokens=lowercase_ , **lowercase_ ) UpperCAmelCase_ : Dict = self.get_lang_id(lowercase_ ) UpperCAmelCase_ : Tuple = tgt_lang_id return inputs def UpperCamelCase__ ( self ): """simple docstring""" self.set_src_lang_special_tokens(self.src_lang ) def UpperCamelCase__ ( self ): """simple docstring""" self.set_tgt_lang_special_tokens(self.tgt_lang ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Dict = self.get_lang_token(lowercase_ ) UpperCAmelCase_ : Union[str, Any] = self.lang_token_to_id[lang_token] UpperCAmelCase_ : Any = [self.cur_lang_id] UpperCAmelCase_ : int = [self.eos_token_id] def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = self.get_lang_token(lowercase_ ) UpperCAmelCase_ : Optional[Any] = self.lang_token_to_id[lang_token] UpperCAmelCase_ : Dict = [self.cur_lang_id] UpperCAmelCase_ : int = [self.eos_token_id] def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.lang_code_to_token[lang] def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = self.get_lang_token(lowercase_ ) return self.lang_token_to_id[lang_token] def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : List[str] = sentencepiece.SentencePieceProcessor(**__lowerCamelCase ) spm.Load(str(__lowerCamelCase ) ) return spm def __a ( __lowerCamelCase ): with open(__lowerCamelCase, "r" ) as f: return json.load(__lowerCamelCase ) def __a ( __lowerCamelCase, __lowerCamelCase ): with open(__lowerCamelCase, "w" ) as f: json.dump(__lowerCamelCase, __lowerCamelCase, indent=2 )
23
"""simple docstring""" import datasets _a = '\\n@InProceedings{conneau2018xnli,\n author = "Conneau, Alexis\n and Rinott, Ruty\n and Lample, Guillaume\n and Williams, Adina\n and Bowman, Samuel R.\n and Schwenk, Holger\n and Stoyanov, Veselin",\n title = "XNLI: Evaluating Cross-lingual Sentence Representations",\n booktitle = "Proceedings of the 2018 Conference on Empirical Methods\n in Natural Language Processing",\n year = "2018",\n publisher = "Association for Computational Linguistics",\n location = "Brussels, Belgium",\n}\n' _a = '\\nXNLI is a subset of a few thousand examples from MNLI which has been translated\ninto a 14 different languages (some low-ish resource). As with MNLI, the goal is\nto predict textual entailment (does sentence A imply/contradict/neither sentence\nB) and is a classification task (given two sentences, predict one of three\nlabels).\n' _a = '\nComputes XNLI score which is just simple accuracy.\nArgs:\n predictions: Predicted labels.\n references: Ground truth labels.\nReturns:\n \'accuracy\': accuracy\nExamples:\n\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> xnli_metric = datasets.load_metric("xnli")\n >>> results = xnli_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n' def __a ( __lowerCamelCase, __lowerCamelCase ): return (preds == labels).mean() @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64" if self.config_name != "sts-b" else "float32" ), "references": datasets.Value("int64" if self.config_name != "sts-b" else "float32" ), } ) , codebase_urls=[] , reference_urls=[] , format="numpy" , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" return {"accuracy": simple_accuracy(lowercase_ , lowercase_ )}
23
1
"""simple docstring""" def __a ( __lowerCamelCase ): if num < 0: return False UpperCAmelCase_ : int = num UpperCAmelCase_ : int = 0 while num > 0: UpperCAmelCase_ : Dict = rev_num * 10 + (num % 10) num //= 10 return num_copy == rev_num if __name__ == "__main__": import doctest doctest.testmod()
23
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy _a = logging.get_logger(__name__) class A_ (lowercase__ ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = feature_size UpperCAmelCase_ : Any = sampling_rate UpperCAmelCase_ : Any = padding_value UpperCAmelCase_ : str = kwargs.pop("padding_side" , "right" ) UpperCAmelCase_ : List[str] = kwargs.pop("return_attention_mask" , lowercase_ ) super().__init__(**lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = True , lowercase_ = None , lowercase_ = False , lowercase_ = None , lowercase_ = None , lowercase_ = None , ): """simple docstring""" # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(lowercase_ , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): UpperCAmelCase_ : Dict = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( "You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`" F""" to this method that includes {self.model_input_names[0]}, but you provided""" F""" {list(processed_features.keys() )}""" ) UpperCAmelCase_ : Tuple = processed_features[self.model_input_names[0]] UpperCAmelCase_ : List[str] = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(lowercase_ ) == 0: if return_attention_mask: UpperCAmelCase_ : Union[str, Any] = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch UpperCAmelCase_ : List[str] = required_input[0] if isinstance(lowercase_ , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. UpperCAmelCase_ : Any = 0 while len(required_input[index] ) == 0: index += 1 if index < len(lowercase_ ): UpperCAmelCase_ : Optional[Any] = required_input[index][0] if return_tensors is None: if is_tf_tensor(lowercase_ ): UpperCAmelCase_ : Dict = "tf" elif is_torch_tensor(lowercase_ ): UpperCAmelCase_ : Any = "pt" elif isinstance(lowercase_ , (int, float, list, tuple, np.ndarray) ): UpperCAmelCase_ : str = "np" else: raise ValueError( F"""type of {first_element} unknown: {type(lowercase_ )}. """ "Should be one of a python, numpy, pytorch or tensorflow object." ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): UpperCAmelCase_ : Optional[int] = to_numpy(lowercase_ ) else: UpperCAmelCase_ : List[str] = [to_numpy(lowercase_ ) for v in value] # Convert padding_strategy in PaddingStrategy UpperCAmelCase_ : Dict = self._get_padding_strategies(padding=lowercase_ , max_length=lowercase_ ) UpperCAmelCase_ : str = processed_features[self.model_input_names[0]] UpperCAmelCase_ : int = len(lowercase_ ) if not all(len(lowercase_ ) == batch_size for v in processed_features.values() ): raise ValueError("Some items in the output dictionary have a different batch size than others." ) UpperCAmelCase_ : int = [] for i in range(lowercase_ ): UpperCAmelCase_ : str = {k: v[i] for k, v in processed_features.items()} # truncation UpperCAmelCase_ : List[str] = self._truncate( lowercase_ , max_length=lowercase_ , pad_to_multiple_of=lowercase_ , truncation=lowercase_ , ) truncated_inputs.append(lowercase_ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length UpperCAmelCase_ : str = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) UpperCAmelCase_ : Dict = PaddingStrategy.MAX_LENGTH UpperCAmelCase_ : List[str] = {} for i in range(lowercase_ ): # padding UpperCAmelCase_ : int = self._pad( truncated_inputs[i] , max_length=lowercase_ , padding_strategy=lowercase_ , pad_to_multiple_of=lowercase_ , return_attention_mask=lowercase_ , ) for key, value in outputs.items(): if key not in batch_outputs: UpperCAmelCase_ : Any = [] if value.dtype is np.dtype(np.floataa ): UpperCAmelCase_ : List[Any] = value.astype(np.floataa ) batch_outputs[key].append(lowercase_ ) return BatchFeature(lowercase_ , tensor_type=lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = PaddingStrategy.DO_NOT_PAD , lowercase_ = None , lowercase_ = None , ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: UpperCAmelCase_ : Tuple = len(lowercase_ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCAmelCase_ : Tuple = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCAmelCase_ : Dict = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(lowercase_ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: UpperCAmelCase_ : Optional[int] = np.ones(len(lowercase_ ) , dtype=np.intaa ) if needs_to_be_padded: UpperCAmelCase_ : Dict = max_length - len(lowercase_ ) if self.padding_side == "right": if return_attention_mask: UpperCAmelCase_ : List[Any] = np.pad( processed_features["attention_mask"] , (0, difference) ) UpperCAmelCase_ : Dict = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) UpperCAmelCase_ : Optional[Any] = np.pad( lowercase_ , lowercase_ , "constant" , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: UpperCAmelCase_ : Optional[Any] = np.pad( processed_features["attention_mask"] , (difference, 0) ) UpperCAmelCase_ : Dict = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) UpperCAmelCase_ : str = np.pad( lowercase_ , lowercase_ , "constant" , constant_values=self.padding_value ) else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return processed_features def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , ): """simple docstring""" if not truncation: return processed_features elif truncation and max_length is None: raise ValueError("When setting ``truncation=True``, make sure that ``max_length`` is defined." ) UpperCAmelCase_ : Optional[int] = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCAmelCase_ : Union[str, Any] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCAmelCase_ : Optional[Any] = len(lowercase_ ) > max_length if needs_to_be_truncated: UpperCAmelCase_ : int = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: UpperCAmelCase_ : Dict = processed_features["attention_mask"][:max_length] return processed_features def UpperCamelCase__ ( self , lowercase_=False , lowercase_=None ): """simple docstring""" # Get padding strategy if padding is not False: if padding is True: UpperCAmelCase_ : Dict = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : Optional[Any] = PaddingStrategy(lowercase_ ) elif isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = padding else: UpperCAmelCase_ : str = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F"""When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined""" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( "Asking to pad but the feature_extractor does not have a padding value. Please select a value to use" " as `padding_value`. For example: `feature_extractor.padding_value = 0.0`." ) return padding_strategy
23
1
"""simple docstring""" from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxCrossAttnUpBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, FlaxUpBlockaD, ) @flax.struct.dataclass class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : jnp.ndarray @flax_register_to_config class A_ (nn.Module ,lowercase__ ,lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = 32 SCREAMING_SNAKE_CASE__ : int = 4 SCREAMING_SNAKE_CASE__ : int = 4 SCREAMING_SNAKE_CASE__ : Tuple[str] = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) SCREAMING_SNAKE_CASE__ : Tuple[str] = ("UpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D", "CrossAttnUpBlock2D") SCREAMING_SNAKE_CASE__ : Union[bool, Tuple[bool]] = False SCREAMING_SNAKE_CASE__ : Tuple[int] = (320, 640, 1280, 1280) SCREAMING_SNAKE_CASE__ : int = 2 SCREAMING_SNAKE_CASE__ : Union[int, Tuple[int]] = 8 SCREAMING_SNAKE_CASE__ : Optional[Union[int, Tuple[int]]] = None SCREAMING_SNAKE_CASE__ : int = 1280 SCREAMING_SNAKE_CASE__ : float = 0.0 SCREAMING_SNAKE_CASE__ : bool = False SCREAMING_SNAKE_CASE__ : jnp.dtype = jnp.floataa SCREAMING_SNAKE_CASE__ : bool = True SCREAMING_SNAKE_CASE__ : int = 0 SCREAMING_SNAKE_CASE__ : bool = False def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" # init input tensors UpperCAmelCase_ : Optional[int] = (1, self.in_channels, self.sample_size, self.sample_size) UpperCAmelCase_ : Optional[Any] = jnp.zeros(lowercase_ , dtype=jnp.floataa ) UpperCAmelCase_ : Dict = jnp.ones((1,) , dtype=jnp.intaa ) UpperCAmelCase_ : List[Any] = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = jax.random.split(lowercase_ ) UpperCAmelCase_ : Optional[int] = {"params": params_rng, "dropout": dropout_rng} return self.init(lowercase_ , lowercase_ , lowercase_ , lowercase_ )["params"] def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Any = self.block_out_channels UpperCAmelCase_ : List[str] = block_out_channels[0] * 4 if self.num_attention_heads is not None: raise ValueError( "At the moment it is not possible to define the number of attention heads via `num_attention_heads` because of a naming issue as described in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131. Passing `num_attention_heads` will only be supported in diffusers v0.19." ) # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. UpperCAmelCase_ : Optional[Any] = self.num_attention_heads or self.attention_head_dim # input UpperCAmelCase_ : List[str] = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time UpperCAmelCase_ : Tuple = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) UpperCAmelCase_ : Tuple = FlaxTimestepEmbedding(lowercase_ , dtype=self.dtype ) UpperCAmelCase_ : Union[str, Any] = self.only_cross_attention if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : Optional[int] = (only_cross_attention,) * len(self.down_block_types ) if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : List[Any] = (num_attention_heads,) * len(self.down_block_types ) # down UpperCAmelCase_ : int = [] UpperCAmelCase_ : Any = block_out_channels[0] for i, down_block_type in enumerate(self.down_block_types ): UpperCAmelCase_ : List[Any] = output_channel UpperCAmelCase_ : str = block_out_channels[i] UpperCAmelCase_ : Optional[Any] = i == len(lowercase_ ) - 1 if down_block_type == "CrossAttnDownBlock2D": UpperCAmelCase_ : Union[str, Any] = FlaxCrossAttnDownBlockaD( in_channels=lowercase_ , out_channels=lowercase_ , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: UpperCAmelCase_ : List[Any] = FlaxDownBlockaD( in_channels=lowercase_ , out_channels=lowercase_ , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(lowercase_ ) UpperCAmelCase_ : Any = down_blocks # mid UpperCAmelCase_ : Any = FlaxUNetMidBlockaDCrossAttn( in_channels=block_out_channels[-1] , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) # up UpperCAmelCase_ : Union[str, Any] = [] UpperCAmelCase_ : Union[str, Any] = list(reversed(lowercase_ ) ) UpperCAmelCase_ : Dict = list(reversed(lowercase_ ) ) UpperCAmelCase_ : Any = list(reversed(lowercase_ ) ) UpperCAmelCase_ : List[str] = reversed_block_out_channels[0] for i, up_block_type in enumerate(self.up_block_types ): UpperCAmelCase_ : Optional[int] = output_channel UpperCAmelCase_ : Tuple = reversed_block_out_channels[i] UpperCAmelCase_ : str = reversed_block_out_channels[min(i + 1 , len(lowercase_ ) - 1 )] UpperCAmelCase_ : Any = i == len(lowercase_ ) - 1 if up_block_type == "CrossAttnUpBlock2D": UpperCAmelCase_ : List[Any] = FlaxCrossAttnUpBlockaD( in_channels=lowercase_ , out_channels=lowercase_ , prev_output_channel=lowercase_ , num_layers=self.layers_per_block + 1 , num_attention_heads=reversed_num_attention_heads[i] , add_upsample=not is_final_block , dropout=self.dropout , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) else: UpperCAmelCase_ : Optional[Any] = FlaxUpBlockaD( in_channels=lowercase_ , out_channels=lowercase_ , prev_output_channel=lowercase_ , num_layers=self.layers_per_block + 1 , add_upsample=not is_final_block , dropout=self.dropout , dtype=self.dtype , ) up_blocks.append(lowercase_ ) UpperCAmelCase_ : Tuple = output_channel UpperCAmelCase_ : Optional[int] = up_blocks # out UpperCAmelCase_ : Dict = nn.GroupNorm(num_groups=32 , epsilon=1E-5 ) UpperCAmelCase_ : Tuple = nn.Conv( self.out_channels , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) def __call__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_=None , lowercase_=None , lowercase_ = True , lowercase_ = False , ): """simple docstring""" # 1. time if not isinstance(lowercase_ , jnp.ndarray ): UpperCAmelCase_ : List[str] = jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(lowercase_ , jnp.ndarray ) and len(timesteps.shape ) == 0: UpperCAmelCase_ : Any = timesteps.astype(dtype=jnp.floataa ) UpperCAmelCase_ : Tuple = jnp.expand_dims(lowercase_ , 0 ) UpperCAmelCase_ : str = self.time_proj(lowercase_ ) UpperCAmelCase_ : str = self.time_embedding(lowercase_ ) # 2. pre-process UpperCAmelCase_ : int = jnp.transpose(lowercase_ , (0, 2, 3, 1) ) UpperCAmelCase_ : List[Any] = self.conv_in(lowercase_ ) # 3. down UpperCAmelCase_ : Tuple = (sample,) for down_block in self.down_blocks: if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = down_block(lowercase_ , lowercase_ , lowercase_ , deterministic=not train ) else: UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = down_block(lowercase_ , lowercase_ , deterministic=not train ) down_block_res_samples += res_samples if down_block_additional_residuals is not None: UpperCAmelCase_ : int = () for down_block_res_sample, down_block_additional_residual in zip( lowercase_ , lowercase_ ): down_block_res_sample += down_block_additional_residual new_down_block_res_samples += (down_block_res_sample,) UpperCAmelCase_ : Dict = new_down_block_res_samples # 4. mid UpperCAmelCase_ : Any = self.mid_block(lowercase_ , lowercase_ , lowercase_ , deterministic=not train ) if mid_block_additional_residual is not None: sample += mid_block_additional_residual # 5. up for up_block in self.up_blocks: UpperCAmelCase_ : Dict = down_block_res_samples[-(self.layers_per_block + 1) :] UpperCAmelCase_ : str = down_block_res_samples[: -(self.layers_per_block + 1)] if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : Union[str, Any] = up_block( lowercase_ , temb=lowercase_ , encoder_hidden_states=lowercase_ , res_hidden_states_tuple=lowercase_ , deterministic=not train , ) else: UpperCAmelCase_ : Optional[int] = up_block(lowercase_ , temb=lowercase_ , res_hidden_states_tuple=lowercase_ , deterministic=not train ) # 6. post-process UpperCAmelCase_ : Tuple = self.conv_norm_out(lowercase_ ) UpperCAmelCase_ : List[Any] = nn.silu(lowercase_ ) UpperCAmelCase_ : Tuple = self.conv_out(lowercase_ ) UpperCAmelCase_ : Any = jnp.transpose(lowercase_ , (0, 3, 1, 2) ) if not return_dict: return (sample,) return FlaxUNetaDConditionOutput(sample=lowercase_ )
23
"""simple docstring""" import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = torch.nn.Linear(10 , 10 ) UpperCAmelCase_ : List[str] = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase_ : Optional[Any] = Accelerator() UpperCAmelCase_ : Tuple = accelerator.prepare(lowercase_ ) try: pickle.loads(pickle.dumps(lowercase_ ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
23
1
"""simple docstring""" from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _a = logging.get_logger(__name__) # pylint: disable=invalid-name _a = '\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained("kandinsky-community/kandinsky-2-2-prior")\n >>> pipe_prior.to("cuda")\n >>> prompt = "red cat, 4k photo"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained("kandinsky-community/kandinsky-2-2-decoder")\n >>> pipe.to("cuda")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save("cat.png")\n ```\n' def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=8 ): UpperCAmelCase_ : List[str] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 UpperCAmelCase_ : Tuple = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class A_ (lowercase__ ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , ): """simple docstring""" super().__init__() self.register_modules( unet=lowercase_ , scheduler=lowercase_ , movq=lowercase_ , ) UpperCAmelCase_ : int = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" if latents is None: UpperCAmelCase_ : Dict = randn_tensor(lowercase_ , generator=lowercase_ , device=lowercase_ , dtype=lowercase_ ) else: if latents.shape != shape: raise ValueError(F"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) UpperCAmelCase_ : str = latents.to(lowercase_ ) UpperCAmelCase_ : Dict = latents * scheduler.init_noise_sigma return latents def UpperCamelCase__ ( self , lowercase_=0 ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) UpperCAmelCase_ : Any = torch.device(F"""cuda:{gpu_id}""" ) UpperCAmelCase_ : int = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowercase_ , lowercase_ ) def UpperCamelCase__ ( self , lowercase_=0 ): """simple docstring""" if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) UpperCAmelCase_ : Any = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=lowercase_ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) UpperCAmelCase_ : List[Any] = None for cpu_offloaded_model in [self.unet, self.movq]: UpperCAmelCase_ , UpperCAmelCase_ : str = cpu_offload_with_hook(lowercase_ , lowercase_ , prev_module_hook=lowercase_ ) # We'll offload the last model manually. UpperCAmelCase_ : Tuple = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCamelCase__ ( self ): """simple docstring""" if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowercase_ , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(lowercase_ ) def __call__( self , lowercase_ , lowercase_ , lowercase_ = 512 , lowercase_ = 512 , lowercase_ = 100 , lowercase_ = 4.0 , lowercase_ = 1 , lowercase_ = None , lowercase_ = None , lowercase_ = "pil" , lowercase_ = True , ): """simple docstring""" UpperCAmelCase_ : str = self._execution_device UpperCAmelCase_ : List[Any] = guidance_scale > 1.0 if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = torch.cat(lowercase_ , dim=0 ) UpperCAmelCase_ : Any = image_embeds.shape[0] * num_images_per_prompt if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : List[Any] = torch.cat(lowercase_ , dim=0 ) if do_classifier_free_guidance: UpperCAmelCase_ : Tuple = image_embeds.repeat_interleave(lowercase_ , dim=0 ) UpperCAmelCase_ : List[str] = negative_image_embeds.repeat_interleave(lowercase_ , dim=0 ) UpperCAmelCase_ : Optional[Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=lowercase_ ) self.scheduler.set_timesteps(lowercase_ , device=lowercase_ ) UpperCAmelCase_ : List[Any] = self.scheduler.timesteps UpperCAmelCase_ : List[str] = self.unet.config.in_channels UpperCAmelCase_ , UpperCAmelCase_ : Dict = downscale_height_and_width(lowercase_ , lowercase_ , self.movq_scale_factor ) # create initial latent UpperCAmelCase_ : int = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , lowercase_ , lowercase_ , lowercase_ , self.scheduler , ) for i, t in enumerate(self.progress_bar(lowercase_ ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase_ : Tuple = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase_ : Union[str, Any] = {"image_embeds": image_embeds} UpperCAmelCase_ : Optional[Any] = self.unet( sample=lowercase_ , timestep=lowercase_ , encoder_hidden_states=lowercase_ , added_cond_kwargs=lowercase_ , return_dict=lowercase_ , )[0] if do_classifier_free_guidance: UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = noise_pred.split(latents.shape[1] , dim=1 ) UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = noise_pred.chunk(2 ) UpperCAmelCase_ , UpperCAmelCase_ : Tuple = variance_pred.chunk(2 ) UpperCAmelCase_ : int = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) UpperCAmelCase_ : str = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): UpperCAmelCase_ , UpperCAmelCase_ : Dict = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase_ : List[str] = self.scheduler.step( lowercase_ , lowercase_ , lowercase_ , generator=lowercase_ , )[0] # post-processing UpperCAmelCase_ : Tuple = self.movq.decode(lowercase_ , force_not_quantize=lowercase_ )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: UpperCAmelCase_ : List[Any] = image * 0.5 + 0.5 UpperCAmelCase_ : int = image.clamp(0 , 1 ) UpperCAmelCase_ : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": UpperCAmelCase_ : Dict = self.numpy_to_pil(lowercase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowercase_ )
23
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = {'ctrl': 'https://huggingface.co/ctrl/resolve/main/config.json'} class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """ctrl""" SCREAMING_SNAKE_CASE__ : Optional[int] = ["""past_key_values"""] SCREAMING_SNAKE_CASE__ : List[str] = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase_=24_6534 , lowercase_=256 , lowercase_=1280 , lowercase_=8192 , lowercase_=48 , lowercase_=16 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=1E-6 , lowercase_=0.02 , lowercase_=True , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Tuple = vocab_size UpperCAmelCase_ : Union[str, Any] = n_positions UpperCAmelCase_ : List[str] = n_embd UpperCAmelCase_ : Dict = n_layer UpperCAmelCase_ : Optional[int] = n_head UpperCAmelCase_ : List[str] = dff UpperCAmelCase_ : Tuple = resid_pdrop UpperCAmelCase_ : Optional[Any] = embd_pdrop UpperCAmelCase_ : str = layer_norm_epsilon UpperCAmelCase_ : List[str] = initializer_range UpperCAmelCase_ : List[str] = use_cache super().__init__(**lowercase_ )
23
1
"""simple docstring""" from __future__ import annotations _a = 10 def __a ( __lowerCamelCase ): UpperCAmelCase_ : Any = 1 UpperCAmelCase_ : Any = max(__lowerCamelCase ) while placement <= max_digit: # declare and initialize empty buckets UpperCAmelCase_ : list[list] = [[] for _ in range(__lowerCamelCase )] # split list_of_ints between the buckets for i in list_of_ints: UpperCAmelCase_ : int = int((i / placement) % RADIX ) buckets[tmp].append(__lowerCamelCase ) # put each buckets' contents into list_of_ints UpperCAmelCase_ : Any = 0 for b in range(__lowerCamelCase ): for i in buckets[b]: UpperCAmelCase_ : str = i a += 1 # move to next placement *= RADIX return list_of_ints if __name__ == "__main__": import doctest doctest.testmod()
23
"""simple docstring""" def __a ( __lowerCamelCase ): assert isinstance(__lowerCamelCase, __lowerCamelCase ), f"""The input value of [n={number}] is not an integer""" if number == 1: return 2 elif number < 1: UpperCAmelCase_ : str = f"""The input value of [n={number}] has to be > 0""" raise ValueError(__lowerCamelCase ) else: UpperCAmelCase_ : List[str] = sylvester(number - 1 ) UpperCAmelCase_ : List[str] = num - 1 UpperCAmelCase_ : List[str] = num return lower * upper + 1 if __name__ == "__main__": print(f"""The 8th number in Sylvester's sequence: {sylvester(8)}""")
23
1
"""simple docstring""" from __future__ import annotations from typing import Any def __a ( __lowerCamelCase ): create_state_space_tree(__lowerCamelCase, [], 0 ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): if index == len(__lowerCamelCase ): print(__lowerCamelCase ) return create_state_space_tree(__lowerCamelCase, __lowerCamelCase, index + 1 ) current_subsequence.append(sequence[index] ) create_state_space_tree(__lowerCamelCase, __lowerCamelCase, index + 1 ) current_subsequence.pop() if __name__ == "__main__": _a = [3, 1, 2, 4] generate_all_subsequences(seq) seq.clear() seq.extend(['A', 'B', 'C']) generate_all_subsequences(seq)
23
"""simple docstring""" import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class A_ (lowercase__ ,lowercase__ ,unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = IFImgaImgSuperResolutionPipeline SCREAMING_SNAKE_CASE__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""width""", """height"""} SCREAMING_SNAKE_CASE__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""original_image"""} ) SCREAMING_SNAKE_CASE__ : List[Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} def UpperCamelCase__ ( self ): """simple docstring""" return self._get_superresolution_dummy_components() def UpperCamelCase__ ( self , lowercase_ , lowercase_=0 ): """simple docstring""" if str(lowercase_ ).startswith("mps" ): UpperCAmelCase_ : Optional[Any] = torch.manual_seed(lowercase_ ) else: UpperCAmelCase_ : Union[str, Any] = torch.Generator(device=lowercase_ ).manual_seed(lowercase_ ) UpperCAmelCase_ : Any = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) UpperCAmelCase_ : Optional[int] = floats_tensor((1, 3, 16, 16) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) UpperCAmelCase_ : int = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def UpperCamelCase__ ( self ): """simple docstring""" # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_save_load_local() def UpperCamelCase__ ( self ): """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
23
1
"""simple docstring""" import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.alt_diffusion.modeling_roberta_series import ( RobertaSeriesConfig, RobertaSeriesModelWithTransformation, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class A_ (lowercase__ ,lowercase__ ,lowercase__ ,unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = AltDiffusionPipeline SCREAMING_SNAKE_CASE__ : str = TEXT_TO_IMAGE_PARAMS SCREAMING_SNAKE_CASE__ : Tuple = TEXT_TO_IMAGE_BATCH_PARAMS SCREAMING_SNAKE_CASE__ : List[Any] = TEXT_TO_IMAGE_IMAGE_PARAMS SCREAMING_SNAKE_CASE__ : Optional[Any] = TEXT_TO_IMAGE_IMAGE_PARAMS def UpperCamelCase__ ( self ): """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase_ : Tuple = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , ) UpperCAmelCase_ : str = DDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="scaled_linear" , clip_sample=lowercase_ , set_alpha_to_one=lowercase_ , ) torch.manual_seed(0 ) UpperCAmelCase_ : int = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) # TODO: address the non-deterministic text encoder (fails for save-load tests) # torch.manual_seed(0) # text_encoder_config = RobertaSeriesConfig( # hidden_size=32, # project_dim=32, # intermediate_size=37, # layer_norm_eps=1e-05, # num_attention_heads=4, # num_hidden_layers=5, # vocab_size=5002, # ) # text_encoder = RobertaSeriesModelWithTransformation(text_encoder_config) torch.manual_seed(0 ) UpperCAmelCase_ : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , projection_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=5002 , ) UpperCAmelCase_ : Tuple = CLIPTextModel(lowercase_ ) UpperCAmelCase_ : Any = XLMRobertaTokenizer.from_pretrained("hf-internal-testing/tiny-xlm-roberta" ) UpperCAmelCase_ : Tuple = 77 UpperCAmelCase_ : Tuple = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCamelCase__ ( self , lowercase_ , lowercase_=0 ): """simple docstring""" if str(lowercase_ ).startswith("mps" ): UpperCAmelCase_ : List[Any] = torch.manual_seed(lowercase_ ) else: UpperCAmelCase_ : Dict = torch.Generator(device=lowercase_ ).manual_seed(lowercase_ ) UpperCAmelCase_ : Tuple = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def UpperCamelCase__ ( self ): """simple docstring""" super().test_attention_slicing_forward_pass(expected_max_diff=3E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ : Tuple = self.get_dummy_components() torch.manual_seed(0 ) UpperCAmelCase_ : Dict = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase_ : Optional[Any] = RobertaSeriesModelWithTransformation(lowercase_ ) UpperCAmelCase_ : int = text_encoder UpperCAmelCase_ : Optional[Any] = AltDiffusionPipeline(**lowercase_ ) UpperCAmelCase_ : Dict = alt_pipe.to(lowercase_ ) alt_pipe.set_progress_bar_config(disable=lowercase_ ) UpperCAmelCase_ : Optional[Any] = self.get_dummy_inputs(lowercase_ ) UpperCAmelCase_ : Dict = "A photo of an astronaut" UpperCAmelCase_ : List[Any] = alt_pipe(**lowercase_ ) UpperCAmelCase_ : int = output.images UpperCAmelCase_ : Any = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase_ : List[Any] = np.array( [0.5_74_81_62, 0.60_44_71_45, 0.48_82_12_17, 0.50_10_06_36, 0.5_43_11_85, 0.45_76_36_83, 0.49_65_76_96, 0.48_13_27_33, 0.47_57_30_93] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[str] = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase_ : Any = self.get_dummy_components() UpperCAmelCase_ : str = PNDMScheduler(skip_prk_steps=lowercase_ ) torch.manual_seed(0 ) UpperCAmelCase_ : Union[str, Any] = RobertaSeriesConfig( hidden_size=32 , project_dim=32 , intermediate_size=37 , layer_norm_eps=1E-0_5 , num_attention_heads=4 , num_hidden_layers=5 , vocab_size=5002 , ) # TODO: remove after fixing the non-deterministic text encoder UpperCAmelCase_ : Dict = RobertaSeriesModelWithTransformation(lowercase_ ) UpperCAmelCase_ : Dict = text_encoder UpperCAmelCase_ : Optional[int] = AltDiffusionPipeline(**lowercase_ ) UpperCAmelCase_ : Any = alt_pipe.to(lowercase_ ) alt_pipe.set_progress_bar_config(disable=lowercase_ ) UpperCAmelCase_ : int = self.get_dummy_inputs(lowercase_ ) UpperCAmelCase_ : Optional[Any] = alt_pipe(**lowercase_ ) UpperCAmelCase_ : int = output.images UpperCAmelCase_ : List[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCAmelCase_ : List[Any] = np.array( [0.51_60_50_93, 0.5_70_72_41, 0.47_36_55_07, 0.50_57_88_86, 0.5_63_38_77, 0.4_64_25_03, 0.5_18_20_81, 0.48_76_34_84, 0.49_08_42_37] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch_gpu class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCamelCase__ ( self ): """simple docstring""" # make sure here that pndm scheduler skips prk UpperCAmelCase_ : List[str] = AltDiffusionPipeline.from_pretrained("BAAI/AltDiffusion" , safety_checker=lowercase_ ) UpperCAmelCase_ : Union[str, Any] = alt_pipe.to(lowercase_ ) alt_pipe.set_progress_bar_config(disable=lowercase_ ) UpperCAmelCase_ : int = "A painting of a squirrel eating a burger" UpperCAmelCase_ : Any = torch.manual_seed(0 ) UpperCAmelCase_ : Optional[int] = alt_pipe([prompt] , generator=lowercase_ , guidance_scale=6.0 , num_inference_steps=20 , output_type="np" ) UpperCAmelCase_ : List[Any] = output.images UpperCAmelCase_ : List[str] = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase_ : Union[str, Any] = np.array([0.10_10, 0.08_00, 0.07_94, 0.08_85, 0.08_43, 0.07_62, 0.07_69, 0.07_29, 0.05_86] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = DDIMScheduler.from_pretrained("BAAI/AltDiffusion" , subfolder="scheduler" ) UpperCAmelCase_ : List[Any] = AltDiffusionPipeline.from_pretrained("BAAI/AltDiffusion" , scheduler=lowercase_ , safety_checker=lowercase_ ) UpperCAmelCase_ : Optional[Any] = alt_pipe.to(lowercase_ ) alt_pipe.set_progress_bar_config(disable=lowercase_ ) UpperCAmelCase_ : str = "A painting of a squirrel eating a burger" UpperCAmelCase_ : Any = torch.manual_seed(0 ) UpperCAmelCase_ : Dict = alt_pipe([prompt] , generator=lowercase_ , num_inference_steps=2 , output_type="numpy" ) UpperCAmelCase_ : Tuple = output.images UpperCAmelCase_ : Tuple = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) UpperCAmelCase_ : Dict = np.array([0.40_19, 0.40_52, 0.38_10, 0.41_19, 0.39_16, 0.39_82, 0.46_51, 0.41_95, 0.53_23] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
23
"""simple docstring""" import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = "ylacombe/bark-small" UpperCAmelCase_ : Union[str, Any] = tempfile.mkdtemp() UpperCAmelCase_ : List[str] = "en_speaker_1" UpperCAmelCase_ : Tuple = "This is a test string" UpperCAmelCase_ : List[Any] = "speaker_embeddings_path.json" UpperCAmelCase_ : Any = "speaker_embeddings" def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" return AutoTokenizer.from_pretrained(self.checkpoint , **lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.get_tokenizer() UpperCAmelCase_ : Union[str, Any] = BarkProcessor(tokenizer=lowercase_ ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase_ : Optional[int] = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) UpperCAmelCase_ : Dict = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) UpperCAmelCase_ : Union[str, Any] = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token="(BOS)" , eos_token="(EOS)" , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) UpperCAmelCase_ : int = 35 UpperCAmelCase_ : Optional[Any] = 2 UpperCAmelCase_ : List[Any] = 8 UpperCAmelCase_ : Optional[Any] = { "semantic_prompt": np.ones(lowercase_ ), "coarse_prompt": np.ones((nb_codebooks_coarse, seq_len) ), "fine_prompt": np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset UpperCAmelCase_ : Dict = processor(text=self.input_string , voice_preset=lowercase_ ) UpperCAmelCase_ : List[str] = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowercase_ , np.array([] ) ).tolist() ) # test loading voice preset from npz file UpperCAmelCase_ : Tuple = os.path.join(self.tmpdirname , "file.npz" ) np.savez(lowercase_ , **lowercase_ ) UpperCAmelCase_ : Optional[int] = processor(text=self.input_string , voice_preset=lowercase_ ) UpperCAmelCase_ : List[str] = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowercase_ , np.array([] ) ).tolist() ) # test loading voice preset from the hub UpperCAmelCase_ : Tuple = processor(text=self.input_string , voice_preset=self.voice_preset ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = self.get_tokenizer() UpperCAmelCase_ : Optional[Any] = BarkProcessor(tokenizer=lowercase_ ) UpperCAmelCase_ : Tuple = processor(text=self.input_string ) UpperCAmelCase_ : Union[str, Any] = tokenizer( self.input_string , padding="max_length" , max_length=256 , add_special_tokens=lowercase_ , return_attention_mask=lowercase_ , return_token_type_ids=lowercase_ , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
23
1
"""simple docstring""" import random def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Dict = [], [], [] for element in data: if element < pivot: less.append(__lowerCamelCase ) elif element > pivot: greater.append(__lowerCamelCase ) else: equal.append(__lowerCamelCase ) return less, equal, greater def __a ( __lowerCamelCase, __lowerCamelCase ): # index = len(items) // 2 when trying to find the median # (value of index when items is sorted) # invalid input if index >= len(__lowerCamelCase ) or index < 0: return None UpperCAmelCase_ : int = items[random.randint(0, len(__lowerCamelCase ) - 1 )] UpperCAmelCase_ : List[str] = 0 UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Dict = _partition(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : int = len(__lowerCamelCase ) UpperCAmelCase_ : Union[str, Any] = len(__lowerCamelCase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(__lowerCamelCase, __lowerCamelCase ) # must be in larger else: return quick_select(__lowerCamelCase, index - (m + count) )
23
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def __a ( __lowerCamelCase, __lowerCamelCase=False ): UpperCAmelCase_ : Optional[int] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""deit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""deit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""deit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""deit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""deit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""deit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""deit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""deit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""deit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""deit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "deit.embeddings.cls_token"), ("dist_token", "deit.embeddings.distillation_token"), ("patch_embed.proj.weight", "deit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "deit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "deit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" UpperCAmelCase_ : Dict = [(pair[0], pair[1][4:]) if pair[1].startswith("deit" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("norm.weight", "deit.layernorm.weight"), ("norm.bias", "deit.layernorm.bias"), ("head.weight", "cls_classifier.weight"), ("head.bias", "cls_classifier.bias"), ("head_dist.weight", "distillation_classifier.weight"), ("head_dist.bias", "distillation_classifier.bias"), ] ) return rename_keys def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=False ): for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase_ : int = "" else: UpperCAmelCase_ : Union[str, Any] = "deit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase_ : Tuple = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) UpperCAmelCase_ : Dict = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ : Union[str, Any] = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase_ : Any = in_proj_bias[: config.hidden_size] UpperCAmelCase_ : Optional[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase_ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase_ : List[Any] = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase_ : str = in_proj_bias[-config.hidden_size :] def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Tuple = dct.pop(__lowerCamelCase ) UpperCAmelCase_ : Tuple = val def __a ( ): UpperCAmelCase_ : Union[str, Any] = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ : str = Image.open(requests.get(__lowerCamelCase, stream=__lowerCamelCase ).raw ) return im @torch.no_grad() def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : List[str] = DeiTConfig() # all deit models have fine-tuned heads UpperCAmelCase_ : Optional[int] = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase_ : Tuple = 1000 UpperCAmelCase_ : str = "huggingface/label-files" UpperCAmelCase_ : str = "imagenet-1k-id2label.json" UpperCAmelCase_ : List[Any] = json.load(open(hf_hub_download(__lowerCamelCase, __lowerCamelCase, repo_type="dataset" ), "r" ) ) UpperCAmelCase_ : List[str] = {int(__lowerCamelCase ): v for k, v in idalabel.items()} UpperCAmelCase_ : Any = idalabel UpperCAmelCase_ : int = {v: k for k, v in idalabel.items()} UpperCAmelCase_ : Any = int(deit_name[-6:-4] ) UpperCAmelCase_ : Dict = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("tiny" ): UpperCAmelCase_ : Any = 192 UpperCAmelCase_ : Union[str, Any] = 768 UpperCAmelCase_ : Union[str, Any] = 12 UpperCAmelCase_ : int = 3 elif deit_name[9:].startswith("small" ): UpperCAmelCase_ : List[str] = 384 UpperCAmelCase_ : List[str] = 1536 UpperCAmelCase_ : Dict = 12 UpperCAmelCase_ : Any = 6 if deit_name[9:].startswith("base" ): pass elif deit_name[4:].startswith("large" ): UpperCAmelCase_ : int = 1024 UpperCAmelCase_ : List[Any] = 4096 UpperCAmelCase_ : Optional[int] = 24 UpperCAmelCase_ : int = 16 # load original model from timm UpperCAmelCase_ : Union[str, Any] = timm.create_model(__lowerCamelCase, pretrained=__lowerCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCAmelCase_ : Optional[Any] = timm_model.state_dict() UpperCAmelCase_ : Tuple = create_rename_keys(__lowerCamelCase, __lowerCamelCase ) for src, dest in rename_keys: rename_key(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) read_in_q_k_v(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) # load HuggingFace model UpperCAmelCase_ : str = DeiTForImageClassificationWithTeacher(__lowerCamelCase ).eval() model.load_state_dict(__lowerCamelCase ) # Check outputs on an image, prepared by DeiTImageProcessor UpperCAmelCase_ : Union[str, Any] = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 UpperCAmelCase_ : Optional[Any] = DeiTImageProcessor(size=__lowerCamelCase, crop_size=config.image_size ) UpperCAmelCase_ : Any = image_processor(images=prepare_img(), return_tensors="pt" ) UpperCAmelCase_ : int = encoding["pixel_values"] UpperCAmelCase_ : Optional[Any] = model(__lowerCamelCase ) UpperCAmelCase_ : Any = timm_model(__lowerCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__lowerCamelCase, outputs.logits, atol=1E-3 ) Path(__lowerCamelCase ).mkdir(exist_ok=__lowerCamelCase ) print(f"""Saving model {deit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCamelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--deit_name', default='vit_deit_base_distilled_patch16_224', type=str, help='Name of the DeiT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _a = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
23
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_camembert import CamembertTokenizer else: _a = None _a = logging.get_logger(__name__) _a = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} _a = { 'vocab_file': { 'camembert-base': 'https://huggingface.co/camembert-base/resolve/main/sentencepiece.bpe.model', }, 'tokenizer_file': { 'camembert-base': 'https://huggingface.co/camembert-base/resolve/main/tokenizer.json', }, } _a = { 'camembert-base': 512, } _a = '▁' class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : str = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : Union[str, Any] = ["""input_ids""", """attention_mask"""] SCREAMING_SNAKE_CASE__ : str = CamembertTokenizer def __init__( self , lowercase_=None , lowercase_=None , lowercase_="<s>" , lowercase_="</s>" , lowercase_="</s>" , lowercase_="<s>" , lowercase_="<unk>" , lowercase_="<pad>" , lowercase_="<mask>" , lowercase_=["<s>NOTUSED", "</s>NOTUSED"] , **lowercase_ , ): """simple docstring""" # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase_ : str = AddedToken(lowercase_ , lstrip=lowercase_ , rstrip=lowercase_ ) if isinstance(lowercase_ , lowercase_ ) else mask_token super().__init__( lowercase_ , tokenizer_file=lowercase_ , bos_token=lowercase_ , eos_token=lowercase_ , sep_token=lowercase_ , cls_token=lowercase_ , unk_token=lowercase_ , pad_token=lowercase_ , mask_token=lowercase_ , additional_special_tokens=lowercase_ , **lowercase_ , ) UpperCAmelCase_ : Any = vocab_file UpperCAmelCase_ : int = False if not self.vocab_file else True def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase_ : str = [self.cls_token_id] UpperCAmelCase_ : List[Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" UpperCAmelCase_ : Optional[int] = [self.sep_token_id] UpperCAmelCase_ : List[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowercase_ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase_ : Any = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase_ ): copyfile(self.vocab_file , lowercase_ ) return (out_vocab_file,)
23
"""simple docstring""" import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights UpperCAmelCase_ : Optional[Any] = FlaxDiffusionPipeline.from_pretrained( "hf-internal-testing/tiny-stable-diffusion-pipe" , safety_checker=lowercase_ , cache_dir=lowercase_ ) UpperCAmelCase_ : List[Any] = [t[-1] for t in os.walk(os.path.join(lowercase_ , os.listdir(lowercase_ )[0] , "snapshots" ) )] UpperCAmelCase_ : Dict = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith(".bin" ) for f in files ) @slow @require_flax class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( "hf-internal-testing/tiny-stable-diffusion-pipe" , safety_checker=lowercase_ ) UpperCAmelCase_ : Tuple = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : List[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase_ : List[str] = 4 UpperCAmelCase_ : Tuple = jax.device_count() UpperCAmelCase_ : Optional[int] = num_samples * [prompt] UpperCAmelCase_ : List[Any] = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : int = replicate(lowercase_ ) UpperCAmelCase_ : str = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : Dict = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_51_47_45 ) < 1E-3 assert np.abs(np.abs(lowercase_ , dtype=np.floataa ).sum() - 4_99_47.8_75 ) < 5E-1 UpperCAmelCase_ : List[Any] = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(lowercase_ ) == num_samples def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="flax" , safety_checker=lowercase_ ) UpperCAmelCase_ : Optional[int] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : str = jax.random.PRNGKey(0 ) UpperCAmelCase_ : Union[str, Any] = 50 UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[str] = num_samples * [prompt] UpperCAmelCase_ : Union[str, Any] = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Any = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : int = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_65_24_01) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_38_38_08.2) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ ) UpperCAmelCase_ : Any = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : str = jax.random.PRNGKey(0 ) UpperCAmelCase_ : str = 50 UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : Any = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Dict = replicate(lowercase_ ) UpperCAmelCase_ : str = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Union[str, Any] = shard(lowercase_ ) UpperCAmelCase_ : List[Any] = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : str = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa ) UpperCAmelCase_ : List[Any] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : Dict = jax.random.PRNGKey(0 ) UpperCAmelCase_ : Optional[int] = 50 UpperCAmelCase_ : Optional[int] = jax.device_count() UpperCAmelCase_ : str = num_samples * [prompt] UpperCAmelCase_ : int = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Union[str, Any] = replicate(lowercase_ ) UpperCAmelCase_ : Union[str, Any] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[int] = shard(lowercase_ ) UpperCAmelCase_ : Any = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = FlaxDDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="scaled_linear" , set_alpha_to_one=lowercase_ , steps_offset=1 , ) UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , scheduler=lowercase_ , safety_checker=lowercase_ , ) UpperCAmelCase_ : List[Any] = scheduler.create_state() UpperCAmelCase_ : int = scheduler_state UpperCAmelCase_ : Union[str, Any] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : Optional[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase_ : int = 50 UpperCAmelCase_ : str = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : int = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : int = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = shard(lowercase_ ) UpperCAmelCase_ : Any = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_45_04_39_45) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_34_76_93.5) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : Union[str, Any] = jax.random.split(jax.random.PRNGKey(0 ) , lowercase_ ) UpperCAmelCase_ , UpperCAmelCase_ : str = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ , ) UpperCAmelCase_ : Any = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = pipeline.prepare_inputs(lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : List[Any] = pipeline(lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase_ : int = images[2, 0, 256, 10:17, 1] # With memory efficient attention UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ , use_memory_efficient_attention=lowercase_ , ) UpperCAmelCase_ : str = replicate(lowercase_ ) UpperCAmelCase_ : str = pipeline.prepare_inputs(lowercase_ ) UpperCAmelCase_ : Optional[int] = shard(lowercase_ ) UpperCAmelCase_ : str = pipeline(lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase_ : Optional[int] = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1E-2
23
1
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { 'asapp/sew-d-tiny-100k': 'https://huggingface.co/asapp/sew-d-tiny-100k/resolve/main/config.json', # See all SEW-D models at https://huggingface.co/models?filter=sew-d } class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = """sew-d""" def __init__( self , lowercase_=32 , lowercase_=768 , lowercase_=12 , lowercase_=12 , lowercase_=3072 , lowercase_=2 , lowercase_=512 , lowercase_=256 , lowercase_=True , lowercase_=True , lowercase_=("p2c", "c2p") , lowercase_="layer_norm" , lowercase_="gelu_python" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=0.0 , lowercase_=0.1 , lowercase_=0.02 , lowercase_=1E-7 , lowercase_=1E-5 , lowercase_="group" , lowercase_="gelu" , lowercase_=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , lowercase_=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , lowercase_=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , lowercase_=False , lowercase_=128 , lowercase_=16 , lowercase_=True , lowercase_=0.05 , lowercase_=10 , lowercase_=2 , lowercase_=0.0 , lowercase_=10 , lowercase_=0 , lowercase_="mean" , lowercase_=False , lowercase_=False , lowercase_=256 , lowercase_=0 , lowercase_=1 , lowercase_=2 , **lowercase_ , ): """simple docstring""" super().__init__(**lowercase_ , pad_token_id=lowercase_ , bos_token_id=lowercase_ , eos_token_id=lowercase_ ) UpperCAmelCase_ : Tuple = hidden_size UpperCAmelCase_ : Dict = feat_extract_norm UpperCAmelCase_ : Tuple = feat_extract_activation UpperCAmelCase_ : str = list(lowercase_ ) UpperCAmelCase_ : Tuple = list(lowercase_ ) UpperCAmelCase_ : List[str] = list(lowercase_ ) UpperCAmelCase_ : Dict = conv_bias UpperCAmelCase_ : int = num_conv_pos_embeddings UpperCAmelCase_ : str = num_conv_pos_embedding_groups UpperCAmelCase_ : int = len(self.conv_dim ) UpperCAmelCase_ : List[str] = num_hidden_layers UpperCAmelCase_ : Optional[int] = intermediate_size UpperCAmelCase_ : Tuple = squeeze_factor UpperCAmelCase_ : Tuple = max_position_embeddings UpperCAmelCase_ : Tuple = position_buckets UpperCAmelCase_ : Optional[int] = share_att_key UpperCAmelCase_ : Dict = relative_attention UpperCAmelCase_ : List[Any] = norm_rel_ebd UpperCAmelCase_ : Optional[int] = list(lowercase_ ) UpperCAmelCase_ : List[str] = hidden_act UpperCAmelCase_ : Union[str, Any] = num_attention_heads UpperCAmelCase_ : List[str] = hidden_dropout UpperCAmelCase_ : Optional[int] = attention_dropout UpperCAmelCase_ : Optional[Any] = activation_dropout UpperCAmelCase_ : Tuple = feat_proj_dropout UpperCAmelCase_ : Optional[int] = final_dropout UpperCAmelCase_ : Optional[int] = layer_norm_eps UpperCAmelCase_ : Union[str, Any] = feature_layer_norm_eps UpperCAmelCase_ : Optional[Any] = initializer_range UpperCAmelCase_ : int = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( "Configuration for convolutional layers is incorrect." "It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`," F"""but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)""" F"""= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCAmelCase_ : Optional[int] = apply_spec_augment UpperCAmelCase_ : Dict = mask_time_prob UpperCAmelCase_ : Optional[Any] = mask_time_length UpperCAmelCase_ : List[str] = mask_time_min_masks UpperCAmelCase_ : Dict = mask_feature_prob UpperCAmelCase_ : int = mask_feature_length UpperCAmelCase_ : Any = mask_feature_min_masks # ctc loss UpperCAmelCase_ : Tuple = ctc_loss_reduction UpperCAmelCase_ : Tuple = ctc_zero_infinity # sequence classification UpperCAmelCase_ : Any = use_weighted_layer_sum UpperCAmelCase_ : Union[str, Any] = classifier_proj_size @property def UpperCamelCase__ ( self ): """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
23
"""simple docstring""" from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean _a = 0 _a = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] _a = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right _a = tuple[int, int] class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ): """simple docstring""" UpperCAmelCase_ : int = pos_x UpperCAmelCase_ : List[Any] = pos_y UpperCAmelCase_ : Union[str, Any] = (pos_y, pos_x) UpperCAmelCase_ : Any = goal_x UpperCAmelCase_ : Dict = goal_y UpperCAmelCase_ : Any = g_cost UpperCAmelCase_ : List[str] = parent UpperCAmelCase_ : int = self.calculate_heuristic() UpperCAmelCase_ : Any = self.g_cost + self.h_cost def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.pos_x - self.goal_x UpperCAmelCase_ : Union[str, Any] = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowercase_ ) + abs(lowercase_ ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self , lowercase_ ): """simple docstring""" return self.f_cost < other.f_cost class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Tuple = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , lowercase_ ) UpperCAmelCase_ : List[Any] = Node(goal[1] , goal[0] , goal[1] , goal[0] , 9_9999 , lowercase_ ) UpperCAmelCase_ : str = [self.start] UpperCAmelCase_ : list[Node] = [] UpperCAmelCase_ : int = False def UpperCamelCase__ ( self ): """simple docstring""" while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() UpperCAmelCase_ : List[str] = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(lowercase_ ) self.closed_nodes.append(lowercase_ ) UpperCAmelCase_ : str = self.get_successors(lowercase_ ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowercase_ ) else: # retrieve the best current path UpperCAmelCase_ : Union[str, Any] = self.open_nodes.pop(self.open_nodes.index(lowercase_ ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowercase_ ) else: self.open_nodes.append(lowercase_ ) return [self.start.pos] def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Any = [] for action in delta: UpperCAmelCase_ : str = parent.pos_x + action[1] UpperCAmelCase_ : int = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowercase_ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowercase_ , lowercase_ , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , lowercase_ , ) ) return successors def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = node UpperCAmelCase_ : int = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) UpperCAmelCase_ : Optional[int] = current_node.parent path.reverse() return path class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Dict = AStar(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = AStar(lowercase_ , lowercase_ ) UpperCAmelCase_ : Tuple = False def UpperCamelCase__ ( self ): """simple docstring""" while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() UpperCAmelCase_ : List[str] = self.fwd_astar.open_nodes.pop(0 ) UpperCAmelCase_ : List[Any] = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowercase_ , lowercase_ ) self.fwd_astar.closed_nodes.append(lowercase_ ) self.bwd_astar.closed_nodes.append(lowercase_ ) UpperCAmelCase_ : Tuple = current_bwd_node UpperCAmelCase_ : str = current_fwd_node UpperCAmelCase_ : Dict = { self.fwd_astar: self.fwd_astar.get_successors(lowercase_ ), self.bwd_astar: self.bwd_astar.get_successors(lowercase_ ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowercase_ ) else: # retrieve the best current path UpperCAmelCase_ : List[Any] = astar.open_nodes.pop( astar.open_nodes.index(lowercase_ ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowercase_ ) else: astar.open_nodes.append(lowercase_ ) return [self.fwd_astar.start.pos] def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.fwd_astar.retrace_path(lowercase_ ) UpperCAmelCase_ : int = self.bwd_astar.retrace_path(lowercase_ ) bwd_path.pop() bwd_path.reverse() UpperCAmelCase_ : Any = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] _a = (0, 0) _a = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) _a = time.time() _a = AStar(init, goal) _a = a_star.search() _a = time.time() - start_time print(f"""AStar execution time = {end_time:f} seconds""") _a = time.time() _a = BidirectionalAStar(init, goal) _a = time.time() - bd_start_time print(f"""BidirectionalAStar execution time = {bd_end_time:f} seconds""")
23
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _a = { 'configuration_clap': [ 'CLAP_PRETRAINED_MODEL_ARCHIVE_LIST', 'ClapAudioConfig', 'ClapConfig', 'ClapTextConfig', ], 'processing_clap': ['ClapProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'CLAP_PRETRAINED_MODEL_ARCHIVE_LIST', 'ClapModel', 'ClapPreTrainedModel', 'ClapTextModel', 'ClapTextModelWithProjection', 'ClapAudioModel', 'ClapAudioModelWithProjection', ] _a = ['ClapFeatureExtractor'] if TYPE_CHECKING: from .configuration_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioConfig, ClapConfig, ClapTextConfig, ) from .processing_clap import ClapProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_clap import ClapFeatureExtractor from .modeling_clap import ( CLAP_PRETRAINED_MODEL_ARCHIVE_LIST, ClapAudioModel, ClapAudioModelWithProjection, ClapModel, ClapPreTrainedModel, ClapTextModel, ClapTextModelWithProjection, ) else: import sys _a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
23
"""simple docstring""" import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = (PNDMScheduler,) SCREAMING_SNAKE_CASE__ : str = (("""num_inference_steps""", 50),) def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : int = { "num_train_timesteps": 1000, "beta_start": 0.00_01, "beta_end": 0.02, "beta_schedule": "linear", } config.update(**lowercase_ ) return config def UpperCamelCase__ ( self , lowercase_=0 , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = dict(self.forward_default_kwargs ) UpperCAmelCase_ : List[str] = kwargs.pop("num_inference_steps" , lowercase_ ) UpperCAmelCase_ : Union[str, Any] = self.dummy_sample UpperCAmelCase_ : Dict = 0.1 * sample UpperCAmelCase_ : Dict = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : List[Any] = self.get_scheduler_config(**lowercase_ ) UpperCAmelCase_ : Dict = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals UpperCAmelCase_ : List[Any] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) UpperCAmelCase_ : Optional[int] = scheduler_class.from_pretrained(lowercase_ ) new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals UpperCAmelCase_ : int = dummy_past_residuals[:] UpperCAmelCase_ : List[str] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : str = new_scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" UpperCAmelCase_ : Optional[int] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Dict = new_scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self , lowercase_=0 , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = dict(self.forward_default_kwargs ) UpperCAmelCase_ : str = kwargs.pop("num_inference_steps" , lowercase_ ) UpperCAmelCase_ : Optional[int] = self.dummy_sample UpperCAmelCase_ : List[str] = 0.1 * sample UpperCAmelCase_ : Tuple = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : str = self.get_scheduler_config() UpperCAmelCase_ : Dict = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals (must be after setting timesteps) UpperCAmelCase_ : List[Any] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) UpperCAmelCase_ : Dict = scheduler_class.from_pretrained(lowercase_ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residual (must be after setting timesteps) UpperCAmelCase_ : Optional[Any] = dummy_past_residuals[:] UpperCAmelCase_ : Union[str, Any] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Dict = new_scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" UpperCAmelCase_ : List[str] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : int = new_scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = self.scheduler_classes[0] UpperCAmelCase_ : Union[str, Any] = self.get_scheduler_config(**lowercase_ ) UpperCAmelCase_ : List[Any] = scheduler_class(**lowercase_ ) UpperCAmelCase_ : Tuple = 10 UpperCAmelCase_ : List[str] = self.dummy_model() UpperCAmelCase_ : str = self.dummy_sample_deter scheduler.set_timesteps(lowercase_ ) for i, t in enumerate(scheduler.prk_timesteps ): UpperCAmelCase_ : Tuple = model(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[int] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): UpperCAmelCase_ : Any = model(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ ).prev_sample return sample def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = dict(self.forward_default_kwargs ) UpperCAmelCase_ : Optional[Any] = kwargs.pop("num_inference_steps" , lowercase_ ) for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : Any = self.get_scheduler_config() UpperCAmelCase_ : Tuple = scheduler_class(**lowercase_ ) UpperCAmelCase_ : str = self.dummy_sample UpperCAmelCase_ : List[Any] = 0.1 * sample if num_inference_steps is not None and hasattr(lowercase_ , "set_timesteps" ): scheduler.set_timesteps(lowercase_ ) elif num_inference_steps is not None and not hasattr(lowercase_ , "set_timesteps" ): UpperCAmelCase_ : List[Any] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) UpperCAmelCase_ : List[str] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] UpperCAmelCase_ : List[str] = dummy_past_residuals[:] UpperCAmelCase_ : str = scheduler.step_prk(lowercase_ , 0 , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Any = scheduler.step_prk(lowercase_ , 1 , lowercase_ , **lowercase_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , 0 , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , 1 , lowercase_ , **lowercase_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def UpperCamelCase__ ( self ): """simple docstring""" for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for steps_offset in [0, 1]: self.check_over_configs(steps_offset=lowercase_ ) UpperCAmelCase_ : Optional[int] = self.scheduler_classes[0] UpperCAmelCase_ : int = self.get_scheduler_config(steps_offset=1 ) UpperCAmelCase_ : Optional[Any] = scheduler_class(**lowercase_ ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [901, 851, 851, 801, 801, 751, 751, 701, 701, 651, 651, 601, 601, 501, 401, 301, 201, 101, 1] ) , ) def UpperCamelCase__ ( self ): """simple docstring""" for beta_start, beta_end in zip([0.00_01, 0.0_01] , [0.0_02, 0.02] ): self.check_over_configs(beta_start=lowercase_ , beta_end=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for t in [1, 5, 10]: self.check_over_forward(time_step=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" # earlier version of set_timesteps() caused an error indexing alpha's with inference steps as power of 3 UpperCAmelCase_ : List[Any] = 27 for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : List[Any] = self.dummy_sample UpperCAmelCase_ : Optional[int] = 0.1 * sample UpperCAmelCase_ : List[str] = self.get_scheduler_config() UpperCAmelCase_ : List[str] = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): UpperCAmelCase_ : List[str] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ ).prev_sample def UpperCamelCase__ ( self ): """simple docstring""" with self.assertRaises(lowercase_ ): UpperCAmelCase_ : List[str] = self.scheduler_classes[0] UpperCAmelCase_ : str = self.get_scheduler_config() UpperCAmelCase_ : Tuple = scheduler_class(**lowercase_ ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.full_loop() UpperCAmelCase_ : Any = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Dict = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 1_98.13_18 ) < 1E-2 assert abs(result_mean.item() - 0.25_80 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = self.full_loop(prediction_type="v_prediction" ) UpperCAmelCase_ : str = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Tuple = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 67.39_86 ) < 1E-2 assert abs(result_mean.item() - 0.08_78 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase_ : Union[str, Any] = self.full_loop(set_alpha_to_one=lowercase_ , beta_start=0.01 ) UpperCAmelCase_ : List[Any] = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : int = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 2_30.03_99 ) < 1E-2 assert abs(result_mean.item() - 0.29_95 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase_ : Tuple = self.full_loop(set_alpha_to_one=lowercase_ , beta_start=0.01 ) UpperCAmelCase_ : int = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Tuple = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 1_86.94_82 ) < 1E-2 assert abs(result_mean.item() - 0.24_34 ) < 1E-3
23
1
"""simple docstring""" import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss _a = pytest.mark.integration @require_faiss class A_ (lowercase__ ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = Dataset.from_dict({"filename": ["my_name-train" + "_" + str(lowercase_ ) for x in np.arange(30 ).tolist()]} ) return dset def UpperCamelCase__ ( self ): """simple docstring""" import faiss UpperCAmelCase_ : Dataset = self._create_dummy_dataset() UpperCAmelCase_ : List[Any] = dset.map( lambda lowercase_ , lowercase_ : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=lowercase_ , keep_in_memory=lowercase_ ) UpperCAmelCase_ : Optional[Any] = dset.add_faiss_index("vecs" , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT ) UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = dset.get_nearest_examples("vecs" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) dset.drop_index("vecs" ) def UpperCamelCase__ ( self ): """simple docstring""" import faiss UpperCAmelCase_ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name="vecs" , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT , ) UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = dset.get_nearest_examples("vecs" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) def UpperCamelCase__ ( self ): """simple docstring""" import faiss UpperCAmelCase_ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name="vecs" , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowercase_ ) as tmp_file: dset.save_faiss_index("vecs" , tmp_file.name ) dset.load_faiss_index("vecs2" , tmp_file.name ) os.unlink(tmp_file.name ) UpperCAmelCase_ , UpperCAmelCase_ : int = dset.get_nearest_examples("vecs2" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dataset = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name="vecs" ) dset.drop_index("vecs" ) self.assertRaises(lowercase_ , partial(dset.get_nearest_examples , "vecs2" , np.ones(5 , dtype=np.floataa ) ) ) def UpperCamelCase__ ( self ): """simple docstring""" from elasticsearch import Elasticsearch UpperCAmelCase_ : Dataset = self._create_dummy_dataset() with patch("elasticsearch.Elasticsearch.search" ) as mocked_search, patch( "elasticsearch.client.IndicesClient.create" ) as mocked_index_create, patch("elasticsearch.helpers.streaming_bulk" ) as mocked_bulk: UpperCAmelCase_ : List[Any] = {"acknowledged": True} mocked_bulk.return_value([(True, None)] * 30 ) UpperCAmelCase_ : List[Any] = {"hits": {"hits": [{"_score": 1, "_id": 29}]}} UpperCAmelCase_ : Any = Elasticsearch() dset.add_elasticsearch_index("filename" , es_client=lowercase_ ) UpperCAmelCase_ , UpperCAmelCase_ : List[str] = dset.get_nearest_examples("filename" , "my_name-train_29" ) self.assertEqual(examples["filename"][0] , "my_name-train_29" ) @require_faiss class A_ (lowercase__ ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" import faiss UpperCAmelCase_ : List[str] = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 10 ) # single query UpperCAmelCase_ : Dict = np.zeros(5 , dtype=np.floataa ) UpperCAmelCase_ : int = 1 UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = index.search(lowercase_ ) self.assertRaises(lowercase_ , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries UpperCAmelCase_ : Any = np.eye(5 , dtype=np.floataa )[::-1] UpperCAmelCase_ , UpperCAmelCase_ : str = index.search_batch(lowercase_ ) self.assertRaises(lowercase_ , index.search_batch , queries[0] ) UpperCAmelCase_ : str = [scores[0] for scores in total_scores] UpperCAmelCase_ : List[str] = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowercase_ ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" import faiss UpperCAmelCase_ : Tuple = FaissIndex(string_factory="Flat" ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) UpperCAmelCase_ : List[Any] = FaissIndex(string_factory="LSH" ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(lowercase_ ): UpperCAmelCase_ : Any = FaissIndex(string_factory="Flat" , custom_index=faiss.IndexFlat(5 ) ) def UpperCamelCase__ ( self ): """simple docstring""" import faiss UpperCAmelCase_ : List[str] = faiss.IndexFlat(5 ) UpperCAmelCase_ : Union[str, Any] = FaissIndex(custom_index=lowercase_ ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def UpperCamelCase__ ( self ): """simple docstring""" import faiss UpperCAmelCase_ : Optional[Any] = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=lowercase_ ) as tmp_file: index.save(tmp_file.name ) UpperCAmelCase_ : str = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) UpperCAmelCase_ : Tuple = np.zeros(5 , dtype=np.floataa ) UpperCAmelCase_ : Optional[Any] = 1 UpperCAmelCase_ , UpperCAmelCase_ : Tuple = index.search(lowercase_ ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def __a ( __lowerCamelCase ): import faiss UpperCAmelCase_ : Optional[int] = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5, dtype=np.floataa ) ) UpperCAmelCase_ : List[str] = "index.faiss" UpperCAmelCase_ : str = f"""mock://{index_name}""" index.save(__lowerCamelCase, storage_options=mockfs.storage_options ) UpperCAmelCase_ : List[str] = FaissIndex.load(__lowerCamelCase, storage_options=mockfs.storage_options ) UpperCAmelCase_ : Union[str, Any] = np.zeros(5, dtype=np.floataa ) UpperCAmelCase_ : List[str] = 1 UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = index.search(__lowerCamelCase ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class A_ (lowercase__ ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" from elasticsearch import Elasticsearch with patch("elasticsearch.Elasticsearch.search" ) as mocked_search, patch( "elasticsearch.client.IndicesClient.create" ) as mocked_index_create, patch("elasticsearch.helpers.streaming_bulk" ) as mocked_bulk: UpperCAmelCase_ : List[Any] = Elasticsearch() UpperCAmelCase_ : int = {"acknowledged": True} UpperCAmelCase_ : List[str] = ElasticSearchIndex(es_client=lowercase_ ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(["foo", "bar", "foobar"] ) # single query UpperCAmelCase_ : Optional[int] = "foo" UpperCAmelCase_ : Union[str, Any] = {"hits": {"hits": [{"_score": 1, "_id": 0}]}} UpperCAmelCase_ , UpperCAmelCase_ : Any = index.search(lowercase_ ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout UpperCAmelCase_ : Union[str, Any] = "foo" UpperCAmelCase_ : Union[str, Any] = {"hits": {"hits": [{"_score": 1, "_id": 0}]}} UpperCAmelCase_ , UpperCAmelCase_ : Tuple = index.search(lowercase_ , request_timeout=30 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries UpperCAmelCase_ : Any = ["foo", "bar", "foobar"] UpperCAmelCase_ : Dict = {"hits": {"hits": [{"_score": 1, "_id": 1}]}} UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = index.search_batch(lowercase_ ) UpperCAmelCase_ : Any = [scores[0] for scores in total_scores] UpperCAmelCase_ : str = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowercase_ ) , 0 ) self.assertListEqual([1, 1, 1] , lowercase_ ) # batched queries with timeout UpperCAmelCase_ : Dict = ["foo", "bar", "foobar"] UpperCAmelCase_ : Any = {"hits": {"hits": [{"_score": 1, "_id": 1}]}} UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = index.search_batch(lowercase_ , request_timeout=30 ) UpperCAmelCase_ : Tuple = [scores[0] for scores in total_scores] UpperCAmelCase_ : List[str] = [indices[0] for indices in total_indices] self.assertGreater(np.min(lowercase_ ) , 0 ) self.assertListEqual([1, 1, 1] , lowercase_ )
23
"""simple docstring""" import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels _a = object() # For specifying empty leaf dict `{}` _a = object() def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Any = tuple((re.compile(x + "$" ) for x in qs) ) for i in range(len(__lowerCamelCase ) - len(__lowerCamelCase ) + 1 ): UpperCAmelCase_ : List[str] = [x.match(__lowerCamelCase ) for x, y in zip(__lowerCamelCase, ks[i:] )] if matches and all(__lowerCamelCase ): return True return False def __a ( __lowerCamelCase ): def replace(__lowerCamelCase, __lowerCamelCase ): for rule, replacement in rules: if _match(__lowerCamelCase, __lowerCamelCase ): return replacement return val return replace def __a ( ): return [ # embeddings (("transformer", "wpe", "embedding"), P("mp", __lowerCamelCase )), (("transformer", "wte", "embedding"), P("mp", __lowerCamelCase )), # atention (("attention", "(q_proj|k_proj|v_proj)", "kernel"), P(__lowerCamelCase, "mp" )), (("attention", "out_proj", "kernel"), P("mp", __lowerCamelCase )), (("attention", "out_proj", "bias"), None), # mlp (("mlp", "c_fc", "kernel"), P(__lowerCamelCase, "mp" )), (("mlp", "c_fc", "bias"), P("mp" )), (("mlp", "c_proj", "kernel"), P("mp", __lowerCamelCase )), (("mlp", "c_proj", "bias"), None), # layer norms ((r"ln_\d+", "bias"), None), ((r"\d+", r"ln_\d+", "scale"), None), (("ln_f", "bias"), None), (("ln_f", "scale"), None), ] def __a ( __lowerCamelCase ): UpperCAmelCase_ : List[str] = _get_partition_rules() UpperCAmelCase_ : Any = _replacement_rules(__lowerCamelCase ) UpperCAmelCase_ : Any = {k: _unmatched for k in flatten_dict(__lowerCamelCase )} UpperCAmelCase_ : Dict = {k: replace(__lowerCamelCase, __lowerCamelCase ) for k, v in initd.items()} assert _unmatched not in result.values(), "Incomplete partition spec." return freeze(unflatten_dict(__lowerCamelCase ) )
23
1
"""simple docstring""" import numpy as np def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Union[str, Any] = int(np.ceil((x_end - xa) / h ) ) UpperCAmelCase_ : int = np.zeros((n + 1,) ) UpperCAmelCase_ : List[Any] = ya UpperCAmelCase_ : Optional[int] = xa for k in range(__lowerCamelCase ): UpperCAmelCase_ : Dict = f(__lowerCamelCase, y[k] ) UpperCAmelCase_ : int = f(x + 0.5 * h, y[k] + 0.5 * h * ka ) UpperCAmelCase_ : Dict = f(x + 0.5 * h, y[k] + 0.5 * h * ka ) UpperCAmelCase_ : str = f(x + h, y[k] + h * ka ) UpperCAmelCase_ : List[str] = y[k] + (1 / 6) * h * (ka + 2 * ka + 2 * ka + ka) x += h return y if __name__ == "__main__": import doctest doctest.testmod()
23
"""simple docstring""" import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow _a = logging.getLogger() @unittest.skip("""Temporarily disable the doc tests.""" ) @require_torch @require_tf @slow class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = True , ): """simple docstring""" UpperCAmelCase_ : List[str] = [file for file in os.listdir(lowercase_ ) if os.path.isfile(os.path.join(lowercase_ , lowercase_ ) )] if identifier is not None: UpperCAmelCase_ : Dict = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(lowercase_ , lowercase_ ): for n_ in n_identifier: UpperCAmelCase_ : str = [file for file in files if n_ not in file] else: UpperCAmelCase_ : Any = [file for file in files if n_identifier not in file] UpperCAmelCase_ : Union[str, Any] = ignore_files or [] ignore_files.append("__init__.py" ) UpperCAmelCase_ : Optional[int] = [file for file in files if file not in ignore_files] for file in files: # Open all files print("Testing" , lowercase_ ) if only_modules: UpperCAmelCase_ : str = file.split("." )[0] try: UpperCAmelCase_ : str = getattr(lowercase_ , lowercase_ ) UpperCAmelCase_ : Tuple = doctest.DocTestSuite(lowercase_ ) UpperCAmelCase_ : int = unittest.TextTestRunner().run(lowercase_ ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: UpperCAmelCase_ : Optional[Any] = doctest.testfile(str(".." / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = Path("src/transformers" ) UpperCAmelCase_ : str = "modeling" UpperCAmelCase_ : Optional[Any] = [ "modeling_ctrl.py", "modeling_tf_ctrl.py", ] self.analyze_directory(lowercase_ , identifier=lowercase_ , ignore_files=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = Path("src/transformers" ) UpperCAmelCase_ : Any = "tokenization" self.analyze_directory(lowercase_ , identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = Path("src/transformers" ) UpperCAmelCase_ : List[Any] = "configuration" self.analyze_directory(lowercase_ , identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = Path("src/transformers" ) UpperCAmelCase_ : List[Any] = ["configuration", "modeling", "tokenization"] self.analyze_directory(lowercase_ , n_identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = Path("docs/source" ) UpperCAmelCase_ : Union[str, Any] = ["favicon.ico"] self.analyze_directory(lowercase_ , ignore_files=lowercase_ , only_modules=lowercase_ )
23
1
"""simple docstring""" import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) _a = pytest.mark.integration @pytest.mark.parametrize("path", ["paws", "csv"] ) def __a ( __lowerCamelCase, __lowerCamelCase ): inspect_dataset(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = path + ".py" assert script_name in os.listdir(__lowerCamelCase ) assert "__pycache__" not in os.listdir(__lowerCamelCase ) @pytest.mark.filterwarnings("ignore:inspect_metric is deprecated:FutureWarning" ) @pytest.mark.filterwarnings("ignore:metric_module_factory is deprecated:FutureWarning" ) @pytest.mark.parametrize("path", ["accuracy"] ) def __a ( __lowerCamelCase, __lowerCamelCase ): inspect_metric(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : Any = path + ".py" assert script_name in os.listdir(__lowerCamelCase ) assert "__pycache__" not in os.listdir(__lowerCamelCase ) @pytest.mark.parametrize( "path, config_name, expected_splits", [ ("squad", "plain_text", ["train", "validation"]), ("dalle-mini/wit", "dalle-mini--wit", ["train"]), ("paws", "labeled_final", ["train", "test", "validation"]), ], ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Optional[int] = get_dataset_config_info(__lowerCamelCase, config_name=__lowerCamelCase ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( "path, config_name, expected_exception", [ ("paws", None, ValueError), ], ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): with pytest.raises(__lowerCamelCase ): get_dataset_config_info(__lowerCamelCase, config_name=__lowerCamelCase ) @pytest.mark.parametrize( "path, expected", [ ("squad", "plain_text"), ("acronym_identification", "default"), ("lhoestq/squad", "plain_text"), ("lhoestq/test", "default"), ("lhoestq/demo1", "lhoestq--demo1"), ("dalle-mini/wit", "dalle-mini--wit"), ], ) def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Optional[Any] = get_dataset_config_names(__lowerCamelCase ) assert expected in config_names @pytest.mark.parametrize( "path, expected_configs, expected_splits_in_first_config", [ ("squad", ["plain_text"], ["train", "validation"]), ("dalle-mini/wit", ["dalle-mini--wit"], ["train"]), ("paws", ["labeled_final", "labeled_swap", "unlabeled_final"], ["train", "test", "validation"]), ], ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Union[str, Any] = get_dataset_infos(__lowerCamelCase ) assert list(infos.keys() ) == expected_configs UpperCAmelCase_ : str = expected_configs[0] assert expected_config in infos UpperCAmelCase_ : List[Any] = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( "path, expected_config, expected_splits", [ ("squad", "plain_text", ["train", "validation"]), ("dalle-mini/wit", "dalle-mini--wit", ["train"]), ("paws", "labeled_final", ["train", "test", "validation"]), ], ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Union[str, Any] = get_dataset_infos(__lowerCamelCase ) assert expected_config in infos UpperCAmelCase_ : Union[str, Any] = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( "path, config_name, expected_exception", [ ("paws", None, ValueError), ], ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): with pytest.raises(__lowerCamelCase ): get_dataset_split_names(__lowerCamelCase, config_name=__lowerCamelCase )
23
"""simple docstring""" import warnings from ...utils import is_sklearn_available, requires_backends if is_sklearn_available(): from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef _a = ( 'This metric will be removed from the library soon, metrics should be handled with the 🤗 Evaluate ' 'library. You can have a look at this example script for pointers: ' 'https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py' ) def __a ( __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) return (preds == labels).mean() def __a ( __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) UpperCAmelCase_ : Optional[Any] = simple_accuracy(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : List[Any] = fa_score(y_true=__lowerCamelCase, y_pred=__lowerCamelCase ) return { "acc": acc, "f1": fa, "acc_and_f1": (acc + fa) / 2, } def __a ( __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) UpperCAmelCase_ : Any = pearsonr(__lowerCamelCase, __lowerCamelCase )[0] UpperCAmelCase_ : Optional[Any] = spearmanr(__lowerCamelCase, __lowerCamelCase )[0] return { "pearson": pearson_corr, "spearmanr": spearman_corr, "corr": (pearson_corr + spearman_corr) / 2, } def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) assert len(__lowerCamelCase ) == len(__lowerCamelCase ), f"""Predictions and labels have mismatched lengths {len(__lowerCamelCase )} and {len(__lowerCamelCase )}""" if task_name == "cola": return {"mcc": matthews_corrcoef(__lowerCamelCase, __lowerCamelCase )} elif task_name == "sst-2": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "mrpc": return acc_and_fa(__lowerCamelCase, __lowerCamelCase ) elif task_name == "sts-b": return pearson_and_spearman(__lowerCamelCase, __lowerCamelCase ) elif task_name == "qqp": return acc_and_fa(__lowerCamelCase, __lowerCamelCase ) elif task_name == "mnli": return {"mnli/acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "mnli-mm": return {"mnli-mm/acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "qnli": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "rte": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "wnli": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "hans": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} else: raise KeyError(__lowerCamelCase ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) if len(__lowerCamelCase ) != len(__lowerCamelCase ): raise ValueError(f"""Predictions and labels have mismatched lengths {len(__lowerCamelCase )} and {len(__lowerCamelCase )}""" ) if task_name == "xnli": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} else: raise KeyError(__lowerCamelCase )
23
1
"""simple docstring""" import inspect import unittest from transformers import ConvNextConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextBackbone, ConvNextForImageClassification, ConvNextModel from transformers.models.convnext.modeling_convnext import CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_=13 , lowercase_=32 , lowercase_=3 , lowercase_=4 , lowercase_=[10, 20, 30, 40] , lowercase_=[2, 2, 3, 2] , lowercase_=True , lowercase_=True , lowercase_=37 , lowercase_="gelu" , lowercase_=10 , lowercase_=0.02 , lowercase_=["stage2", "stage3", "stage4"] , lowercase_=[2, 3, 4] , lowercase_=None , ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = parent UpperCAmelCase_ : List[Any] = batch_size UpperCAmelCase_ : Dict = image_size UpperCAmelCase_ : int = num_channels UpperCAmelCase_ : Union[str, Any] = num_stages UpperCAmelCase_ : Optional[Any] = hidden_sizes UpperCAmelCase_ : List[str] = depths UpperCAmelCase_ : Optional[int] = is_training UpperCAmelCase_ : Optional[Any] = use_labels UpperCAmelCase_ : Any = intermediate_size UpperCAmelCase_ : str = hidden_act UpperCAmelCase_ : Any = num_labels UpperCAmelCase_ : List[Any] = initializer_range UpperCAmelCase_ : str = out_features UpperCAmelCase_ : Optional[Any] = out_indices UpperCAmelCase_ : List[str] = scope def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase_ : List[Any] = None if self.use_labels: UpperCAmelCase_ : Union[str, Any] = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase_ : Tuple = self.get_config() return config, pixel_values, labels def UpperCamelCase__ ( self ): """simple docstring""" return ConvNextConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=lowercase_ , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Any = ConvNextModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() UpperCAmelCase_ : List[Any] = model(lowercase_ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = ConvNextForImageClassification(lowercase_ ) model.to(lowercase_ ) model.eval() UpperCAmelCase_ : Any = model(lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = ConvNextBackbone(config=lowercase_ ) model.to(lowercase_ ) model.eval() UpperCAmelCase_ : Dict = model(lowercase_ ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None UpperCAmelCase_ : int = None UpperCAmelCase_ : Union[str, Any] = ConvNextBackbone(config=lowercase_ ) model.to(lowercase_ ) model.eval() UpperCAmelCase_ : List[Any] = model(lowercase_ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.prepare_config_and_inputs() UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : List[Any] = config_and_inputs UpperCAmelCase_ : int = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class A_ (lowercase__ ,lowercase__ ,unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = ( ( ConvNextModel, ConvNextForImageClassification, ConvNextBackbone, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE__ : Any = ( {"""feature-extraction""": ConvNextModel, """image-classification""": ConvNextForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE__ : List[Any] = True SCREAMING_SNAKE_CASE__ : Union[str, Any] = False SCREAMING_SNAKE_CASE__ : Dict = False SCREAMING_SNAKE_CASE__ : Dict = False SCREAMING_SNAKE_CASE__ : str = False def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = ConvNextModelTester(self ) UpperCAmelCase_ : int = ConfigTester(self , config_class=lowercase_ , has_text_modality=lowercase_ , hidden_size=37 ) def UpperCamelCase__ ( self ): """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def UpperCamelCase__ ( self ): """simple docstring""" return @unittest.skip(reason="ConvNext does not use inputs_embeds" ) def UpperCamelCase__ ( self ): """simple docstring""" pass @unittest.skip(reason="ConvNext does not support input and output embeddings" ) def UpperCamelCase__ ( self ): """simple docstring""" pass @unittest.skip(reason="ConvNext does not use feedforward chunking" ) def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : Any = model_class(lowercase_ ) UpperCAmelCase_ : Tuple = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase_ : Union[str, Any] = [*signature.parameters.keys()] UpperCAmelCase_ : List[Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" def check_hidden_states_output(lowercase_ , lowercase_ , lowercase_ ): UpperCAmelCase_ : List[Any] = model_class(lowercase_ ) model.to(lowercase_ ) model.eval() with torch.no_grad(): UpperCAmelCase_ : Tuple = model(**self._prepare_for_class(lowercase_ , lowercase_ ) ) UpperCAmelCase_ : Any = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCAmelCase_ : int = self.model_tester.num_stages self.assertEqual(len(lowercase_ ) , expected_num_stages + 1 ) # ConvNext's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) UpperCAmelCase_ , UpperCAmelCase_ : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase_ : List[Any] = True check_hidden_states_output(lowercase_ , lowercase_ , lowercase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase_ : List[Any] = True check_hidden_states_output(lowercase_ , lowercase_ , lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase_ ) @slow def UpperCamelCase__ ( self ): """simple docstring""" for model_name in CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : Optional[int] = ConvNextModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) def __a ( ): UpperCAmelCase_ : Optional[int] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class A_ (unittest.TestCase ): '''simple docstring''' @cached_property def UpperCamelCase__ ( self ): """simple docstring""" return AutoImageProcessor.from_pretrained("facebook/convnext-tiny-224" ) if is_vision_available() else None @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = ConvNextForImageClassification.from_pretrained("facebook/convnext-tiny-224" ).to(lowercase_ ) UpperCAmelCase_ : List[Any] = self.default_image_processor UpperCAmelCase_ : List[Any] = prepare_img() UpperCAmelCase_ : int = image_processor(images=lowercase_ , return_tensors="pt" ).to(lowercase_ ) # forward pass with torch.no_grad(): UpperCAmelCase_ : List[Any] = model(**lowercase_ ) # verify the logits UpperCAmelCase_ : Tuple = torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , lowercase_ ) UpperCAmelCase_ : str = torch.tensor([-0.02_60, -0.47_39, 0.19_11] ).to(lowercase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase_ , atol=1E-4 ) ) @require_torch class A_ (unittest.TestCase ,lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = (ConvNextBackbone,) if is_torch_available() else () SCREAMING_SNAKE_CASE__ : Tuple = ConvNextConfig SCREAMING_SNAKE_CASE__ : Optional[Any] = False def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = ConvNextModelTester(self )
23
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a = logging.get_logger(__name__) _a = {'vocab_file': 'vocab.json'} _a = { 'vocab_file': { 'mgp-str': 'https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json', } } _a = {'mgp-str': 27} class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : List[str] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , lowercase_ , lowercase_="[GO]" , lowercase_="[GO]" , lowercase_="[s]" , lowercase_="[GO]" , **lowercase_ ): """simple docstring""" super().__init__( unk_token=lowercase_ , bos_token=lowercase_ , eos_token=lowercase_ , pad_token=lowercase_ , **lowercase_ , ) with open(lowercase_ , encoding="utf-8" ) as vocab_handle: UpperCAmelCase_ : Dict = json.load(lowercase_ ) UpperCAmelCase_ : Dict = {v: k for k, v in self.vocab.items()} @property def UpperCamelCase__ ( self ): """simple docstring""" return len(self.vocab ) def UpperCamelCase__ ( self ): """simple docstring""" return dict(self.vocab , **self.added_tokens_encoder ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = [] for s in text: char_tokens.extend(lowercase_ ) return char_tokens def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.vocab.get(lowercase_ , self.vocab.get(self.unk_token ) ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.decoder.get(lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if not os.path.isdir(lowercase_ ): logger.error("Vocabulary path ({}) should be a directory".format(lowercase_ ) ) return UpperCAmelCase_ : Optional[int] = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) with open(lowercase_ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=lowercase_ , ensure_ascii=lowercase_ ) + "\n" ) return (vocab_file,)
23
1
"""simple docstring""" from __future__ import annotations def __a ( __lowerCamelCase ): UpperCAmelCase_ : Dict = [True] * limit UpperCAmelCase_ : Optional[int] = False UpperCAmelCase_ : int = False UpperCAmelCase_ : int = True for i in range(3, int(limit**0.5 + 1 ), 2 ): UpperCAmelCase_ : List[Any] = i * 2 while index < limit: UpperCAmelCase_ : Union[str, Any] = False UpperCAmelCase_ : str = index + i UpperCAmelCase_ : Union[str, Any] = [2] for i in range(3, __lowerCamelCase, 2 ): if is_prime[i]: primes.append(__lowerCamelCase ) return primes def __a ( __lowerCamelCase = 100_0000 ): UpperCAmelCase_ : Optional[Any] = prime_sieve(__lowerCamelCase ) UpperCAmelCase_ : int = 0 UpperCAmelCase_ : Optional[int] = 0 for i in range(len(__lowerCamelCase ) ): for j in range(i + length, len(__lowerCamelCase ) ): UpperCAmelCase_ : Optional[Any] = sum(primes[i:j] ) if sol >= ceiling: break if sol in primes: UpperCAmelCase_ : List[str] = j - i UpperCAmelCase_ : List[Any] = sol return largest if __name__ == "__main__": print(f"""{solution() = }""")
23
"""simple docstring""" import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency _a = { 'E': 12.70, 'T': 9.06, 'A': 8.17, 'O': 7.51, 'I': 6.97, 'N': 6.75, 'S': 6.33, 'H': 6.09, 'R': 5.99, 'D': 4.25, 'L': 4.03, 'C': 2.78, 'U': 2.76, 'M': 2.41, 'W': 2.36, 'F': 2.23, 'G': 2.02, 'Y': 1.97, 'P': 1.93, 'B': 1.29, 'V': 0.98, 'K': 0.77, 'J': 0.15, 'X': 0.15, 'Q': 0.10, 'Z': 0.07, } _a = 'ETAOINSHRDLCUMWFGYPBVKJXQZ' _a = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def __a ( __lowerCamelCase ): UpperCAmelCase_ : Any = {letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def __a ( __lowerCamelCase ): return x[0] def __a ( __lowerCamelCase ): UpperCAmelCase_ : Any = get_letter_count(__lowerCamelCase ) UpperCAmelCase_ : dict[int, list[str]] = { freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(__lowerCamelCase ) UpperCAmelCase_ : dict[int, str] = {} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find, reverse=__lowerCamelCase ) UpperCAmelCase_ : Any = "".join(freq_to_letter[freq] ) UpperCAmelCase_ : str = list(freq_to_letter_str.items() ) freq_pairs.sort(key=__lowerCamelCase, reverse=__lowerCamelCase ) UpperCAmelCase_ : list[str] = [freq_pair[1] for freq_pair in freq_pairs] return "".join(__lowerCamelCase ) def __a ( __lowerCamelCase ): UpperCAmelCase_ : Any = get_frequency_order(__lowerCamelCase ) UpperCAmelCase_ : int = 0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
23
1
"""simple docstring""" import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Optional[Any] = checkpoint UpperCAmelCase_ : Optional[int] = {} UpperCAmelCase_ : int = vae_state_dict["encoder.conv_in.weight"] UpperCAmelCase_ : Optional[int] = vae_state_dict["encoder.conv_in.bias"] UpperCAmelCase_ : Any = vae_state_dict["encoder.conv_out.weight"] UpperCAmelCase_ : Dict = vae_state_dict["encoder.conv_out.bias"] UpperCAmelCase_ : str = vae_state_dict["encoder.norm_out.weight"] UpperCAmelCase_ : Optional[Any] = vae_state_dict["encoder.norm_out.bias"] UpperCAmelCase_ : Dict = vae_state_dict["decoder.conv_in.weight"] UpperCAmelCase_ : Optional[Any] = vae_state_dict["decoder.conv_in.bias"] UpperCAmelCase_ : Tuple = vae_state_dict["decoder.conv_out.weight"] UpperCAmelCase_ : Tuple = vae_state_dict["decoder.conv_out.bias"] UpperCAmelCase_ : Union[str, Any] = vae_state_dict["decoder.norm_out.weight"] UpperCAmelCase_ : str = vae_state_dict["decoder.norm_out.bias"] UpperCAmelCase_ : List[str] = vae_state_dict["quant_conv.weight"] UpperCAmelCase_ : int = vae_state_dict["quant_conv.bias"] UpperCAmelCase_ : Dict = vae_state_dict["post_quant_conv.weight"] UpperCAmelCase_ : List[Any] = vae_state_dict["post_quant_conv.bias"] # Retrieves the keys for the encoder down blocks only UpperCAmelCase_ : Optional[int] = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "encoder.down" in layer} ) UpperCAmelCase_ : str = { layer_id: [key for key in vae_state_dict if f"""down.{layer_id}""" in key] for layer_id in range(__lowerCamelCase ) } # Retrieves the keys for the decoder up blocks only UpperCAmelCase_ : int = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "decoder.up" in layer} ) UpperCAmelCase_ : Dict = { layer_id: [key for key in vae_state_dict if f"""up.{layer_id}""" in key] for layer_id in range(__lowerCamelCase ) } for i in range(__lowerCamelCase ): UpperCAmelCase_ : str = [key for key in down_blocks[i] if f"""down.{i}""" in key and f"""down.{i}.downsample""" not in key] if f"""encoder.down.{i}.downsample.conv.weight""" in vae_state_dict: UpperCAmelCase_ : Dict = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.weight""" ) UpperCAmelCase_ : List[Any] = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.bias""" ) UpperCAmelCase_ : str = renew_vae_resnet_paths(__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = {"old": f"""down.{i}.block""", "new": f"""down_blocks.{i}.resnets"""} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) UpperCAmelCase_ : List[Any] = [key for key in vae_state_dict if "encoder.mid.block" in key] UpperCAmelCase_ : List[Any] = 2 for i in range(1, num_mid_res_blocks + 1 ): UpperCAmelCase_ : Optional[int] = [key for key in mid_resnets if f"""encoder.mid.block_{i}""" in key] UpperCAmelCase_ : List[Any] = renew_vae_resnet_paths(__lowerCamelCase ) UpperCAmelCase_ : Dict = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) UpperCAmelCase_ : Any = [key for key in vae_state_dict if "encoder.mid.attn" in key] UpperCAmelCase_ : Any = renew_vae_attention_paths(__lowerCamelCase ) UpperCAmelCase_ : List[Any] = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) conv_attn_to_linear(__lowerCamelCase ) for i in range(__lowerCamelCase ): UpperCAmelCase_ : str = num_up_blocks - 1 - i UpperCAmelCase_ : Optional[Any] = [ key for key in up_blocks[block_id] if f"""up.{block_id}""" in key and f"""up.{block_id}.upsample""" not in key ] if f"""decoder.up.{block_id}.upsample.conv.weight""" in vae_state_dict: UpperCAmelCase_ : Any = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.weight""" ] UpperCAmelCase_ : Optional[Any] = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.bias""" ] UpperCAmelCase_ : List[str] = renew_vae_resnet_paths(__lowerCamelCase ) UpperCAmelCase_ : Tuple = {"old": f"""up.{block_id}.block""", "new": f"""up_blocks.{i}.resnets"""} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) UpperCAmelCase_ : Dict = [key for key in vae_state_dict if "decoder.mid.block" in key] UpperCAmelCase_ : int = 2 for i in range(1, num_mid_res_blocks + 1 ): UpperCAmelCase_ : Optional[Any] = [key for key in mid_resnets if f"""decoder.mid.block_{i}""" in key] UpperCAmelCase_ : Any = renew_vae_resnet_paths(__lowerCamelCase ) UpperCAmelCase_ : Any = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) UpperCAmelCase_ : Tuple = [key for key in vae_state_dict if "decoder.mid.attn" in key] UpperCAmelCase_ : Optional[Any] = renew_vae_attention_paths(__lowerCamelCase ) UpperCAmelCase_ : str = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, additional_replacements=[meta_path], config=__lowerCamelCase ) conv_attn_to_linear(__lowerCamelCase ) return new_checkpoint def __a ( __lowerCamelCase, __lowerCamelCase, ): # Only support V1 UpperCAmelCase_ : int = requests.get( " https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml" ) UpperCAmelCase_ : Dict = io.BytesIO(r.content ) UpperCAmelCase_ : Tuple = OmegaConf.load(__lowerCamelCase ) UpperCAmelCase_ : int = 512 UpperCAmelCase_ : Optional[int] = "cuda" if torch.cuda.is_available() else "cpu" if checkpoint_path.endswith("safetensors" ): from safetensors import safe_open UpperCAmelCase_ : Union[str, Any] = {} with safe_open(__lowerCamelCase, framework="pt", device="cpu" ) as f: for key in f.keys(): UpperCAmelCase_ : Union[str, Any] = f.get_tensor(__lowerCamelCase ) else: UpperCAmelCase_ : Union[str, Any] = torch.load(__lowerCamelCase, map_location=__lowerCamelCase )["state_dict"] # Convert the VAE model. UpperCAmelCase_ : Any = create_vae_diffusers_config(__lowerCamelCase, image_size=__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = custom_convert_ldm_vae_checkpoint(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : str = AutoencoderKL(**__lowerCamelCase ) vae.load_state_dict(__lowerCamelCase ) vae.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument('--vae_pt_path', default=None, type=str, required=True, help='Path to the VAE.pt to convert.') parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the VAE.pt to convert.') _a = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
23
"""simple docstring""" import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) _a = logging.getLogger() def __a ( ): UpperCAmelCase_ : Tuple = argparse.ArgumentParser() parser.add_argument("-f" ) UpperCAmelCase_ : Dict = parser.parse_args() return args.f class A_ (lowercase__ ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = logging.StreamHandler(sys.stdout ) logger.addHandler(lowercase_ ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , "run_glue_deebert.py" ) with patch.object(lowercase_ , "argv" , lowercase_ ): UpperCAmelCase_ : List[str] = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(lowercase_ , 0.6_66 ) @slow @require_torch_non_multi_gpu def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = "\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n ".split() self.run_and_check(lowercase_ ) UpperCAmelCase_ : Optional[Any] = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowercase_ ) UpperCAmelCase_ : Dict = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowercase_ )
23
1
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_mbart import MBartTokenizer else: _a = None _a = logging.get_logger(__name__) _a = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} _a = { 'vocab_file': { 'facebook/mbart-large-en-ro': ( 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model' ), 'facebook/mbart-large-cc25': ( 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model' ), }, 'tokenizer_file': { 'facebook/mbart-large-en-ro': 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/tokenizer.json', 'facebook/mbart-large-cc25': 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/tokenizer.json', }, } _a = { 'facebook/mbart-large-en-ro': 1_024, 'facebook/mbart-large-cc25': 1_024, } # fmt: off _a = ['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN'] class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : List[str] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : List[Any] = ["""input_ids""", """attention_mask"""] SCREAMING_SNAKE_CASE__ : Any = MBartTokenizer SCREAMING_SNAKE_CASE__ : List[int] = [] SCREAMING_SNAKE_CASE__ : List[int] = [] def __init__( self , lowercase_=None , lowercase_=None , lowercase_="<s>" , lowercase_="</s>" , lowercase_="</s>" , lowercase_="<s>" , lowercase_="<unk>" , lowercase_="<pad>" , lowercase_="<mask>" , lowercase_=None , lowercase_=None , lowercase_=None , **lowercase_ , ): """simple docstring""" # Mask token behave like a normal word, i.e. include the space before it UpperCAmelCase_ : int = AddedToken(lowercase_ , lstrip=lowercase_ , rstrip=lowercase_ ) if isinstance(lowercase_ , lowercase_ ) else mask_token super().__init__( vocab_file=lowercase_ , tokenizer_file=lowercase_ , bos_token=lowercase_ , eos_token=lowercase_ , sep_token=lowercase_ , cls_token=lowercase_ , unk_token=lowercase_ , pad_token=lowercase_ , mask_token=lowercase_ , src_lang=lowercase_ , tgt_lang=lowercase_ , additional_special_tokens=lowercase_ , **lowercase_ , ) UpperCAmelCase_ : str = vocab_file UpperCAmelCase_ : Dict = False if not self.vocab_file else True UpperCAmelCase_ : Any = FAIRSEQ_LANGUAGE_CODES.copy() if additional_special_tokens is not None: # Only add those special tokens if they are not already there. _additional_special_tokens.extend( [t for t in additional_special_tokens if t not in _additional_special_tokens] ) self.add_special_tokens({"additional_special_tokens": _additional_special_tokens} ) UpperCAmelCase_ : List[str] = { lang_code: self.convert_tokens_to_ids(lowercase_ ) for lang_code in FAIRSEQ_LANGUAGE_CODES } UpperCAmelCase_ : Optional[Any] = src_lang if src_lang is not None else "en_XX" UpperCAmelCase_ : Any = self.convert_tokens_to_ids(self._src_lang ) UpperCAmelCase_ : Union[str, Any] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def UpperCamelCase__ ( self ): """simple docstring""" return self._src_lang @src_lang.setter def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = [self.sep_token_id] UpperCAmelCase_ : Tuple = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) UpperCAmelCase_ : str = src_lang UpperCAmelCase_ : Optional[Any] = self(lowercase_ , add_special_tokens=lowercase_ , return_tensors=lowercase_ , **lowercase_ ) UpperCAmelCase_ : Optional[int] = self.convert_tokens_to_ids(lowercase_ ) UpperCAmelCase_ : Union[str, Any] = tgt_lang_id return inputs def UpperCamelCase__ ( self , lowercase_ , lowercase_ = "en_XX" , lowercase_ = None , lowercase_ = "ro_RO" , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = src_lang UpperCAmelCase_ : Optional[Any] = tgt_lang return super().prepare_seqaseq_batch(lowercase_ , lowercase_ , **lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" return self.set_src_lang_special_tokens(self.src_lang ) def UpperCamelCase__ ( self ): """simple docstring""" return self.set_tgt_lang_special_tokens(self.tgt_lang ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = self.convert_tokens_to_ids(lowercase_ ) UpperCAmelCase_ : List[str] = [] UpperCAmelCase_ : Dict = [self.eos_token_id, self.cur_lang_code] UpperCAmelCase_ : List[str] = self.convert_ids_to_tokens(self.prefix_tokens ) UpperCAmelCase_ : Optional[Any] = self.convert_ids_to_tokens(self.suffix_tokens ) UpperCAmelCase_ : Union[str, Any] = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : int = self.convert_tokens_to_ids(lowercase_ ) UpperCAmelCase_ : Union[str, Any] = [] UpperCAmelCase_ : Dict = [self.eos_token_id, self.cur_lang_code] UpperCAmelCase_ : Tuple = self.convert_ids_to_tokens(self.prefix_tokens ) UpperCAmelCase_ : int = self.convert_ids_to_tokens(self.suffix_tokens ) UpperCAmelCase_ : str = processors.TemplateProcessing( single=prefix_tokens_str + ["$A"] + suffix_tokens_str , pair=prefix_tokens_str + ["$A", "$B"] + suffix_tokens_str , special_tokens=list(zip(prefix_tokens_str + suffix_tokens_str , self.prefix_tokens + self.suffix_tokens ) ) , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(lowercase_ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory.""" ) return UpperCAmelCase_ : Optional[Any] = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase_ ): copyfile(self.vocab_file , lowercase_ ) return (out_vocab_file,)
23
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _a = {'configuration_unispeech': ['UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP', 'UniSpeechConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST', 'UniSpeechForCTC', 'UniSpeechForPreTraining', 'UniSpeechForSequenceClassification', 'UniSpeechModel', 'UniSpeechPreTrainedModel', ] if TYPE_CHECKING: from .configuration_unispeech import UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP, UniSpeechConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_unispeech import ( UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST, UniSpeechForCTC, UniSpeechForPreTraining, UniSpeechForSequenceClassification, UniSpeechModel, UniSpeechPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
23
1
"""simple docstring""" from maths.is_square_free import is_square_free from maths.prime_factors import prime_factors def __a ( __lowerCamelCase ): UpperCAmelCase_ : str = prime_factors(__lowerCamelCase ) if is_square_free(__lowerCamelCase ): return -1 if len(__lowerCamelCase ) % 2 else 1 return 0 if __name__ == "__main__": import doctest doctest.testmod()
23
"""simple docstring""" from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _a = logging.get_logger(__name__) # pylint: disable=invalid-name _a = '\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained("kandinsky-community/kandinsky-2-2-prior")\n >>> pipe_prior.to("cuda")\n >>> prompt = "red cat, 4k photo"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained("kandinsky-community/kandinsky-2-2-decoder")\n >>> pipe.to("cuda")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save("cat.png")\n ```\n' def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=8 ): UpperCAmelCase_ : List[str] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 UpperCAmelCase_ : Tuple = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class A_ (lowercase__ ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , ): """simple docstring""" super().__init__() self.register_modules( unet=lowercase_ , scheduler=lowercase_ , movq=lowercase_ , ) UpperCAmelCase_ : int = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" if latents is None: UpperCAmelCase_ : Dict = randn_tensor(lowercase_ , generator=lowercase_ , device=lowercase_ , dtype=lowercase_ ) else: if latents.shape != shape: raise ValueError(F"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) UpperCAmelCase_ : str = latents.to(lowercase_ ) UpperCAmelCase_ : Dict = latents * scheduler.init_noise_sigma return latents def UpperCamelCase__ ( self , lowercase_=0 ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) UpperCAmelCase_ : Any = torch.device(F"""cuda:{gpu_id}""" ) UpperCAmelCase_ : int = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowercase_ , lowercase_ ) def UpperCamelCase__ ( self , lowercase_=0 ): """simple docstring""" if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) UpperCAmelCase_ : Any = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=lowercase_ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) UpperCAmelCase_ : List[Any] = None for cpu_offloaded_model in [self.unet, self.movq]: UpperCAmelCase_ , UpperCAmelCase_ : str = cpu_offload_with_hook(lowercase_ , lowercase_ , prev_module_hook=lowercase_ ) # We'll offload the last model manually. UpperCAmelCase_ : Tuple = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCamelCase__ ( self ): """simple docstring""" if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowercase_ , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(lowercase_ ) def __call__( self , lowercase_ , lowercase_ , lowercase_ = 512 , lowercase_ = 512 , lowercase_ = 100 , lowercase_ = 4.0 , lowercase_ = 1 , lowercase_ = None , lowercase_ = None , lowercase_ = "pil" , lowercase_ = True , ): """simple docstring""" UpperCAmelCase_ : str = self._execution_device UpperCAmelCase_ : List[Any] = guidance_scale > 1.0 if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = torch.cat(lowercase_ , dim=0 ) UpperCAmelCase_ : Any = image_embeds.shape[0] * num_images_per_prompt if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : List[Any] = torch.cat(lowercase_ , dim=0 ) if do_classifier_free_guidance: UpperCAmelCase_ : Tuple = image_embeds.repeat_interleave(lowercase_ , dim=0 ) UpperCAmelCase_ : List[str] = negative_image_embeds.repeat_interleave(lowercase_ , dim=0 ) UpperCAmelCase_ : Optional[Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=lowercase_ ) self.scheduler.set_timesteps(lowercase_ , device=lowercase_ ) UpperCAmelCase_ : List[Any] = self.scheduler.timesteps UpperCAmelCase_ : List[str] = self.unet.config.in_channels UpperCAmelCase_ , UpperCAmelCase_ : Dict = downscale_height_and_width(lowercase_ , lowercase_ , self.movq_scale_factor ) # create initial latent UpperCAmelCase_ : int = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , lowercase_ , lowercase_ , lowercase_ , self.scheduler , ) for i, t in enumerate(self.progress_bar(lowercase_ ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase_ : Tuple = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase_ : Union[str, Any] = {"image_embeds": image_embeds} UpperCAmelCase_ : Optional[Any] = self.unet( sample=lowercase_ , timestep=lowercase_ , encoder_hidden_states=lowercase_ , added_cond_kwargs=lowercase_ , return_dict=lowercase_ , )[0] if do_classifier_free_guidance: UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = noise_pred.split(latents.shape[1] , dim=1 ) UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = noise_pred.chunk(2 ) UpperCAmelCase_ , UpperCAmelCase_ : Tuple = variance_pred.chunk(2 ) UpperCAmelCase_ : int = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) UpperCAmelCase_ : str = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): UpperCAmelCase_ , UpperCAmelCase_ : Dict = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase_ : List[str] = self.scheduler.step( lowercase_ , lowercase_ , lowercase_ , generator=lowercase_ , )[0] # post-processing UpperCAmelCase_ : Tuple = self.movq.decode(lowercase_ , force_not_quantize=lowercase_ )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: UpperCAmelCase_ : List[Any] = image * 0.5 + 0.5 UpperCAmelCase_ : int = image.clamp(0 , 1 ) UpperCAmelCase_ : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": UpperCAmelCase_ : Dict = self.numpy_to_pil(lowercase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowercase_ )
23
1
"""simple docstring""" from math import ceil def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Tuple = list(range(0, __lowerCamelCase ) ) UpperCAmelCase_ : Any = [item for sublist in list(device_map.values() ) for item in sublist] # Duplicate check UpperCAmelCase_ : int = [] for i in device_map_blocks: if device_map_blocks.count(__lowerCamelCase ) > 1 and i not in duplicate_blocks: duplicate_blocks.append(__lowerCamelCase ) # Missing blocks UpperCAmelCase_ : List[Any] = [i for i in blocks if i not in device_map_blocks] UpperCAmelCase_ : Dict = [i for i in device_map_blocks if i not in blocks] if len(__lowerCamelCase ) != 0: raise ValueError( "Duplicate attention blocks specified in device_map. Attention blocks must be specified to one device." " These attention blocks were specified more than once: " + str(__lowerCamelCase ) ) if len(__lowerCamelCase ) != 0: raise ValueError( "There are attention blocks for this model that are not specified in the device_map. Add these attention " "blocks to a device on the device_map: " + str(__lowerCamelCase ) ) if len(__lowerCamelCase ) != 0: raise ValueError( "The device_map contains more attention blocks than this model has. Remove these from the device_map:" + str(__lowerCamelCase ) ) def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Union[str, Any] = list(range(__lowerCamelCase ) ) UpperCAmelCase_ : Dict = int(ceil(n_layers / len(__lowerCamelCase ) ) ) UpperCAmelCase_ : str = [layers[i : i + n_blocks] for i in range(0, __lowerCamelCase, __lowerCamelCase )] return dict(zip(__lowerCamelCase, __lowerCamelCase ) )
23
"""simple docstring""" import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _a = logging.get_logger(__name__) _a = { 'facebook/detr-resnet-50': 'https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json', # See all DETR models at https://huggingface.co/models?filter=detr } class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """detr""" SCREAMING_SNAKE_CASE__ : str = ["""past_key_values"""] SCREAMING_SNAKE_CASE__ : Union[str, Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , lowercase_=True , lowercase_=None , lowercase_=3 , lowercase_=100 , lowercase_=6 , lowercase_=2048 , lowercase_=8 , lowercase_=6 , lowercase_=2048 , lowercase_=8 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=True , lowercase_="relu" , lowercase_=256 , lowercase_=0.1 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=0.02 , lowercase_=1.0 , lowercase_=False , lowercase_="sine" , lowercase_="resnet50" , lowercase_=True , lowercase_=False , lowercase_=1 , lowercase_=5 , lowercase_=2 , lowercase_=1 , lowercase_=1 , lowercase_=5 , lowercase_=2 , lowercase_=0.1 , **lowercase_ , ): """simple docstring""" if backbone_config is not None and use_timm_backbone: raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`." ) if not use_timm_backbone: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) UpperCAmelCase_ : Union[str, Any] = CONFIG_MAPPING["resnet"](out_features=["stage4"] ) elif isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = backbone_config.get("model_type" ) UpperCAmelCase_ : int = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase_ : Any = config_class.from_dict(lowercase_ ) # set timm attributes to None UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = None, None, None UpperCAmelCase_ : int = use_timm_backbone UpperCAmelCase_ : int = backbone_config UpperCAmelCase_ : List[Any] = num_channels UpperCAmelCase_ : int = num_queries UpperCAmelCase_ : Union[str, Any] = d_model UpperCAmelCase_ : str = encoder_ffn_dim UpperCAmelCase_ : Tuple = encoder_layers UpperCAmelCase_ : List[Any] = encoder_attention_heads UpperCAmelCase_ : Union[str, Any] = decoder_ffn_dim UpperCAmelCase_ : Optional[Any] = decoder_layers UpperCAmelCase_ : Union[str, Any] = decoder_attention_heads UpperCAmelCase_ : Optional[int] = dropout UpperCAmelCase_ : List[str] = attention_dropout UpperCAmelCase_ : Any = activation_dropout UpperCAmelCase_ : str = activation_function UpperCAmelCase_ : Tuple = init_std UpperCAmelCase_ : Optional[Any] = init_xavier_std UpperCAmelCase_ : Optional[Any] = encoder_layerdrop UpperCAmelCase_ : Optional[int] = decoder_layerdrop UpperCAmelCase_ : Tuple = encoder_layers UpperCAmelCase_ : int = auxiliary_loss UpperCAmelCase_ : Optional[Any] = position_embedding_type UpperCAmelCase_ : Tuple = backbone UpperCAmelCase_ : Optional[int] = use_pretrained_backbone UpperCAmelCase_ : Dict = dilation # Hungarian matcher UpperCAmelCase_ : Union[str, Any] = class_cost UpperCAmelCase_ : Any = bbox_cost UpperCAmelCase_ : int = giou_cost # Loss coefficients UpperCAmelCase_ : str = mask_loss_coefficient UpperCAmelCase_ : Any = dice_loss_coefficient UpperCAmelCase_ : Optional[Any] = bbox_loss_coefficient UpperCAmelCase_ : List[str] = giou_loss_coefficient UpperCAmelCase_ : List[Any] = eos_coefficient super().__init__(is_encoder_decoder=lowercase_ , **lowercase_ ) @property def UpperCamelCase__ ( self ): """simple docstring""" return self.encoder_attention_heads @property def UpperCamelCase__ ( self ): """simple docstring""" return self.d_model @classmethod def UpperCamelCase__ ( cls , lowercase_ , **lowercase_ ): """simple docstring""" return cls(backbone_config=lowercase_ , **lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: UpperCAmelCase_ : Union[str, Any] = self.backbone_config.to_dict() UpperCAmelCase_ : str = self.__class__.model_type return output class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = version.parse("""1.11""" ) @property def UpperCamelCase__ ( self ): """simple docstring""" return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("pixel_mask", {0: "batch"}), ] ) @property def UpperCamelCase__ ( self ): """simple docstring""" return 1E-5 @property def UpperCamelCase__ ( self ): """simple docstring""" return 12
23
1
"""simple docstring""" import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, require_tf, slow if is_tf_available(): import tensorflow as tf if is_tensorflow_text_available(): from transformers.models.bert import TFBertTokenizer _a = ['bert-base-uncased', 'bert-base-cased'] _a = 'hf-internal-testing/tiny-bert-tf-only' if is_tf_available(): class A_ (tf.keras.Model ): '''simple docstring''' def __init__( self , lowercase_ ): """simple docstring""" super().__init__() UpperCAmelCase_ : Any = tokenizer UpperCAmelCase_ : List[Any] = AutoConfig.from_pretrained(lowercase_ ) UpperCAmelCase_ : int = TFAutoModel.from_config(lowercase_ ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = self.tokenizer(lowercase_ ) UpperCAmelCase_ : str = self.bert(**lowercase_ ) return out["pooler_output"] @require_tf @require_tensorflow_text class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" super().setUp() UpperCAmelCase_ : Any = [ BertTokenizer.from_pretrained(lowercase_ ) for checkpoint in (TOKENIZER_CHECKPOINTS * 2) ] # repeat for when fast_bert_tokenizer=false UpperCAmelCase_ : str = [TFBertTokenizer.from_pretrained(lowercase_ ) for checkpoint in TOKENIZER_CHECKPOINTS] + [ TFBertTokenizer.from_pretrained(lowercase_ , use_fast_bert_tokenizer=lowercase_ ) for checkpoint in TOKENIZER_CHECKPOINTS ] assert len(self.tokenizers ) == len(self.tf_tokenizers ) UpperCAmelCase_ : Optional[Any] = [ "This is a straightforward English test sentence.", "This one has some weird characters\rto\nsee\r\nif those\u00E9break things.", "Now we're going to add some Chinese: 一 二 三 一二三", "And some much more rare Chinese: 齉 堃 齉堃", "Je vais aussi écrire en français pour tester les accents", "Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ", ] UpperCAmelCase_ : List[Any] = list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def UpperCamelCase__ ( self ): """simple docstring""" for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in (self.test_sentences, self.paired_sentences): UpperCAmelCase_ : str = tokenizer(lowercase_ , return_tensors="tf" , padding="longest" ) UpperCAmelCase_ : Tuple = tf_tokenizer(lowercase_ ) for key in python_outputs.keys(): self.assertTrue(tf.reduce_all(python_outputs[key].shape == tf_outputs[key].shape ) ) self.assertTrue(tf.reduce_all(tf.cast(python_outputs[key] , tf.intaa ) == tf_outputs[key] ) ) @slow def UpperCamelCase__ ( self ): """simple docstring""" for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase_ : Optional[int] = tf_tokenizer(self.paired_sentences ) UpperCAmelCase_ : Any = tf_tokenizer( text=[sentence[0] for sentence in self.paired_sentences] , text_pair=[sentence[1] for sentence in self.paired_sentences] , ) for key in merged_outputs.keys(): self.assertTrue(tf.reduce_all(tf.cast(merged_outputs[key] , tf.intaa ) == separated_outputs[key] ) ) @slow def UpperCamelCase__ ( self ): """simple docstring""" for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase_ : int = tf.function(lowercase_ ) for test_inputs in (self.test_sentences, self.paired_sentences): UpperCAmelCase_ : str = tf.constant(lowercase_ ) UpperCAmelCase_ : Any = compiled_tokenizer(lowercase_ ) UpperCAmelCase_ : Tuple = tf_tokenizer(lowercase_ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def UpperCamelCase__ ( self ): """simple docstring""" for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase_ : List[Any] = ModelToSave(tokenizer=lowercase_ ) UpperCAmelCase_ : Union[str, Any] = tf.convert_to_tensor(self.test_sentences ) UpperCAmelCase_ : str = model(lowercase_ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: UpperCAmelCase_ : Optional[int] = Path(lowercase_ ) / "saved.model" model.save(lowercase_ ) UpperCAmelCase_ : str = tf.keras.models.load_model(lowercase_ ) UpperCAmelCase_ : Dict = loaded_model(lowercase_ ) # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertLessEqual(tf.reduce_max(tf.abs(out - loaded_output ) ) , 1E-5 )
23
"""simple docstring""" _a = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(100_000)] def __a ( __lowerCamelCase ): UpperCAmelCase_ : Optional[int] = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 10_0000] number //= 10_0000 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution _a = [None] * 10_000_000 _a = True _a = False def __a ( __lowerCamelCase ): if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore UpperCAmelCase_ : Dict = chain(next_number(__lowerCamelCase ) ) UpperCAmelCase_ : List[str] = number_chain while number < 1000_0000: UpperCAmelCase_ : List[Any] = number_chain number *= 10 return number_chain def __a ( __lowerCamelCase = 1000_0000 ): for i in range(1, __lowerCamelCase ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(__lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() print(f"""{solution() = }""")
23
1
"""simple docstring""" import io import json import fsspec import pytest from datasets import Dataset, DatasetDict, Features, NamedSplit, Value from datasets.io.json import JsonDatasetReader, JsonDatasetWriter from ..utils import assert_arrow_memory_doesnt_increase, assert_arrow_memory_increases def __a ( __lowerCamelCase, __lowerCamelCase ): assert isinstance(__lowerCamelCase, __lowerCamelCase ) assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory", [False, True] ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Union[str, Any] = tmp_path / "cache" UpperCAmelCase_ : str = {"col_1": "string", "col_2": "int64", "col_3": "float64"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase_ : str = JsonDatasetReader(__lowerCamelCase, cache_dir=__lowerCamelCase, keep_in_memory=__lowerCamelCase ).read() _check_json_dataset(__lowerCamelCase, __lowerCamelCase ) @pytest.mark.parametrize( "features", [ None, {"col_1": "string", "col_2": "int64", "col_3": "float64"}, {"col_1": "string", "col_2": "string", "col_3": "string"}, {"col_1": "int32", "col_2": "int32", "col_3": "int32"}, {"col_1": "float32", "col_2": "float32", "col_3": "float32"}, ], ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Optional[Any] = tmp_path / "cache" UpperCAmelCase_ : Optional[int] = {"col_1": "string", "col_2": "int64", "col_3": "float64"} UpperCAmelCase_ : int = features.copy() if features else default_expected_features UpperCAmelCase_ : Optional[Any] = ( Features({feature: Value(__lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase_ : Any = JsonDatasetReader(__lowerCamelCase, features=__lowerCamelCase, cache_dir=__lowerCamelCase ).read() _check_json_dataset(__lowerCamelCase, __lowerCamelCase ) @pytest.mark.parametrize( "features", [ None, {"col_3": "float64", "col_1": "string", "col_2": "int64"}, ], ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : List[str] = tmp_path / "cache" UpperCAmelCase_ : Dict = {"col_3": "float64", "col_1": "string", "col_2": "int64"} UpperCAmelCase_ : Any = features.copy() if features else default_expected_features UpperCAmelCase_ : Dict = ( Features({feature: Value(__lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase_ : Dict = JsonDatasetReader(__lowerCamelCase, features=__lowerCamelCase, cache_dir=__lowerCamelCase ).read() assert isinstance(__lowerCamelCase, __lowerCamelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_3", "col_1", "col_2"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype def __a ( __lowerCamelCase, __lowerCamelCase ): # jsonl_312_path features are {"col_3": "float64", "col_1": "string", "col_2": "int64"} UpperCAmelCase_ : str = {"col_2": "int64", "col_3": "float64", "col_1": "string"} UpperCAmelCase_ : str = features.copy() UpperCAmelCase_ : Union[str, Any] = ( Features({feature: Value(__lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase_ : Any = tmp_path / "cache" UpperCAmelCase_ : str = JsonDatasetReader(__lowerCamelCase, features=__lowerCamelCase, cache_dir=__lowerCamelCase ).read() assert isinstance(__lowerCamelCase, __lowerCamelCase ) assert dataset.num_rows == 2 assert dataset.num_columns == 3 assert dataset.column_names == ["col_2", "col_3", "col_1"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("split", [None, NamedSplit("train" ), "train", "test"] ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Dict = tmp_path / "cache" UpperCAmelCase_ : str = {"col_1": "string", "col_2": "int64", "col_3": "float64"} UpperCAmelCase_ : Dict = JsonDatasetReader(__lowerCamelCase, cache_dir=__lowerCamelCase, split=__lowerCamelCase ).read() _check_json_dataset(__lowerCamelCase, __lowerCamelCase ) assert dataset.split == split if split else "train" @pytest.mark.parametrize("path_type", [str, list] ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): if issubclass(__lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Dict = jsonl_path elif issubclass(__lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : List[Any] = [jsonl_path] UpperCAmelCase_ : str = tmp_path / "cache" UpperCAmelCase_ : str = {"col_1": "string", "col_2": "int64", "col_3": "float64"} UpperCAmelCase_ : str = JsonDatasetReader(__lowerCamelCase, cache_dir=__lowerCamelCase ).read() _check_json_dataset(__lowerCamelCase, __lowerCamelCase ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=("train",) ): assert isinstance(__lowerCamelCase, __lowerCamelCase ) for split in splits: UpperCAmelCase_ : List[str] = dataset_dict[split] assert dataset.num_rows == 4 assert dataset.num_columns == 3 assert dataset.column_names == ["col_1", "col_2", "col_3"] for feature, expected_dtype in expected_features.items(): assert dataset.features[feature].dtype == expected_dtype @pytest.mark.parametrize("keep_in_memory", [False, True] ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Tuple = tmp_path / "cache" UpperCAmelCase_ : Optional[Any] = {"col_1": "string", "col_2": "int64", "col_3": "float64"} with assert_arrow_memory_increases() if keep_in_memory else assert_arrow_memory_doesnt_increase(): UpperCAmelCase_ : List[str] = JsonDatasetReader({"train": jsonl_path}, cache_dir=__lowerCamelCase, keep_in_memory=__lowerCamelCase ).read() _check_json_datasetdict(__lowerCamelCase, __lowerCamelCase ) @pytest.mark.parametrize( "features", [ None, {"col_1": "string", "col_2": "int64", "col_3": "float64"}, {"col_1": "string", "col_2": "string", "col_3": "string"}, {"col_1": "int32", "col_2": "int32", "col_3": "int32"}, {"col_1": "float32", "col_2": "float32", "col_3": "float32"}, ], ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : int = tmp_path / "cache" UpperCAmelCase_ : str = {"col_1": "string", "col_2": "int64", "col_3": "float64"} UpperCAmelCase_ : Any = features.copy() if features else default_expected_features UpperCAmelCase_ : Union[str, Any] = ( Features({feature: Value(__lowerCamelCase ) for feature, dtype in features.items()} ) if features is not None else None ) UpperCAmelCase_ : Optional[Any] = JsonDatasetReader({"train": jsonl_path}, features=__lowerCamelCase, cache_dir=__lowerCamelCase ).read() _check_json_datasetdict(__lowerCamelCase, __lowerCamelCase ) @pytest.mark.parametrize("split", [None, NamedSplit("train" ), "train", "test"] ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): if split: UpperCAmelCase_ : Dict = {split: jsonl_path} else: UpperCAmelCase_ : Any = "train" UpperCAmelCase_ : Union[str, Any] = {"train": jsonl_path, "test": jsonl_path} UpperCAmelCase_ : Any = tmp_path / "cache" UpperCAmelCase_ : str = {"col_1": "string", "col_2": "int64", "col_3": "float64"} UpperCAmelCase_ : Dict = JsonDatasetReader(__lowerCamelCase, cache_dir=__lowerCamelCase ).read() _check_json_datasetdict(__lowerCamelCase, __lowerCamelCase, splits=list(path.keys() ) ) assert all(dataset[split].split == split for split in path.keys() ) def __a ( __lowerCamelCase ): return json.load(__lowerCamelCase ) def __a ( __lowerCamelCase ): return [json.loads(__lowerCamelCase ) for line in buffer] class A_ : '''simple docstring''' @pytest.mark.parametrize("lines, load_json_function" , [(True, load_json_lines), (False, load_json)] ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , lines=lowercase_ ).write() buffer.seek(0 ) UpperCAmelCase_ : Dict = load_json_function(lowercase_ ) assert isinstance(lowercase_ , lowercase_ ) assert isinstance(exported_content[0] , lowercase_ ) assert len(lowercase_ ) == 10 @pytest.mark.parametrize( "orient, container, keys, len_at" , [ ("records", list, {"tokens", "labels", "answers", "id"}, None), ("split", dict, {"columns", "data"}, "data"), ("index", dict, set("0123456789" ), None), ("columns", dict, {"tokens", "labels", "answers", "id"}, "tokens"), ("values", list, None, None), ("table", dict, {"schema", "data"}, "data"), ] , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , lines=lowercase_ , orient=lowercase_ ).write() buffer.seek(0 ) UpperCAmelCase_ : Optional[Any] = load_json(lowercase_ ) assert isinstance(lowercase_ , lowercase_ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(lowercase_ , "keys" ) and not hasattr(exported_content[0] , "keys" ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(lowercase_ ) == 10 @pytest.mark.parametrize("lines, load_json_function" , [(True, load_json_lines), (False, load_json)] ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , lines=lowercase_ , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase_ : Optional[Any] = load_json_function(lowercase_ ) assert isinstance(lowercase_ , lowercase_ ) assert isinstance(exported_content[0] , lowercase_ ) assert len(lowercase_ ) == 10 @pytest.mark.parametrize( "orient, container, keys, len_at" , [ ("records", list, {"tokens", "labels", "answers", "id"}, None), ("split", dict, {"columns", "data"}, "data"), ("index", dict, set("0123456789" ), None), ("columns", dict, {"tokens", "labels", "answers", "id"}, "tokens"), ("values", list, None, None), ("table", dict, {"schema", "data"}, "data"), ] , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , lines=lowercase_ , orient=lowercase_ , num_proc=2 ).write() buffer.seek(0 ) UpperCAmelCase_ : List[Any] = load_json(lowercase_ ) assert isinstance(lowercase_ , lowercase_ ) if keys: if container is dict: assert exported_content.keys() == keys else: assert exported_content[0].keys() == keys else: assert not hasattr(lowercase_ , "keys" ) and not hasattr(exported_content[0] , "keys" ) if len_at: assert len(exported_content[len_at] ) == 10 else: assert len(lowercase_ ) == 10 def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" with pytest.raises(lowercase_ ): with io.BytesIO() as buffer: JsonDatasetWriter(lowercase_ , lowercase_ , num_proc=0 ) @pytest.mark.parametrize("compression, extension" , [("gzip", "gz"), ("bz2", "bz2"), ("xz", "xz")] ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = tmp_path_factory.mktemp("data" ) / F"""test.json.{extension}""" UpperCAmelCase_ : str = str(shared_datadir / F"""test_file.json.{extension}""" ) JsonDatasetWriter(lowercase_ , lowercase_ , compression=lowercase_ ).write() with fsspec.open(lowercase_ , "rb" , compression="infer" ) as f: UpperCAmelCase_ : List[str] = f.read() with fsspec.open(lowercase_ , "rb" , compression="infer" ) as f: UpperCAmelCase_ : Dict = f.read() assert exported_content == original_content
23
"""simple docstring""" def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # Return True if there is node that has not iterated. UpperCAmelCase_ : List[Any] = [False] * len(__lowerCamelCase ) UpperCAmelCase_ : Any = [] queue.append(__lowerCamelCase ) UpperCAmelCase_ : Tuple = True while queue: UpperCAmelCase_ : str = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(__lowerCamelCase ) UpperCAmelCase_ : Any = True UpperCAmelCase_ : Union[str, Any] = u return visited[t] def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # This array is filled by BFS and to store path UpperCAmelCase_ : List[str] = [-1] * (len(__lowerCamelCase )) UpperCAmelCase_ : Any = 0 while bfs(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : int = float("Inf" ) UpperCAmelCase_ : Tuple = sink while s != source: # Find the minimum value in select path UpperCAmelCase_ : Tuple = min(__lowerCamelCase, graph[parent[s]][s] ) UpperCAmelCase_ : Dict = parent[s] max_flow += path_flow UpperCAmelCase_ : Optional[Any] = sink while v != source: UpperCAmelCase_ : List[str] = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow UpperCAmelCase_ : Optional[int] = parent[v] return max_flow _a = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] _a , _a = 0, 5 print(ford_fulkerson(graph, source, sink))
23
1
"""simple docstring""" import torch from transformers import CamembertForMaskedLM, CamembertTokenizer def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=5 ): # Adapted from https://github.com/pytorch/fairseq/blob/master/fairseq/models/roberta/hub_interface.py assert masked_input.count("<mask>" ) == 1 UpperCAmelCase_ : str = torch.tensor(tokenizer.encode(__lowerCamelCase, add_special_tokens=__lowerCamelCase ) ).unsqueeze(0 ) # Batch size 1 UpperCAmelCase_ : Any = model(__lowerCamelCase )[0] # The last hidden-state is the first element of the output tuple UpperCAmelCase_ : Optional[int] = (input_ids.squeeze() == tokenizer.mask_token_id).nonzero().item() UpperCAmelCase_ : List[str] = logits[0, masked_index, :] UpperCAmelCase_ : int = logits.softmax(dim=0 ) UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = prob.topk(k=__lowerCamelCase, dim=0 ) UpperCAmelCase_ : int = " ".join( [tokenizer.convert_ids_to_tokens(indices[i].item() ) for i in range(len(__lowerCamelCase ) )] ) UpperCAmelCase_ : int = tokenizer.mask_token UpperCAmelCase_ : Optional[int] = [] for index, predicted_token_bpe in enumerate(topk_predicted_token_bpe.split(" " ) ): UpperCAmelCase_ : Optional[Any] = predicted_token_bpe.replace("\u2581", " " ) if " {0}".format(__lowerCamelCase ) in masked_input: topk_filled_outputs.append( ( masked_input.replace(" {0}".format(__lowerCamelCase ), __lowerCamelCase ), values[index].item(), predicted_token, ) ) else: topk_filled_outputs.append( ( masked_input.replace(__lowerCamelCase, __lowerCamelCase ), values[index].item(), predicted_token, ) ) return topk_filled_outputs _a = CamembertTokenizer.from_pretrained('camembert-base') _a = CamembertForMaskedLM.from_pretrained('camembert-base') model.eval() _a = 'Le camembert est <mask> :)' print(fill_mask(masked_input, model, tokenizer, topk=3))
23
"""simple docstring""" import datasets _a = '\\n@InProceedings{conneau2018xnli,\n author = "Conneau, Alexis\n and Rinott, Ruty\n and Lample, Guillaume\n and Williams, Adina\n and Bowman, Samuel R.\n and Schwenk, Holger\n and Stoyanov, Veselin",\n title = "XNLI: Evaluating Cross-lingual Sentence Representations",\n booktitle = "Proceedings of the 2018 Conference on Empirical Methods\n in Natural Language Processing",\n year = "2018",\n publisher = "Association for Computational Linguistics",\n location = "Brussels, Belgium",\n}\n' _a = '\\nXNLI is a subset of a few thousand examples from MNLI which has been translated\ninto a 14 different languages (some low-ish resource). As with MNLI, the goal is\nto predict textual entailment (does sentence A imply/contradict/neither sentence\nB) and is a classification task (given two sentences, predict one of three\nlabels).\n' _a = '\nComputes XNLI score which is just simple accuracy.\nArgs:\n predictions: Predicted labels.\n references: Ground truth labels.\nReturns:\n \'accuracy\': accuracy\nExamples:\n\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> xnli_metric = datasets.load_metric("xnli")\n >>> results = xnli_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n' def __a ( __lowerCamelCase, __lowerCamelCase ): return (preds == labels).mean() @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64" if self.config_name != "sts-b" else "float32" ), "references": datasets.Value("int64" if self.config_name != "sts-b" else "float32" ), } ) , codebase_urls=[] , reference_urls=[] , format="numpy" , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" return {"accuracy": simple_accuracy(lowercase_ , lowercase_ )}
23
1
"""simple docstring""" from __future__ import annotations def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): if (voltage, current, resistance).count(0 ) != 1: raise ValueError("One and only one argument must be 0" ) if resistance < 0: raise ValueError("Resistance cannot be negative" ) if voltage == 0: return {"voltage": float(current * resistance )} elif current == 0: return {"current": voltage / resistance} elif resistance == 0: return {"resistance": voltage / current} else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
23
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy _a = logging.get_logger(__name__) class A_ (lowercase__ ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = feature_size UpperCAmelCase_ : Any = sampling_rate UpperCAmelCase_ : Any = padding_value UpperCAmelCase_ : str = kwargs.pop("padding_side" , "right" ) UpperCAmelCase_ : List[str] = kwargs.pop("return_attention_mask" , lowercase_ ) super().__init__(**lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = True , lowercase_ = None , lowercase_ = False , lowercase_ = None , lowercase_ = None , lowercase_ = None , ): """simple docstring""" # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(lowercase_ , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): UpperCAmelCase_ : Dict = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( "You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`" F""" to this method that includes {self.model_input_names[0]}, but you provided""" F""" {list(processed_features.keys() )}""" ) UpperCAmelCase_ : Tuple = processed_features[self.model_input_names[0]] UpperCAmelCase_ : List[str] = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(lowercase_ ) == 0: if return_attention_mask: UpperCAmelCase_ : Union[str, Any] = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch UpperCAmelCase_ : List[str] = required_input[0] if isinstance(lowercase_ , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. UpperCAmelCase_ : Any = 0 while len(required_input[index] ) == 0: index += 1 if index < len(lowercase_ ): UpperCAmelCase_ : Optional[Any] = required_input[index][0] if return_tensors is None: if is_tf_tensor(lowercase_ ): UpperCAmelCase_ : Dict = "tf" elif is_torch_tensor(lowercase_ ): UpperCAmelCase_ : Any = "pt" elif isinstance(lowercase_ , (int, float, list, tuple, np.ndarray) ): UpperCAmelCase_ : str = "np" else: raise ValueError( F"""type of {first_element} unknown: {type(lowercase_ )}. """ "Should be one of a python, numpy, pytorch or tensorflow object." ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): UpperCAmelCase_ : Optional[int] = to_numpy(lowercase_ ) else: UpperCAmelCase_ : List[str] = [to_numpy(lowercase_ ) for v in value] # Convert padding_strategy in PaddingStrategy UpperCAmelCase_ : Dict = self._get_padding_strategies(padding=lowercase_ , max_length=lowercase_ ) UpperCAmelCase_ : str = processed_features[self.model_input_names[0]] UpperCAmelCase_ : int = len(lowercase_ ) if not all(len(lowercase_ ) == batch_size for v in processed_features.values() ): raise ValueError("Some items in the output dictionary have a different batch size than others." ) UpperCAmelCase_ : int = [] for i in range(lowercase_ ): UpperCAmelCase_ : str = {k: v[i] for k, v in processed_features.items()} # truncation UpperCAmelCase_ : List[str] = self._truncate( lowercase_ , max_length=lowercase_ , pad_to_multiple_of=lowercase_ , truncation=lowercase_ , ) truncated_inputs.append(lowercase_ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length UpperCAmelCase_ : str = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) UpperCAmelCase_ : Dict = PaddingStrategy.MAX_LENGTH UpperCAmelCase_ : List[str] = {} for i in range(lowercase_ ): # padding UpperCAmelCase_ : int = self._pad( truncated_inputs[i] , max_length=lowercase_ , padding_strategy=lowercase_ , pad_to_multiple_of=lowercase_ , return_attention_mask=lowercase_ , ) for key, value in outputs.items(): if key not in batch_outputs: UpperCAmelCase_ : Any = [] if value.dtype is np.dtype(np.floataa ): UpperCAmelCase_ : List[Any] = value.astype(np.floataa ) batch_outputs[key].append(lowercase_ ) return BatchFeature(lowercase_ , tensor_type=lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = PaddingStrategy.DO_NOT_PAD , lowercase_ = None , lowercase_ = None , ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: UpperCAmelCase_ : Tuple = len(lowercase_ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCAmelCase_ : Tuple = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCAmelCase_ : Dict = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(lowercase_ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: UpperCAmelCase_ : Optional[int] = np.ones(len(lowercase_ ) , dtype=np.intaa ) if needs_to_be_padded: UpperCAmelCase_ : Dict = max_length - len(lowercase_ ) if self.padding_side == "right": if return_attention_mask: UpperCAmelCase_ : List[Any] = np.pad( processed_features["attention_mask"] , (0, difference) ) UpperCAmelCase_ : Dict = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) UpperCAmelCase_ : Optional[Any] = np.pad( lowercase_ , lowercase_ , "constant" , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: UpperCAmelCase_ : Optional[Any] = np.pad( processed_features["attention_mask"] , (difference, 0) ) UpperCAmelCase_ : Dict = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) UpperCAmelCase_ : str = np.pad( lowercase_ , lowercase_ , "constant" , constant_values=self.padding_value ) else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return processed_features def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , ): """simple docstring""" if not truncation: return processed_features elif truncation and max_length is None: raise ValueError("When setting ``truncation=True``, make sure that ``max_length`` is defined." ) UpperCAmelCase_ : Optional[int] = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCAmelCase_ : Union[str, Any] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCAmelCase_ : Optional[Any] = len(lowercase_ ) > max_length if needs_to_be_truncated: UpperCAmelCase_ : int = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: UpperCAmelCase_ : Dict = processed_features["attention_mask"][:max_length] return processed_features def UpperCamelCase__ ( self , lowercase_=False , lowercase_=None ): """simple docstring""" # Get padding strategy if padding is not False: if padding is True: UpperCAmelCase_ : Dict = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : Optional[Any] = PaddingStrategy(lowercase_ ) elif isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = padding else: UpperCAmelCase_ : str = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F"""When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined""" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( "Asking to pad but the feature_extractor does not have a padding value. Please select a value to use" " as `padding_value`. For example: `feature_extractor.padding_value = 0.0`." ) return padding_strategy
23
1
"""simple docstring""" import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast _a = datasets.utils.logging.get_logger(__name__) @dataclass class A_ (datasets.BuilderConfig ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = 10000 SCREAMING_SNAKE_CASE__ : Optional[List[str]] = None SCREAMING_SNAKE_CASE__ : Optional[datasets.Features] = None class A_ (datasets.ArrowBasedBuilder ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = ParquetConfig def UpperCamelCase__ ( self ): """simple docstring""" return datasets.DatasetInfo(features=self.config.features ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" if not self.config.data_files: raise ValueError(F"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) UpperCAmelCase_ : Any = dl_manager.download_and_extract(self.config.data_files ) if isinstance(lowercase_ , (str, list, tuple) ): UpperCAmelCase_ : Optional[Any] = data_files if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : Dict = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive UpperCAmelCase_ : Tuple = [dl_manager.iter_files(lowercase_ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] UpperCAmelCase_ : Any = [] for split_name, files in data_files.items(): if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : str = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive UpperCAmelCase_ : List[str] = [dl_manager.iter_files(lowercase_ ) for file in files] # Infer features is they are stoed in the arrow schema if self.info.features is None: for file in itertools.chain.from_iterable(lowercase_ ): with open(lowercase_ , "rb" ) as f: UpperCAmelCase_ : Optional[int] = datasets.Features.from_arrow_schema(pq.read_schema(lowercase_ ) ) break splits.append(datasets.SplitGenerator(name=lowercase_ , gen_kwargs={"files": files} ) ) return splits def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" if self.info.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example UpperCAmelCase_ : Tuple = table_cast(lowercase_ , self.info.features.arrow_schema ) return pa_table def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = self.info.features.arrow_schema if self.info.features is not None else None if self.info.features is not None and self.config.columns is not None: if sorted(field.name for field in schema ) != sorted(self.config.columns ): raise ValueError( F"""Tried to load parquet data with columns '{self.config.columns}' with mismatching features '{self.info.features}'""" ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowercase_ ) ): with open(lowercase_ , "rb" ) as f: UpperCAmelCase_ : int = pq.ParquetFile(lowercase_ ) try: for batch_idx, record_batch in enumerate( parquet_file.iter_batches(batch_size=self.config.batch_size , columns=self.config.columns ) ): UpperCAmelCase_ : Union[str, Any] = pa.Table.from_batches([record_batch] ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield F"""{file_idx}_{batch_idx}""", self._cast_table(lowercase_ ) except ValueError as e: logger.error(F"""Failed to read file '{file}' with error {type(lowercase_ )}: {e}""" ) raise
23
"""simple docstring""" import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = torch.nn.Linear(10 , 10 ) UpperCAmelCase_ : List[str] = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase_ : Optional[Any] = Accelerator() UpperCAmelCase_ : Tuple = accelerator.prepare(lowercase_ ) try: pickle.loads(pickle.dumps(lowercase_ ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
23
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _a = { 'configuration_bert': ['BERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BertConfig', 'BertOnnxConfig'], 'tokenization_bert': ['BasicTokenizer', 'BertTokenizer', 'WordpieceTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ['BertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'BERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BertForMaskedLM', 'BertForMultipleChoice', 'BertForNextSentencePrediction', 'BertForPreTraining', 'BertForQuestionAnswering', 'BertForSequenceClassification', 'BertForTokenClassification', 'BertLayer', 'BertLMHeadModel', 'BertModel', 'BertPreTrainedModel', 'load_tf_weights_in_bert', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFBertEmbeddings', 'TFBertForMaskedLM', 'TFBertForMultipleChoice', 'TFBertForNextSentencePrediction', 'TFBertForPreTraining', 'TFBertForQuestionAnswering', 'TFBertForSequenceClassification', 'TFBertForTokenClassification', 'TFBertLMHeadModel', 'TFBertMainLayer', 'TFBertModel', 'TFBertPreTrainedModel', ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ['TFBertTokenizer'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'FlaxBertForCausalLM', 'FlaxBertForMaskedLM', 'FlaxBertForMultipleChoice', 'FlaxBertForNextSentencePrediction', 'FlaxBertForPreTraining', 'FlaxBertForQuestionAnswering', 'FlaxBertForSequenceClassification', 'FlaxBertForTokenClassification', 'FlaxBertModel', 'FlaxBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
23
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = {'ctrl': 'https://huggingface.co/ctrl/resolve/main/config.json'} class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """ctrl""" SCREAMING_SNAKE_CASE__ : Optional[int] = ["""past_key_values"""] SCREAMING_SNAKE_CASE__ : List[str] = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase_=24_6534 , lowercase_=256 , lowercase_=1280 , lowercase_=8192 , lowercase_=48 , lowercase_=16 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=1E-6 , lowercase_=0.02 , lowercase_=True , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Tuple = vocab_size UpperCAmelCase_ : Union[str, Any] = n_positions UpperCAmelCase_ : List[str] = n_embd UpperCAmelCase_ : Dict = n_layer UpperCAmelCase_ : Optional[int] = n_head UpperCAmelCase_ : List[str] = dff UpperCAmelCase_ : Tuple = resid_pdrop UpperCAmelCase_ : Optional[Any] = embd_pdrop UpperCAmelCase_ : str = layer_norm_epsilon UpperCAmelCase_ : List[str] = initializer_range UpperCAmelCase_ : List[str] = use_cache super().__init__(**lowercase_ )
23
1
"""simple docstring""" from typing import Optional from urllib.parse import quote import huggingface_hub as hfh from packaging import version def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase = None ): if version.parse(hfh.__version__ ).release < version.parse("0.11.0" ).release: # old versions of hfh don't url-encode the file path UpperCAmelCase_ : str = quote(__lowerCamelCase ) return hfh.hf_hub_url(__lowerCamelCase, __lowerCamelCase, repo_type="dataset", revision=__lowerCamelCase )
23
"""simple docstring""" def __a ( __lowerCamelCase ): assert isinstance(__lowerCamelCase, __lowerCamelCase ), f"""The input value of [n={number}] is not an integer""" if number == 1: return 2 elif number < 1: UpperCAmelCase_ : str = f"""The input value of [n={number}] has to be > 0""" raise ValueError(__lowerCamelCase ) else: UpperCAmelCase_ : List[str] = sylvester(number - 1 ) UpperCAmelCase_ : List[str] = num - 1 UpperCAmelCase_ : List[str] = num return lower * upper + 1 if __name__ == "__main__": print(f"""The 8th number in Sylvester's sequence: {sylvester(8)}""")
23
1
"""simple docstring""" import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class A_ (unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = inspect.getfile(accelerate.test_utils ) SCREAMING_SNAKE_CASE__ : Tuple = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["""scripts""", """test_cli.py"""] ) SCREAMING_SNAKE_CASE__ : Dict = ["""accelerate""", """launch"""] SCREAMING_SNAKE_CASE__ : Optional[int] = Path.home() / """.cache/huggingface/accelerate""" SCREAMING_SNAKE_CASE__ : Tuple = """default_config.yaml""" SCREAMING_SNAKE_CASE__ : Any = config_folder / config_file SCREAMING_SNAKE_CASE__ : List[str] = config_folder / """_default_config.yaml""" SCREAMING_SNAKE_CASE__ : Optional[Any] = Path("""tests/test_configs""" ) @classmethod def UpperCamelCase__ ( cls ): """simple docstring""" if cls.config_path.is_file(): cls.config_path.rename(cls.changed_path ) @classmethod def UpperCamelCase__ ( cls ): """simple docstring""" if cls.changed_path.is_file(): cls.changed_path.rename(cls.config_path ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = self.base_cmd if torch.cuda.is_available() and (torch.cuda.device_count() > 1): cmd += ["--multi_gpu"] execute_subprocess_async(cmd + [self.test_file_path] , env=os.environ.copy() ) def UpperCamelCase__ ( self ): """simple docstring""" for config in sorted(self.test_config_path.glob("**/*.yaml" ) ): with self.subTest(config_file=lowercase_ ): execute_subprocess_async( self.base_cmd + ["--config_file", str(lowercase_ ), self.test_file_path] , env=os.environ.copy() ) def UpperCamelCase__ ( self ): """simple docstring""" execute_subprocess_async(["accelerate", "test"] , env=os.environ.copy() ) class A_ (unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """test-tpu""" SCREAMING_SNAKE_CASE__ : Dict = """us-central1-a""" SCREAMING_SNAKE_CASE__ : str = """ls""" SCREAMING_SNAKE_CASE__ : Any = ["""accelerate""", """tpu-config"""] SCREAMING_SNAKE_CASE__ : Dict = """cd /usr/share""" SCREAMING_SNAKE_CASE__ : Tuple = """tests/test_samples/test_command_file.sh""" SCREAMING_SNAKE_CASE__ : str = """Running gcloud compute tpus tpu-vm ssh""" def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = run_command( self.cmd + ["--command", self.command, "--tpu_zone", self.tpu_zone, "--tpu_name", self.tpu_name, "--debug"] , return_stdout=lowercase_ , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" , lowercase_ , ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = run_command( self.cmd + [ "--config_file", "tests/test_configs/0_12_0.yaml", "--command", self.command, "--tpu_zone", self.tpu_zone, "--tpu_name", self.tpu_name, "--debug", ] , return_stdout=lowercase_ , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" , lowercase_ , ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--debug"] , return_stdout=lowercase_ ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" , lowercase_ , ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--command", self.command, "--debug"] , return_stdout=lowercase_ , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls --worker all""" , lowercase_ , ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = run_command( self.cmd + [ "--config_file", "tests/test_configs/latest.yaml", "--command", self.command, "--command", "echo \"Hello World\"", "--debug", ] , return_stdout=lowercase_ , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; ls; echo \"Hello World\" --worker all""" , lowercase_ , ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--command_file", self.command_file, "--debug"] , return_stdout=lowercase_ , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" , lowercase_ , ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = run_command( self.cmd + [ "--config_file", "tests/test_configs/0_12_0.yaml", "--command_file", self.command_file, "--tpu_zone", self.tpu_zone, "--tpu_name", self.tpu_name, "--debug", ] , return_stdout=lowercase_ , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; echo \"hello world\"; echo \"this is a second command\" --worker all""" , lowercase_ , ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = run_command( self.cmd + ["--config_file", "tests/test_configs/latest.yaml", "--install_accelerate", "--debug"] , return_stdout=lowercase_ , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate -U; echo \"hello world\"; echo \"this is a second command\" --worker all""" , lowercase_ , ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[str] = run_command( self.cmd + [ "--config_file", "tests/test_configs/latest.yaml", "--install_accelerate", "--accelerate_version", "12.0.0", "--debug", ] , return_stdout=lowercase_ , ) self.assertIn( F"""{self.gcloud} test-tpu --zone us-central1-a --command {self.base_output}; pip install accelerate==12.0.0; echo \"hello world\"; echo \"this is a second command\" --worker all""" , lowercase_ , )
23
"""simple docstring""" import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class A_ (lowercase__ ,lowercase__ ,unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = IFImgaImgSuperResolutionPipeline SCREAMING_SNAKE_CASE__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""width""", """height"""} SCREAMING_SNAKE_CASE__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""original_image"""} ) SCREAMING_SNAKE_CASE__ : List[Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} def UpperCamelCase__ ( self ): """simple docstring""" return self._get_superresolution_dummy_components() def UpperCamelCase__ ( self , lowercase_ , lowercase_=0 ): """simple docstring""" if str(lowercase_ ).startswith("mps" ): UpperCAmelCase_ : Optional[Any] = torch.manual_seed(lowercase_ ) else: UpperCAmelCase_ : Union[str, Any] = torch.Generator(device=lowercase_ ).manual_seed(lowercase_ ) UpperCAmelCase_ : Any = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) UpperCAmelCase_ : Optional[int] = floats_tensor((1, 3, 16, 16) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) UpperCAmelCase_ : int = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def UpperCamelCase__ ( self ): """simple docstring""" # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_save_load_local() def UpperCamelCase__ ( self ): """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
23
1
"""simple docstring""" import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = torch.nn.Linear(10 , 10 ) UpperCAmelCase_ : List[str] = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase_ : Optional[Any] = Accelerator() UpperCAmelCase_ : Tuple = accelerator.prepare(lowercase_ ) try: pickle.loads(pickle.dumps(lowercase_ ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
23
"""simple docstring""" import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = "ylacombe/bark-small" UpperCAmelCase_ : Union[str, Any] = tempfile.mkdtemp() UpperCAmelCase_ : List[str] = "en_speaker_1" UpperCAmelCase_ : Tuple = "This is a test string" UpperCAmelCase_ : List[Any] = "speaker_embeddings_path.json" UpperCAmelCase_ : Any = "speaker_embeddings" def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" return AutoTokenizer.from_pretrained(self.checkpoint , **lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.get_tokenizer() UpperCAmelCase_ : Union[str, Any] = BarkProcessor(tokenizer=lowercase_ ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase_ : Optional[int] = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) UpperCAmelCase_ : Dict = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) UpperCAmelCase_ : Union[str, Any] = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token="(BOS)" , eos_token="(EOS)" , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) UpperCAmelCase_ : int = 35 UpperCAmelCase_ : Optional[Any] = 2 UpperCAmelCase_ : List[Any] = 8 UpperCAmelCase_ : Optional[Any] = { "semantic_prompt": np.ones(lowercase_ ), "coarse_prompt": np.ones((nb_codebooks_coarse, seq_len) ), "fine_prompt": np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset UpperCAmelCase_ : Dict = processor(text=self.input_string , voice_preset=lowercase_ ) UpperCAmelCase_ : List[str] = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowercase_ , np.array([] ) ).tolist() ) # test loading voice preset from npz file UpperCAmelCase_ : Tuple = os.path.join(self.tmpdirname , "file.npz" ) np.savez(lowercase_ , **lowercase_ ) UpperCAmelCase_ : Optional[int] = processor(text=self.input_string , voice_preset=lowercase_ ) UpperCAmelCase_ : List[str] = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowercase_ , np.array([] ) ).tolist() ) # test loading voice preset from the hub UpperCAmelCase_ : Tuple = processor(text=self.input_string , voice_preset=self.voice_preset ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = self.get_tokenizer() UpperCAmelCase_ : Optional[Any] = BarkProcessor(tokenizer=lowercase_ ) UpperCAmelCase_ : Tuple = processor(text=self.input_string ) UpperCAmelCase_ : Union[str, Any] = tokenizer( self.input_string , padding="max_length" , max_length=256 , add_special_tokens=lowercase_ , return_attention_mask=lowercase_ , return_token_type_ids=lowercase_ , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
23
1
"""simple docstring""" import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = (PNDMScheduler,) SCREAMING_SNAKE_CASE__ : str = (("""num_inference_steps""", 50),) def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : int = { "num_train_timesteps": 1000, "beta_start": 0.00_01, "beta_end": 0.02, "beta_schedule": "linear", } config.update(**lowercase_ ) return config def UpperCamelCase__ ( self , lowercase_=0 , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = dict(self.forward_default_kwargs ) UpperCAmelCase_ : List[str] = kwargs.pop("num_inference_steps" , lowercase_ ) UpperCAmelCase_ : Union[str, Any] = self.dummy_sample UpperCAmelCase_ : Dict = 0.1 * sample UpperCAmelCase_ : Dict = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : List[Any] = self.get_scheduler_config(**lowercase_ ) UpperCAmelCase_ : Dict = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals UpperCAmelCase_ : List[Any] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) UpperCAmelCase_ : Optional[int] = scheduler_class.from_pretrained(lowercase_ ) new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals UpperCAmelCase_ : int = dummy_past_residuals[:] UpperCAmelCase_ : List[str] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : str = new_scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" UpperCAmelCase_ : Optional[int] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Dict = new_scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self , lowercase_=0 , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = dict(self.forward_default_kwargs ) UpperCAmelCase_ : str = kwargs.pop("num_inference_steps" , lowercase_ ) UpperCAmelCase_ : Optional[int] = self.dummy_sample UpperCAmelCase_ : List[str] = 0.1 * sample UpperCAmelCase_ : Tuple = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : str = self.get_scheduler_config() UpperCAmelCase_ : Dict = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals (must be after setting timesteps) UpperCAmelCase_ : List[Any] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) UpperCAmelCase_ : Dict = scheduler_class.from_pretrained(lowercase_ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residual (must be after setting timesteps) UpperCAmelCase_ : Optional[Any] = dummy_past_residuals[:] UpperCAmelCase_ : Union[str, Any] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Dict = new_scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" UpperCAmelCase_ : List[str] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : int = new_scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = self.scheduler_classes[0] UpperCAmelCase_ : Union[str, Any] = self.get_scheduler_config(**lowercase_ ) UpperCAmelCase_ : List[Any] = scheduler_class(**lowercase_ ) UpperCAmelCase_ : Tuple = 10 UpperCAmelCase_ : List[str] = self.dummy_model() UpperCAmelCase_ : str = self.dummy_sample_deter scheduler.set_timesteps(lowercase_ ) for i, t in enumerate(scheduler.prk_timesteps ): UpperCAmelCase_ : Tuple = model(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[int] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): UpperCAmelCase_ : Any = model(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ ).prev_sample return sample def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = dict(self.forward_default_kwargs ) UpperCAmelCase_ : Optional[Any] = kwargs.pop("num_inference_steps" , lowercase_ ) for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : Any = self.get_scheduler_config() UpperCAmelCase_ : Tuple = scheduler_class(**lowercase_ ) UpperCAmelCase_ : str = self.dummy_sample UpperCAmelCase_ : List[Any] = 0.1 * sample if num_inference_steps is not None and hasattr(lowercase_ , "set_timesteps" ): scheduler.set_timesteps(lowercase_ ) elif num_inference_steps is not None and not hasattr(lowercase_ , "set_timesteps" ): UpperCAmelCase_ : List[Any] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) UpperCAmelCase_ : List[str] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] UpperCAmelCase_ : List[str] = dummy_past_residuals[:] UpperCAmelCase_ : str = scheduler.step_prk(lowercase_ , 0 , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Any = scheduler.step_prk(lowercase_ , 1 , lowercase_ , **lowercase_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , 0 , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , 1 , lowercase_ , **lowercase_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def UpperCamelCase__ ( self ): """simple docstring""" for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for steps_offset in [0, 1]: self.check_over_configs(steps_offset=lowercase_ ) UpperCAmelCase_ : Optional[int] = self.scheduler_classes[0] UpperCAmelCase_ : int = self.get_scheduler_config(steps_offset=1 ) UpperCAmelCase_ : Optional[Any] = scheduler_class(**lowercase_ ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [901, 851, 851, 801, 801, 751, 751, 701, 701, 651, 651, 601, 601, 501, 401, 301, 201, 101, 1] ) , ) def UpperCamelCase__ ( self ): """simple docstring""" for beta_start, beta_end in zip([0.00_01, 0.0_01] , [0.0_02, 0.02] ): self.check_over_configs(beta_start=lowercase_ , beta_end=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for t in [1, 5, 10]: self.check_over_forward(time_step=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" # earlier version of set_timesteps() caused an error indexing alpha's with inference steps as power of 3 UpperCAmelCase_ : List[Any] = 27 for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : List[Any] = self.dummy_sample UpperCAmelCase_ : Optional[int] = 0.1 * sample UpperCAmelCase_ : List[str] = self.get_scheduler_config() UpperCAmelCase_ : List[str] = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): UpperCAmelCase_ : List[str] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ ).prev_sample def UpperCamelCase__ ( self ): """simple docstring""" with self.assertRaises(lowercase_ ): UpperCAmelCase_ : List[str] = self.scheduler_classes[0] UpperCAmelCase_ : str = self.get_scheduler_config() UpperCAmelCase_ : Tuple = scheduler_class(**lowercase_ ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.full_loop() UpperCAmelCase_ : Any = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Dict = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 1_98.13_18 ) < 1E-2 assert abs(result_mean.item() - 0.25_80 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = self.full_loop(prediction_type="v_prediction" ) UpperCAmelCase_ : str = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Tuple = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 67.39_86 ) < 1E-2 assert abs(result_mean.item() - 0.08_78 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase_ : Union[str, Any] = self.full_loop(set_alpha_to_one=lowercase_ , beta_start=0.01 ) UpperCAmelCase_ : List[Any] = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : int = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 2_30.03_99 ) < 1E-2 assert abs(result_mean.item() - 0.29_95 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase_ : Tuple = self.full_loop(set_alpha_to_one=lowercase_ , beta_start=0.01 ) UpperCAmelCase_ : int = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Tuple = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 1_86.94_82 ) < 1E-2 assert abs(result_mean.item() - 0.24_34 ) < 1E-3
23
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def __a ( __lowerCamelCase, __lowerCamelCase=False ): UpperCAmelCase_ : Optional[int] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""deit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""deit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""deit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""deit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""deit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""deit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""deit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""deit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""deit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""deit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "deit.embeddings.cls_token"), ("dist_token", "deit.embeddings.distillation_token"), ("patch_embed.proj.weight", "deit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "deit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "deit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" UpperCAmelCase_ : Dict = [(pair[0], pair[1][4:]) if pair[1].startswith("deit" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("norm.weight", "deit.layernorm.weight"), ("norm.bias", "deit.layernorm.bias"), ("head.weight", "cls_classifier.weight"), ("head.bias", "cls_classifier.bias"), ("head_dist.weight", "distillation_classifier.weight"), ("head_dist.bias", "distillation_classifier.bias"), ] ) return rename_keys def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=False ): for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase_ : int = "" else: UpperCAmelCase_ : Union[str, Any] = "deit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase_ : Tuple = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) UpperCAmelCase_ : Dict = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ : Union[str, Any] = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase_ : Any = in_proj_bias[: config.hidden_size] UpperCAmelCase_ : Optional[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase_ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase_ : List[Any] = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase_ : str = in_proj_bias[-config.hidden_size :] def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Tuple = dct.pop(__lowerCamelCase ) UpperCAmelCase_ : Tuple = val def __a ( ): UpperCAmelCase_ : Union[str, Any] = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ : str = Image.open(requests.get(__lowerCamelCase, stream=__lowerCamelCase ).raw ) return im @torch.no_grad() def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : List[str] = DeiTConfig() # all deit models have fine-tuned heads UpperCAmelCase_ : Optional[int] = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase_ : Tuple = 1000 UpperCAmelCase_ : str = "huggingface/label-files" UpperCAmelCase_ : str = "imagenet-1k-id2label.json" UpperCAmelCase_ : List[Any] = json.load(open(hf_hub_download(__lowerCamelCase, __lowerCamelCase, repo_type="dataset" ), "r" ) ) UpperCAmelCase_ : List[str] = {int(__lowerCamelCase ): v for k, v in idalabel.items()} UpperCAmelCase_ : Any = idalabel UpperCAmelCase_ : int = {v: k for k, v in idalabel.items()} UpperCAmelCase_ : Any = int(deit_name[-6:-4] ) UpperCAmelCase_ : Dict = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("tiny" ): UpperCAmelCase_ : Any = 192 UpperCAmelCase_ : Union[str, Any] = 768 UpperCAmelCase_ : Union[str, Any] = 12 UpperCAmelCase_ : int = 3 elif deit_name[9:].startswith("small" ): UpperCAmelCase_ : List[str] = 384 UpperCAmelCase_ : List[str] = 1536 UpperCAmelCase_ : Dict = 12 UpperCAmelCase_ : Any = 6 if deit_name[9:].startswith("base" ): pass elif deit_name[4:].startswith("large" ): UpperCAmelCase_ : int = 1024 UpperCAmelCase_ : List[Any] = 4096 UpperCAmelCase_ : Optional[int] = 24 UpperCAmelCase_ : int = 16 # load original model from timm UpperCAmelCase_ : Union[str, Any] = timm.create_model(__lowerCamelCase, pretrained=__lowerCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCAmelCase_ : Optional[Any] = timm_model.state_dict() UpperCAmelCase_ : Tuple = create_rename_keys(__lowerCamelCase, __lowerCamelCase ) for src, dest in rename_keys: rename_key(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) read_in_q_k_v(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) # load HuggingFace model UpperCAmelCase_ : str = DeiTForImageClassificationWithTeacher(__lowerCamelCase ).eval() model.load_state_dict(__lowerCamelCase ) # Check outputs on an image, prepared by DeiTImageProcessor UpperCAmelCase_ : Union[str, Any] = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 UpperCAmelCase_ : Optional[Any] = DeiTImageProcessor(size=__lowerCamelCase, crop_size=config.image_size ) UpperCAmelCase_ : Any = image_processor(images=prepare_img(), return_tensors="pt" ) UpperCAmelCase_ : int = encoding["pixel_values"] UpperCAmelCase_ : Optional[Any] = model(__lowerCamelCase ) UpperCAmelCase_ : Any = timm_model(__lowerCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__lowerCamelCase, outputs.logits, atol=1E-3 ) Path(__lowerCamelCase ).mkdir(exist_ok=__lowerCamelCase ) print(f"""Saving model {deit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCamelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--deit_name', default='vit_deit_base_distilled_patch16_224', type=str, help='Name of the DeiT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _a = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
23
1
"""simple docstring""" def __a ( __lowerCamelCase ): UpperCAmelCase_ , UpperCAmelCase_ : List[str] = [], [] while len(__lowerCamelCase ) > 1: UpperCAmelCase_ , UpperCAmelCase_ : str = min(__lowerCamelCase ), max(__lowerCamelCase ) start.append(__lowerCamelCase ) end.append(__lowerCamelCase ) collection.remove(__lowerCamelCase ) collection.remove(__lowerCamelCase ) end.reverse() return start + collection + end if __name__ == "__main__": _a = input('Enter numbers separated by a comma:\n').strip() _a = [int(item) for item in user_input.split(',')] print(*merge_sort(unsorted), sep=',')
23
"""simple docstring""" import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights UpperCAmelCase_ : Optional[Any] = FlaxDiffusionPipeline.from_pretrained( "hf-internal-testing/tiny-stable-diffusion-pipe" , safety_checker=lowercase_ , cache_dir=lowercase_ ) UpperCAmelCase_ : List[Any] = [t[-1] for t in os.walk(os.path.join(lowercase_ , os.listdir(lowercase_ )[0] , "snapshots" ) )] UpperCAmelCase_ : Dict = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith(".bin" ) for f in files ) @slow @require_flax class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( "hf-internal-testing/tiny-stable-diffusion-pipe" , safety_checker=lowercase_ ) UpperCAmelCase_ : Tuple = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : List[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase_ : List[str] = 4 UpperCAmelCase_ : Tuple = jax.device_count() UpperCAmelCase_ : Optional[int] = num_samples * [prompt] UpperCAmelCase_ : List[Any] = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : int = replicate(lowercase_ ) UpperCAmelCase_ : str = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : Dict = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_51_47_45 ) < 1E-3 assert np.abs(np.abs(lowercase_ , dtype=np.floataa ).sum() - 4_99_47.8_75 ) < 5E-1 UpperCAmelCase_ : List[Any] = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(lowercase_ ) == num_samples def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="flax" , safety_checker=lowercase_ ) UpperCAmelCase_ : Optional[int] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : str = jax.random.PRNGKey(0 ) UpperCAmelCase_ : Union[str, Any] = 50 UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[str] = num_samples * [prompt] UpperCAmelCase_ : Union[str, Any] = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Any = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : int = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_65_24_01) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_38_38_08.2) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ ) UpperCAmelCase_ : Any = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : str = jax.random.PRNGKey(0 ) UpperCAmelCase_ : str = 50 UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : Any = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Dict = replicate(lowercase_ ) UpperCAmelCase_ : str = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Union[str, Any] = shard(lowercase_ ) UpperCAmelCase_ : List[Any] = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : str = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa ) UpperCAmelCase_ : List[Any] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : Dict = jax.random.PRNGKey(0 ) UpperCAmelCase_ : Optional[int] = 50 UpperCAmelCase_ : Optional[int] = jax.device_count() UpperCAmelCase_ : str = num_samples * [prompt] UpperCAmelCase_ : int = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Union[str, Any] = replicate(lowercase_ ) UpperCAmelCase_ : Union[str, Any] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[int] = shard(lowercase_ ) UpperCAmelCase_ : Any = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = FlaxDDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="scaled_linear" , set_alpha_to_one=lowercase_ , steps_offset=1 , ) UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , scheduler=lowercase_ , safety_checker=lowercase_ , ) UpperCAmelCase_ : List[Any] = scheduler.create_state() UpperCAmelCase_ : int = scheduler_state UpperCAmelCase_ : Union[str, Any] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : Optional[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase_ : int = 50 UpperCAmelCase_ : str = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : int = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : int = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = shard(lowercase_ ) UpperCAmelCase_ : Any = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_45_04_39_45) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_34_76_93.5) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : Union[str, Any] = jax.random.split(jax.random.PRNGKey(0 ) , lowercase_ ) UpperCAmelCase_ , UpperCAmelCase_ : str = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ , ) UpperCAmelCase_ : Any = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = pipeline.prepare_inputs(lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : List[Any] = pipeline(lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase_ : int = images[2, 0, 256, 10:17, 1] # With memory efficient attention UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ , use_memory_efficient_attention=lowercase_ , ) UpperCAmelCase_ : str = replicate(lowercase_ ) UpperCAmelCase_ : str = pipeline.prepare_inputs(lowercase_ ) UpperCAmelCase_ : Optional[int] = shard(lowercase_ ) UpperCAmelCase_ : str = pipeline(lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase_ : Optional[int] = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1E-2
23
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = {'ctrl': 'https://huggingface.co/ctrl/resolve/main/config.json'} class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """ctrl""" SCREAMING_SNAKE_CASE__ : Optional[int] = ["""past_key_values"""] SCREAMING_SNAKE_CASE__ : List[str] = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase_=24_6534 , lowercase_=256 , lowercase_=1280 , lowercase_=8192 , lowercase_=48 , lowercase_=16 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=1E-6 , lowercase_=0.02 , lowercase_=True , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Tuple = vocab_size UpperCAmelCase_ : Union[str, Any] = n_positions UpperCAmelCase_ : List[str] = n_embd UpperCAmelCase_ : Dict = n_layer UpperCAmelCase_ : Optional[int] = n_head UpperCAmelCase_ : List[str] = dff UpperCAmelCase_ : Tuple = resid_pdrop UpperCAmelCase_ : Optional[Any] = embd_pdrop UpperCAmelCase_ : str = layer_norm_epsilon UpperCAmelCase_ : List[str] = initializer_range UpperCAmelCase_ : List[str] = use_cache super().__init__(**lowercase_ )
23
"""simple docstring""" from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean _a = 0 _a = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] _a = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right _a = tuple[int, int] class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ): """simple docstring""" UpperCAmelCase_ : int = pos_x UpperCAmelCase_ : List[Any] = pos_y UpperCAmelCase_ : Union[str, Any] = (pos_y, pos_x) UpperCAmelCase_ : Any = goal_x UpperCAmelCase_ : Dict = goal_y UpperCAmelCase_ : Any = g_cost UpperCAmelCase_ : List[str] = parent UpperCAmelCase_ : int = self.calculate_heuristic() UpperCAmelCase_ : Any = self.g_cost + self.h_cost def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.pos_x - self.goal_x UpperCAmelCase_ : Union[str, Any] = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowercase_ ) + abs(lowercase_ ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self , lowercase_ ): """simple docstring""" return self.f_cost < other.f_cost class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Tuple = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , lowercase_ ) UpperCAmelCase_ : List[Any] = Node(goal[1] , goal[0] , goal[1] , goal[0] , 9_9999 , lowercase_ ) UpperCAmelCase_ : str = [self.start] UpperCAmelCase_ : list[Node] = [] UpperCAmelCase_ : int = False def UpperCamelCase__ ( self ): """simple docstring""" while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() UpperCAmelCase_ : List[str] = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(lowercase_ ) self.closed_nodes.append(lowercase_ ) UpperCAmelCase_ : str = self.get_successors(lowercase_ ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowercase_ ) else: # retrieve the best current path UpperCAmelCase_ : Union[str, Any] = self.open_nodes.pop(self.open_nodes.index(lowercase_ ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowercase_ ) else: self.open_nodes.append(lowercase_ ) return [self.start.pos] def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Any = [] for action in delta: UpperCAmelCase_ : str = parent.pos_x + action[1] UpperCAmelCase_ : int = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowercase_ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowercase_ , lowercase_ , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , lowercase_ , ) ) return successors def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = node UpperCAmelCase_ : int = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) UpperCAmelCase_ : Optional[int] = current_node.parent path.reverse() return path class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Dict = AStar(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = AStar(lowercase_ , lowercase_ ) UpperCAmelCase_ : Tuple = False def UpperCamelCase__ ( self ): """simple docstring""" while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() UpperCAmelCase_ : List[str] = self.fwd_astar.open_nodes.pop(0 ) UpperCAmelCase_ : List[Any] = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowercase_ , lowercase_ ) self.fwd_astar.closed_nodes.append(lowercase_ ) self.bwd_astar.closed_nodes.append(lowercase_ ) UpperCAmelCase_ : Tuple = current_bwd_node UpperCAmelCase_ : str = current_fwd_node UpperCAmelCase_ : Dict = { self.fwd_astar: self.fwd_astar.get_successors(lowercase_ ), self.bwd_astar: self.bwd_astar.get_successors(lowercase_ ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowercase_ ) else: # retrieve the best current path UpperCAmelCase_ : List[Any] = astar.open_nodes.pop( astar.open_nodes.index(lowercase_ ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowercase_ ) else: astar.open_nodes.append(lowercase_ ) return [self.fwd_astar.start.pos] def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.fwd_astar.retrace_path(lowercase_ ) UpperCAmelCase_ : int = self.bwd_astar.retrace_path(lowercase_ ) bwd_path.pop() bwd_path.reverse() UpperCAmelCase_ : Any = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] _a = (0, 0) _a = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) _a = time.time() _a = AStar(init, goal) _a = a_star.search() _a = time.time() - start_time print(f"""AStar execution time = {end_time:f} seconds""") _a = time.time() _a = BidirectionalAStar(init, goal) _a = time.time() - bd_start_time print(f"""BidirectionalAStar execution time = {bd_end_time:f} seconds""")
23
1
"""simple docstring""" import argparse import json import os import pickle import shutil import numpy as np import torch from distiller import Distiller from lm_seqs_dataset import LmSeqsDataset from transformers import ( BertConfig, BertForMaskedLM, BertTokenizer, DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer, GPTaConfig, GPTaLMHeadModel, GPTaTokenizer, RobertaConfig, RobertaForMaskedLM, RobertaTokenizer, ) from utils import git_log, init_gpu_params, logger, set_seed _a = { 'distilbert': (DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer), 'roberta': (RobertaConfig, RobertaForMaskedLM, RobertaTokenizer), 'bert': (BertConfig, BertForMaskedLM, BertTokenizer), 'gpt2': (GPTaConfig, GPTaLMHeadModel, GPTaTokenizer), } def __a ( __lowerCamelCase ): assert (args.mlm and args.alpha_mlm > 0.0) or (not args.mlm and args.alpha_mlm == 0.0) assert (args.alpha_mlm > 0.0 and args.alpha_clm == 0.0) or (args.alpha_mlm == 0.0 and args.alpha_clm > 0.0) if args.mlm: assert os.path.isfile(args.token_counts ) assert (args.student_type in ["roberta", "distilbert"]) and (args.teacher_type in ["roberta", "bert"]) else: assert (args.student_type in ["gpt2"]) and (args.teacher_type in ["gpt2"]) assert args.teacher_type == args.student_type or ( args.student_type == "distilbert" and args.teacher_type == "bert" ) assert os.path.isfile(args.student_config ) if args.student_pretrained_weights is not None: assert os.path.isfile(args.student_pretrained_weights ) if args.freeze_token_type_embds: assert args.student_type in ["roberta"] assert args.alpha_ce >= 0.0 assert args.alpha_mlm >= 0.0 assert args.alpha_clm >= 0.0 assert args.alpha_mse >= 0.0 assert args.alpha_cos >= 0.0 assert args.alpha_ce + args.alpha_mlm + args.alpha_clm + args.alpha_mse + args.alpha_cos > 0.0 def __a ( __lowerCamelCase, __lowerCamelCase ): if args.student_type == "roberta": UpperCAmelCase_ : Optional[int] = False elif args.student_type == "gpt2": UpperCAmelCase_ : List[Any] = False def __a ( __lowerCamelCase, __lowerCamelCase ): if args.student_type == "roberta": UpperCAmelCase_ : str = False def __a ( ): UpperCAmelCase_ : Union[str, Any] = argparse.ArgumentParser(description="Training" ) parser.add_argument("--force", action="store_true", help="Overwrite dump_path if it already exists." ) parser.add_argument( "--dump_path", type=__lowerCamelCase, required=__lowerCamelCase, help="The output directory (log, checkpoints, parameters, etc.)" ) parser.add_argument( "--data_file", type=__lowerCamelCase, required=__lowerCamelCase, help="The binarized file (tokenized + tokens_to_ids) and grouped by sequence.", ) parser.add_argument( "--student_type", type=__lowerCamelCase, choices=["distilbert", "roberta", "gpt2"], required=__lowerCamelCase, help="The student type (DistilBERT, RoBERTa).", ) parser.add_argument("--student_config", type=__lowerCamelCase, required=__lowerCamelCase, help="Path to the student configuration." ) parser.add_argument( "--student_pretrained_weights", default=__lowerCamelCase, type=__lowerCamelCase, help="Load student initialization checkpoint." ) parser.add_argument( "--teacher_type", choices=["bert", "roberta", "gpt2"], required=__lowerCamelCase, help="Teacher type (BERT, RoBERTa)." ) parser.add_argument("--teacher_name", type=__lowerCamelCase, required=__lowerCamelCase, help="The teacher model." ) parser.add_argument("--temperature", default=2.0, type=__lowerCamelCase, help="Temperature for the softmax temperature." ) parser.add_argument( "--alpha_ce", default=0.5, type=__lowerCamelCase, help="Linear weight for the distillation loss. Must be >=0." ) parser.add_argument( "--alpha_mlm", default=0.0, type=__lowerCamelCase, help="Linear weight for the MLM loss. Must be >=0. Should be used in conjunction with `mlm` flag.", ) parser.add_argument("--alpha_clm", default=0.5, type=__lowerCamelCase, help="Linear weight for the CLM loss. Must be >=0." ) parser.add_argument("--alpha_mse", default=0.0, type=__lowerCamelCase, help="Linear weight of the MSE loss. Must be >=0." ) parser.add_argument( "--alpha_cos", default=0.0, type=__lowerCamelCase, help="Linear weight of the cosine embedding loss. Must be >=0." ) parser.add_argument( "--mlm", action="store_true", help="The LM step: MLM or CLM. If `mlm` is True, the MLM is used over CLM." ) parser.add_argument( "--mlm_mask_prop", default=0.15, type=__lowerCamelCase, help="Proportion of tokens for which we need to make a prediction.", ) parser.add_argument("--word_mask", default=0.8, type=__lowerCamelCase, help="Proportion of tokens to mask out." ) parser.add_argument("--word_keep", default=0.1, type=__lowerCamelCase, help="Proportion of tokens to keep." ) parser.add_argument("--word_rand", default=0.1, type=__lowerCamelCase, help="Proportion of tokens to randomly replace." ) parser.add_argument( "--mlm_smoothing", default=0.7, type=__lowerCamelCase, help="Smoothing parameter to emphasize more rare tokens (see XLM, similar to word2vec).", ) parser.add_argument("--token_counts", type=__lowerCamelCase, help="The token counts in the data_file for MLM." ) parser.add_argument( "--restrict_ce_to_mask", action="store_true", help="If true, compute the distillation loss only the [MLM] prediction distribution.", ) parser.add_argument( "--freeze_pos_embs", action="store_true", help="Freeze positional embeddings during distillation. For student_type in ['roberta', 'gpt2'] only.", ) parser.add_argument( "--freeze_token_type_embds", action="store_true", help="Freeze token type embeddings during distillation if existent. For student_type in ['roberta'] only.", ) parser.add_argument("--n_epoch", type=__lowerCamelCase, default=3, help="Number of pass on the whole dataset." ) parser.add_argument("--batch_size", type=__lowerCamelCase, default=5, help="Batch size (for each process)." ) parser.add_argument( "--group_by_size", action="store_false", help="If true, group sequences that have similar length into the same batch. Default is true.", ) parser.add_argument( "--gradient_accumulation_steps", type=__lowerCamelCase, default=50, help="Gradient accumulation for larger training batches.", ) parser.add_argument("--warmup_prop", default=0.05, type=__lowerCamelCase, help="Linear warmup proportion." ) parser.add_argument("--weight_decay", default=0.0, type=__lowerCamelCase, help="Weight decay if we apply some." ) parser.add_argument("--learning_rate", default=5E-4, type=__lowerCamelCase, help="The initial learning rate for Adam." ) parser.add_argument("--adam_epsilon", default=1E-6, type=__lowerCamelCase, help="Epsilon for Adam optimizer." ) parser.add_argument("--max_grad_norm", default=5.0, type=__lowerCamelCase, help="Max gradient norm." ) parser.add_argument("--initializer_range", default=0.02, type=__lowerCamelCase, help="Random initialization range." ) parser.add_argument( "--fp16", action="store_true", help="Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit", ) parser.add_argument( "--fp16_opt_level", type=__lowerCamelCase, default="O1", help=( "For fp16: Apex AMP optimization level selected in ['O0', 'O1', 'O2', and 'O3']." "See details at https://nvidia.github.io/apex/amp.html" ), ) parser.add_argument("--n_gpu", type=__lowerCamelCase, default=1, help="Number of GPUs in the node." ) parser.add_argument("--local_rank", type=__lowerCamelCase, default=-1, help="Distributed training - Local rank" ) parser.add_argument("--seed", type=__lowerCamelCase, default=56, help="Random seed" ) parser.add_argument("--log_interval", type=__lowerCamelCase, default=500, help="Tensorboard logging interval." ) parser.add_argument("--checkpoint_interval", type=__lowerCamelCase, default=4000, help="Checkpoint interval." ) UpperCAmelCase_ : Tuple = parser.parse_args() sanity_checks(__lowerCamelCase ) # ARGS # init_gpu_params(__lowerCamelCase ) set_seed(__lowerCamelCase ) if args.is_master: if os.path.exists(args.dump_path ): if not args.force: raise ValueError( f"""Serialization dir {args.dump_path} already exists, but you have not precised wheter to overwrite""" " itUse `--force` if you want to overwrite it" ) else: shutil.rmtree(args.dump_path ) if not os.path.exists(args.dump_path ): os.makedirs(args.dump_path ) logger.info(f"""Experiment will be dumped and logged in {args.dump_path}""" ) # SAVE PARAMS # logger.info(f"""Param: {args}""" ) with open(os.path.join(args.dump_path, "parameters.json" ), "w" ) as f: json.dump(vars(__lowerCamelCase ), __lowerCamelCase, indent=4 ) git_log(args.dump_path ) UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Tuple = MODEL_CLASSES[args.student_type] UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Tuple = MODEL_CLASSES[args.teacher_type] # TOKENIZER # UpperCAmelCase_ : str = teacher_tokenizer_class.from_pretrained(args.teacher_name ) UpperCAmelCase_ : List[Any] = {} for tok_name, tok_symbol in tokenizer.special_tokens_map.items(): UpperCAmelCase_ : Dict = tokenizer.all_special_tokens.index(__lowerCamelCase ) UpperCAmelCase_ : List[str] = tokenizer.all_special_ids[idx] logger.info(f"""Special tokens {special_tok_ids}""" ) UpperCAmelCase_ : Tuple = special_tok_ids UpperCAmelCase_ : List[str] = tokenizer.max_model_input_sizes[args.teacher_name] # DATA LOADER # logger.info(f"""Loading data from {args.data_file}""" ) with open(args.data_file, "rb" ) as fp: UpperCAmelCase_ : List[Any] = pickle.load(__lowerCamelCase ) if args.mlm: logger.info(f"""Loading token counts from {args.token_counts} (already pre-computed)""" ) with open(args.token_counts, "rb" ) as fp: UpperCAmelCase_ : List[Any] = pickle.load(__lowerCamelCase ) UpperCAmelCase_ : List[Any] = np.maximum(__lowerCamelCase, 1 ) ** -args.mlm_smoothing for idx in special_tok_ids.values(): UpperCAmelCase_ : Any = 0.0 # do not predict special tokens UpperCAmelCase_ : str = torch.from_numpy(__lowerCamelCase ) else: UpperCAmelCase_ : Optional[int] = None UpperCAmelCase_ : Dict = LmSeqsDataset(params=__lowerCamelCase, data=__lowerCamelCase ) logger.info("Data loader created." ) # STUDENT # logger.info(f"""Loading student config from {args.student_config}""" ) UpperCAmelCase_ : List[str] = student_config_class.from_pretrained(args.student_config ) UpperCAmelCase_ : Dict = True if args.student_pretrained_weights is not None: logger.info(f"""Loading pretrained weights from {args.student_pretrained_weights}""" ) UpperCAmelCase_ : Union[str, Any] = student_model_class.from_pretrained(args.student_pretrained_weights, config=__lowerCamelCase ) else: UpperCAmelCase_ : Optional[int] = student_model_class(__lowerCamelCase ) if args.n_gpu > 0: student.to(f"""cuda:{args.local_rank}""" ) logger.info("Student loaded." ) # TEACHER # UpperCAmelCase_ : str = teacher_model_class.from_pretrained(args.teacher_name, output_hidden_states=__lowerCamelCase ) if args.n_gpu > 0: teacher.to(f"""cuda:{args.local_rank}""" ) logger.info(f"""Teacher loaded from {args.teacher_name}.""" ) # FREEZING # if args.freeze_pos_embs: freeze_pos_embeddings(__lowerCamelCase, __lowerCamelCase ) if args.freeze_token_type_embds: freeze_token_type_embeddings(__lowerCamelCase, __lowerCamelCase ) # SANITY CHECKS # assert student.config.vocab_size == teacher.config.vocab_size assert student.config.hidden_size == teacher.config.hidden_size assert student.config.max_position_embeddings == teacher.config.max_position_embeddings if args.mlm: assert token_probs.size(0 ) == stu_architecture_config.vocab_size # DISTILLER # torch.cuda.empty_cache() UpperCAmelCase_ : int = Distiller( params=__lowerCamelCase, dataset=__lowerCamelCase, token_probs=__lowerCamelCase, student=__lowerCamelCase, teacher=__lowerCamelCase ) distiller.train() logger.info("Let's go get some drinks." ) if __name__ == "__main__": main()
23
"""simple docstring""" import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = (PNDMScheduler,) SCREAMING_SNAKE_CASE__ : str = (("""num_inference_steps""", 50),) def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : int = { "num_train_timesteps": 1000, "beta_start": 0.00_01, "beta_end": 0.02, "beta_schedule": "linear", } config.update(**lowercase_ ) return config def UpperCamelCase__ ( self , lowercase_=0 , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = dict(self.forward_default_kwargs ) UpperCAmelCase_ : List[str] = kwargs.pop("num_inference_steps" , lowercase_ ) UpperCAmelCase_ : Union[str, Any] = self.dummy_sample UpperCAmelCase_ : Dict = 0.1 * sample UpperCAmelCase_ : Dict = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : List[Any] = self.get_scheduler_config(**lowercase_ ) UpperCAmelCase_ : Dict = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals UpperCAmelCase_ : List[Any] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) UpperCAmelCase_ : Optional[int] = scheduler_class.from_pretrained(lowercase_ ) new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals UpperCAmelCase_ : int = dummy_past_residuals[:] UpperCAmelCase_ : List[str] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : str = new_scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" UpperCAmelCase_ : Optional[int] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Dict = new_scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self , lowercase_=0 , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = dict(self.forward_default_kwargs ) UpperCAmelCase_ : str = kwargs.pop("num_inference_steps" , lowercase_ ) UpperCAmelCase_ : Optional[int] = self.dummy_sample UpperCAmelCase_ : List[str] = 0.1 * sample UpperCAmelCase_ : Tuple = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : str = self.get_scheduler_config() UpperCAmelCase_ : Dict = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals (must be after setting timesteps) UpperCAmelCase_ : List[Any] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) UpperCAmelCase_ : Dict = scheduler_class.from_pretrained(lowercase_ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residual (must be after setting timesteps) UpperCAmelCase_ : Optional[Any] = dummy_past_residuals[:] UpperCAmelCase_ : Union[str, Any] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Dict = new_scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" UpperCAmelCase_ : List[str] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : int = new_scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = self.scheduler_classes[0] UpperCAmelCase_ : Union[str, Any] = self.get_scheduler_config(**lowercase_ ) UpperCAmelCase_ : List[Any] = scheduler_class(**lowercase_ ) UpperCAmelCase_ : Tuple = 10 UpperCAmelCase_ : List[str] = self.dummy_model() UpperCAmelCase_ : str = self.dummy_sample_deter scheduler.set_timesteps(lowercase_ ) for i, t in enumerate(scheduler.prk_timesteps ): UpperCAmelCase_ : Tuple = model(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[int] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): UpperCAmelCase_ : Any = model(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ ).prev_sample return sample def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = dict(self.forward_default_kwargs ) UpperCAmelCase_ : Optional[Any] = kwargs.pop("num_inference_steps" , lowercase_ ) for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : Any = self.get_scheduler_config() UpperCAmelCase_ : Tuple = scheduler_class(**lowercase_ ) UpperCAmelCase_ : str = self.dummy_sample UpperCAmelCase_ : List[Any] = 0.1 * sample if num_inference_steps is not None and hasattr(lowercase_ , "set_timesteps" ): scheduler.set_timesteps(lowercase_ ) elif num_inference_steps is not None and not hasattr(lowercase_ , "set_timesteps" ): UpperCAmelCase_ : List[Any] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) UpperCAmelCase_ : List[str] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] UpperCAmelCase_ : List[str] = dummy_past_residuals[:] UpperCAmelCase_ : str = scheduler.step_prk(lowercase_ , 0 , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Any = scheduler.step_prk(lowercase_ , 1 , lowercase_ , **lowercase_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , 0 , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , 1 , lowercase_ , **lowercase_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def UpperCamelCase__ ( self ): """simple docstring""" for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for steps_offset in [0, 1]: self.check_over_configs(steps_offset=lowercase_ ) UpperCAmelCase_ : Optional[int] = self.scheduler_classes[0] UpperCAmelCase_ : int = self.get_scheduler_config(steps_offset=1 ) UpperCAmelCase_ : Optional[Any] = scheduler_class(**lowercase_ ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [901, 851, 851, 801, 801, 751, 751, 701, 701, 651, 651, 601, 601, 501, 401, 301, 201, 101, 1] ) , ) def UpperCamelCase__ ( self ): """simple docstring""" for beta_start, beta_end in zip([0.00_01, 0.0_01] , [0.0_02, 0.02] ): self.check_over_configs(beta_start=lowercase_ , beta_end=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for t in [1, 5, 10]: self.check_over_forward(time_step=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" # earlier version of set_timesteps() caused an error indexing alpha's with inference steps as power of 3 UpperCAmelCase_ : List[Any] = 27 for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : List[Any] = self.dummy_sample UpperCAmelCase_ : Optional[int] = 0.1 * sample UpperCAmelCase_ : List[str] = self.get_scheduler_config() UpperCAmelCase_ : List[str] = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): UpperCAmelCase_ : List[str] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ ).prev_sample def UpperCamelCase__ ( self ): """simple docstring""" with self.assertRaises(lowercase_ ): UpperCAmelCase_ : List[str] = self.scheduler_classes[0] UpperCAmelCase_ : str = self.get_scheduler_config() UpperCAmelCase_ : Tuple = scheduler_class(**lowercase_ ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.full_loop() UpperCAmelCase_ : Any = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Dict = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 1_98.13_18 ) < 1E-2 assert abs(result_mean.item() - 0.25_80 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = self.full_loop(prediction_type="v_prediction" ) UpperCAmelCase_ : str = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Tuple = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 67.39_86 ) < 1E-2 assert abs(result_mean.item() - 0.08_78 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase_ : Union[str, Any] = self.full_loop(set_alpha_to_one=lowercase_ , beta_start=0.01 ) UpperCAmelCase_ : List[Any] = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : int = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 2_30.03_99 ) < 1E-2 assert abs(result_mean.item() - 0.29_95 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase_ : Tuple = self.full_loop(set_alpha_to_one=lowercase_ , beta_start=0.01 ) UpperCAmelCase_ : int = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Tuple = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 1_86.94_82 ) < 1E-2 assert abs(result_mean.item() - 0.24_34 ) < 1E-3
23
1
"""simple docstring""" import argparse from tax import checkpoints from transformers import AutoConfig, FlaxAutoModelForSeqaSeqLM def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Tuple = AutoConfig.from_pretrained(__lowerCamelCase ) UpperCAmelCase_ : str = FlaxAutoModelForSeqaSeqLM.from_config(config=__lowerCamelCase ) UpperCAmelCase_ : int = checkpoints.load_tax_checkpoint(__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = "wi_0" in tax_model["target"]["encoder"]["layers_0"]["mlp"] if config.model_type == "t5": UpperCAmelCase_ : str = "SelfAttention" if config.model_type == "longt5" and config.encoder_attention_type == "local": UpperCAmelCase_ : Tuple = "LocalSelfAttention" elif config.model_type == "longt5" and config.encoder_attention_type == "transient-global": UpperCAmelCase_ : Any = "TransientGlobalSelfAttention" else: raise ValueError( "Given config is expected to have `model_type='t5'`, or `model_type='longt5` with `encoder_attention_type`" " attribute with a value from ['local', 'transient-global]." ) # Encoder for layer_index in range(config.num_layers ): UpperCAmelCase_ : str = f"""layers_{str(__lowerCamelCase )}""" # Self-Attention UpperCAmelCase_ : Optional[Any] = tax_model["target"]["encoder"][layer_name]["attention"]["key"]["kernel"] UpperCAmelCase_ : Union[str, Any] = tax_model["target"]["encoder"][layer_name]["attention"]["out"]["kernel"] UpperCAmelCase_ : Optional[Any] = tax_model["target"]["encoder"][layer_name]["attention"]["query"]["kernel"] UpperCAmelCase_ : Optional[int] = tax_model["target"]["encoder"][layer_name]["attention"]["value"]["kernel"] # Global input layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": UpperCAmelCase_ : List[str] = tax_model["target"]["encoder"][layer_name]["attention"]["T5LayerNorm_0"]["scale"] # Layer Normalization UpperCAmelCase_ : Dict = tax_model["target"]["encoder"][layer_name]["pre_attention_layer_norm"]["scale"] if split_mlp_wi: UpperCAmelCase_ : Union[str, Any] = tax_model["target"]["encoder"][layer_name]["mlp"]["wi_0"]["kernel"] UpperCAmelCase_ : Any = tax_model["target"]["encoder"][layer_name]["mlp"]["wi_1"]["kernel"] else: UpperCAmelCase_ : str = tax_model["target"]["encoder"][layer_name]["mlp"]["wi"]["kernel"] UpperCAmelCase_ : int = tax_model["target"]["encoder"][layer_name]["mlp"]["wo"]["kernel"] # Layer Normalization UpperCAmelCase_ : List[Any] = tax_model["target"]["encoder"][layer_name]["pre_mlp_layer_norm"]["scale"] # Assigning UpperCAmelCase_ : str = flax_model.params["encoder"]["block"][str(__lowerCamelCase )]["layer"] UpperCAmelCase_ : int = tax_attention_key UpperCAmelCase_ : int = tax_attention_out UpperCAmelCase_ : str = tax_attention_query UpperCAmelCase_ : Union[str, Any] = tax_attention_value UpperCAmelCase_ : str = tax_attention_layer_norm # Global input layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": UpperCAmelCase_ : str = tax_global_layer_norm if split_mlp_wi: UpperCAmelCase_ : List[Any] = tax_mlp_wi_a UpperCAmelCase_ : int = tax_mlp_wi_a else: UpperCAmelCase_ : Any = tax_mlp_wi UpperCAmelCase_ : Dict = tax_mlp_wo UpperCAmelCase_ : Optional[int] = tax_mlp_layer_norm UpperCAmelCase_ : List[Any] = flax_model_encoder_layer_block # Only for layer 0: UpperCAmelCase_ : Tuple = tax_model["target"]["encoder"]["relpos_bias"]["rel_embedding"].T UpperCAmelCase_ : str = tax_encoder_rel_embedding # Side/global relative position_bias + layer norm if config.model_type == "longt5" and config.encoder_attention_type == "transient-global": UpperCAmelCase_ : List[str] = tax_model["target"]["encoder"]["side_relpos_bias"]["rel_embedding"].T UpperCAmelCase_ : Union[str, Any] = tax_encoder_global_rel_embedding # Assigning UpperCAmelCase_ : Dict = tax_model["target"]["encoder"]["encoder_norm"]["scale"] UpperCAmelCase_ : str = tax_encoder_norm # Decoder for layer_index in range(config.num_layers ): UpperCAmelCase_ : Optional[Any] = f"""layers_{str(__lowerCamelCase )}""" # Self-Attention UpperCAmelCase_ : Any = tax_model["target"]["decoder"][layer_name]["self_attention"]["key"]["kernel"] UpperCAmelCase_ : Tuple = tax_model["target"]["decoder"][layer_name]["self_attention"]["out"]["kernel"] UpperCAmelCase_ : Tuple = tax_model["target"]["decoder"][layer_name]["self_attention"]["query"]["kernel"] UpperCAmelCase_ : List[str] = tax_model["target"]["decoder"][layer_name]["self_attention"]["value"]["kernel"] # Layer Normalization UpperCAmelCase_ : Union[str, Any] = tax_model["target"]["decoder"][layer_name]["pre_self_attention_layer_norm"][ "scale" ] # Encoder-Decoder-Attention UpperCAmelCase_ : List[str] = tax_model["target"]["decoder"][layer_name]["encoder_decoder_attention"] UpperCAmelCase_ : Tuple = tax_enc_dec_attention_module["key"]["kernel"] UpperCAmelCase_ : Any = tax_enc_dec_attention_module["out"]["kernel"] UpperCAmelCase_ : Tuple = tax_enc_dec_attention_module["query"]["kernel"] UpperCAmelCase_ : List[str] = tax_enc_dec_attention_module["value"]["kernel"] # Layer Normalization UpperCAmelCase_ : List[Any] = tax_model["target"]["decoder"][layer_name]["pre_cross_attention_layer_norm"]["scale"] # MLP if split_mlp_wi: UpperCAmelCase_ : Dict = tax_model["target"]["decoder"][layer_name]["mlp"]["wi_0"]["kernel"] UpperCAmelCase_ : Tuple = tax_model["target"]["decoder"][layer_name]["mlp"]["wi_1"]["kernel"] else: UpperCAmelCase_ : Dict = tax_model["target"]["decoder"][layer_name]["mlp"]["wi"]["kernel"] UpperCAmelCase_ : str = tax_model["target"]["decoder"][layer_name]["mlp"]["wo"]["kernel"] # Layer Normalization UpperCAmelCase_ : Tuple = tax_model["target"]["decoder"][layer_name]["pre_mlp_layer_norm"]["scale"] # Assigning UpperCAmelCase_ : Optional[Any] = flax_model.params["decoder"]["block"][str(__lowerCamelCase )]["layer"] UpperCAmelCase_ : List[str] = tax_attention_key UpperCAmelCase_ : Optional[Any] = tax_attention_out UpperCAmelCase_ : Optional[int] = tax_attention_query UpperCAmelCase_ : str = tax_attention_value UpperCAmelCase_ : List[Any] = tax_pre_attention_layer_norm UpperCAmelCase_ : Dict = tax_enc_dec_attention_key UpperCAmelCase_ : Any = tax_enc_dec_attention_out UpperCAmelCase_ : List[Any] = tax_enc_dec_attention_query UpperCAmelCase_ : Tuple = tax_enc_dec_attention_value UpperCAmelCase_ : List[Any] = tax_cross_layer_norm if split_mlp_wi: UpperCAmelCase_ : List[Any] = tax_mlp_wi_a UpperCAmelCase_ : Any = tax_mlp_wi_a else: UpperCAmelCase_ : Any = tax_mlp_wi UpperCAmelCase_ : Dict = tax_mlp_wo UpperCAmelCase_ : Optional[int] = txa_mlp_layer_norm UpperCAmelCase_ : List[Any] = flax_model_decoder_layer_block # Decoder Normalization UpperCAmelCase_ : Tuple = tax_model["target"]["decoder"]["decoder_norm"]["scale"] UpperCAmelCase_ : str = txa_decoder_norm # Only for layer 0: UpperCAmelCase_ : Tuple = tax_model["target"]["decoder"]["relpos_bias"]["rel_embedding"].T UpperCAmelCase_ : Tuple = tax_decoder_rel_embedding # Token Embeddings UpperCAmelCase_ : Optional[Any] = tax_model["target"]["token_embedder"]["embedding"] UpperCAmelCase_ : Any = txa_token_embeddings # LM Head (only in v1.1 and LongT5 checkpoints) if "logits_dense" in tax_model["target"]["decoder"]: UpperCAmelCase_ : Tuple = tax_model["target"]["decoder"]["logits_dense"]["kernel"] flax_model.save_pretrained(__lowerCamelCase ) print("T5X Model was sucessfully converted!" ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--t5x_checkpoint_path', default=None, type=str, required=True, help='Path the T5X checkpoint.' ) parser.add_argument('--config_name', default=None, type=str, required=True, help='Config name of LongT5/T5 model.') parser.add_argument( '--flax_dump_folder_path', default=None, type=str, required=True, help='Path to the output FLAX model.' ) _a = parser.parse_args() convert_tax_checkpoint_to_flax(args.tax_checkpoint_path, args.config_name, args.flax_dump_folder_path)
23
"""simple docstring""" import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels _a = object() # For specifying empty leaf dict `{}` _a = object() def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Any = tuple((re.compile(x + "$" ) for x in qs) ) for i in range(len(__lowerCamelCase ) - len(__lowerCamelCase ) + 1 ): UpperCAmelCase_ : List[str] = [x.match(__lowerCamelCase ) for x, y in zip(__lowerCamelCase, ks[i:] )] if matches and all(__lowerCamelCase ): return True return False def __a ( __lowerCamelCase ): def replace(__lowerCamelCase, __lowerCamelCase ): for rule, replacement in rules: if _match(__lowerCamelCase, __lowerCamelCase ): return replacement return val return replace def __a ( ): return [ # embeddings (("transformer", "wpe", "embedding"), P("mp", __lowerCamelCase )), (("transformer", "wte", "embedding"), P("mp", __lowerCamelCase )), # atention (("attention", "(q_proj|k_proj|v_proj)", "kernel"), P(__lowerCamelCase, "mp" )), (("attention", "out_proj", "kernel"), P("mp", __lowerCamelCase )), (("attention", "out_proj", "bias"), None), # mlp (("mlp", "c_fc", "kernel"), P(__lowerCamelCase, "mp" )), (("mlp", "c_fc", "bias"), P("mp" )), (("mlp", "c_proj", "kernel"), P("mp", __lowerCamelCase )), (("mlp", "c_proj", "bias"), None), # layer norms ((r"ln_\d+", "bias"), None), ((r"\d+", r"ln_\d+", "scale"), None), (("ln_f", "bias"), None), (("ln_f", "scale"), None), ] def __a ( __lowerCamelCase ): UpperCAmelCase_ : List[str] = _get_partition_rules() UpperCAmelCase_ : Any = _replacement_rules(__lowerCamelCase ) UpperCAmelCase_ : Any = {k: _unmatched for k in flatten_dict(__lowerCamelCase )} UpperCAmelCase_ : Dict = {k: replace(__lowerCamelCase, __lowerCamelCase ) for k, v in initd.items()} assert _unmatched not in result.values(), "Incomplete partition spec." return freeze(unflatten_dict(__lowerCamelCase ) )
23
1
"""simple docstring""" import argparse import gc import json import os import shutil import warnings import torch from transformers import LlamaConfig, LlamaForCausalLM, LlamaTokenizer try: from transformers import LlamaTokenizerFast except ImportError as e: warnings.warn(e) warnings.warn( 'The converted tokenizer will be the `slow` tokenizer. To use the fast, update your `tokenizers` library and re-run the tokenizer conversion' ) _a = None _a = { '7B': 11_008, '13B': 13_824, '30B': 17_920, '65B': 22_016, '70B': 28_672, } _a = { '7B': 1, '7Bf': 1, '13B': 2, '13Bf': 2, '30B': 4, '65B': 8, '70B': 8, '70Bf': 8, } def __a ( __lowerCamelCase, __lowerCamelCase=1, __lowerCamelCase=256 ): return multiple_of * ((int(ffn_dim_multiplier * int(8 * n / 3 ) ) + multiple_of - 1) // multiple_of) def __a ( __lowerCamelCase ): with open(__lowerCamelCase, "r" ) as f: return json.load(__lowerCamelCase ) def __a ( __lowerCamelCase, __lowerCamelCase ): with open(__lowerCamelCase, "w" ) as f: json.dump(__lowerCamelCase, __lowerCamelCase ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=True ): os.makedirs(__lowerCamelCase, exist_ok=__lowerCamelCase ) UpperCAmelCase_ : Union[str, Any] = os.path.join(__lowerCamelCase, "tmp" ) os.makedirs(__lowerCamelCase, exist_ok=__lowerCamelCase ) UpperCAmelCase_ : List[Any] = read_json(os.path.join(__lowerCamelCase, "params.json" ) ) UpperCAmelCase_ : str = NUM_SHARDS[model_size] UpperCAmelCase_ : Union[str, Any] = params["n_layers"] UpperCAmelCase_ : Union[str, Any] = params["n_heads"] UpperCAmelCase_ : Union[str, Any] = n_heads // num_shards UpperCAmelCase_ : int = params["dim"] UpperCAmelCase_ : Tuple = dim // n_heads UpperCAmelCase_ : Any = 1_0000.0 UpperCAmelCase_ : Optional[int] = 1.0 / (base ** (torch.arange(0, __lowerCamelCase, 2 ).float() / dims_per_head)) if "n_kv_heads" in params: UpperCAmelCase_ : List[str] = params["n_kv_heads"] # for GQA / MQA UpperCAmelCase_ : List[Any] = n_heads_per_shard // num_key_value_heads UpperCAmelCase_ : List[str] = dim // num_key_value_heads else: # compatibility with other checkpoints UpperCAmelCase_ : Dict = n_heads UpperCAmelCase_ : List[Any] = n_heads_per_shard UpperCAmelCase_ : Any = dim # permute for sliced rotary def permute(__lowerCamelCase, __lowerCamelCase=n_heads, __lowerCamelCase=dim, __lowerCamelCase=dim ): return w.view(__lowerCamelCase, dima // n_heads // 2, 2, __lowerCamelCase ).transpose(1, 2 ).reshape(__lowerCamelCase, __lowerCamelCase ) print(f"""Fetching all parameters from the checkpoint at {input_base_path}.""" ) # Load weights if model_size == "7B": # Not sharded # (The sharded implementation would also work, but this is simpler.) UpperCAmelCase_ : Dict = torch.load(os.path.join(__lowerCamelCase, "consolidated.00.pth" ), map_location="cpu" ) else: # Sharded UpperCAmelCase_ : List[Any] = [ torch.load(os.path.join(__lowerCamelCase, f"""consolidated.{i:02d}.pth""" ), map_location="cpu" ) for i in range(__lowerCamelCase ) ] UpperCAmelCase_ : Optional[Any] = 0 UpperCAmelCase_ : Optional[int] = {"weight_map": {}} for layer_i in range(__lowerCamelCase ): UpperCAmelCase_ : List[str] = f"""pytorch_model-{layer_i + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase_ : List[Any] = { f"""model.layers.{layer_i}.self_attn.q_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wq.weight"""] ), f"""model.layers.{layer_i}.self_attn.k_proj.weight""": permute( loaded[f"""layers.{layer_i}.attention.wk.weight"""] ), f"""model.layers.{layer_i}.self_attn.v_proj.weight""": loaded[f"""layers.{layer_i}.attention.wv.weight"""], f"""model.layers.{layer_i}.self_attn.o_proj.weight""": loaded[f"""layers.{layer_i}.attention.wo.weight"""], f"""model.layers.{layer_i}.mlp.gate_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w1.weight"""], f"""model.layers.{layer_i}.mlp.down_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w2.weight"""], f"""model.layers.{layer_i}.mlp.up_proj.weight""": loaded[f"""layers.{layer_i}.feed_forward.w3.weight"""], f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[f"""layers.{layer_i}.attention_norm.weight"""], f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[f"""layers.{layer_i}.ffn_norm.weight"""], } else: # Sharded # Note that attention.w{q,k,v,o}, feed_fordward.w[1,2,3], attention_norm.weight and ffn_norm.weight share # the same storage object, saving attention_norm and ffn_norm will save other weights too, which is # redundant as other weights will be stitched from multiple shards. To avoid that, they are cloned. UpperCAmelCase_ : Union[str, Any] = { f"""model.layers.{layer_i}.input_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.attention_norm.weight""" ].clone(), f"""model.layers.{layer_i}.post_attention_layernorm.weight""": loaded[0][ f"""layers.{layer_i}.ffn_norm.weight""" ].clone(), } UpperCAmelCase_ : Dict = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wq.weight"""].view(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) for i in range(__lowerCamelCase ) ], dim=0, ).reshape(__lowerCamelCase, __lowerCamelCase ) ) UpperCAmelCase_ : List[str] = permute( torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wk.weight"""].view( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) for i in range(__lowerCamelCase ) ], dim=0, ).reshape(__lowerCamelCase, __lowerCamelCase ), __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, ) UpperCAmelCase_ : int = torch.cat( [ loaded[i][f"""layers.{layer_i}.attention.wv.weight"""].view( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) for i in range(__lowerCamelCase ) ], dim=0, ).reshape(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : List[str] = torch.cat( [loaded[i][f"""layers.{layer_i}.attention.wo.weight"""] for i in range(__lowerCamelCase )], dim=1 ) UpperCAmelCase_ : str = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w1.weight"""] for i in range(__lowerCamelCase )], dim=0 ) UpperCAmelCase_ : Dict = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w2.weight"""] for i in range(__lowerCamelCase )], dim=1 ) UpperCAmelCase_ : Optional[Any] = torch.cat( [loaded[i][f"""layers.{layer_i}.feed_forward.w3.weight"""] for i in range(__lowerCamelCase )], dim=0 ) UpperCAmelCase_ : List[Any] = inv_freq for k, v in state_dict.items(): UpperCAmelCase_ : List[str] = filename param_count += v.numel() torch.save(__lowerCamelCase, os.path.join(__lowerCamelCase, __lowerCamelCase ) ) UpperCAmelCase_ : str = f"""pytorch_model-{n_layers + 1}-of-{n_layers + 1}.bin""" if model_size == "7B": # Unsharded UpperCAmelCase_ : List[Any] = { "model.embed_tokens.weight": loaded["tok_embeddings.weight"], "model.norm.weight": loaded["norm.weight"], "lm_head.weight": loaded["output.weight"], } else: UpperCAmelCase_ : Tuple = { "model.norm.weight": loaded[0]["norm.weight"], "model.embed_tokens.weight": torch.cat( [loaded[i]["tok_embeddings.weight"] for i in range(__lowerCamelCase )], dim=1 ), "lm_head.weight": torch.cat([loaded[i]["output.weight"] for i in range(__lowerCamelCase )], dim=0 ), } for k, v in state_dict.items(): UpperCAmelCase_ : Dict = filename param_count += v.numel() torch.save(__lowerCamelCase, os.path.join(__lowerCamelCase, __lowerCamelCase ) ) # Write configs UpperCAmelCase_ : Optional[Any] = {"total_size": param_count * 2} write_json(__lowerCamelCase, os.path.join(__lowerCamelCase, "pytorch_model.bin.index.json" ) ) UpperCAmelCase_ : Optional[Any] = params["ffn_dim_multiplier"] if "ffn_dim_multiplier" in params else 1 UpperCAmelCase_ : List[str] = params["multiple_of"] if "multiple_of" in params else 256 UpperCAmelCase_ : str = LlamaConfig( hidden_size=__lowerCamelCase, intermediate_size=compute_intermediate_size(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ), num_attention_heads=params["n_heads"], num_hidden_layers=params["n_layers"], rms_norm_eps=params["norm_eps"], num_key_value_heads=__lowerCamelCase, ) config.save_pretrained(__lowerCamelCase ) # Make space so we can load the model properly now. del state_dict del loaded gc.collect() print("Loading the checkpoint in a Llama model." ) UpperCAmelCase_ : Tuple = LlamaForCausalLM.from_pretrained(__lowerCamelCase, torch_dtype=torch.floataa, low_cpu_mem_usage=__lowerCamelCase ) # Avoid saving this as part of the config. del model.config._name_or_path print("Saving in the Transformers format." ) model.save_pretrained(__lowerCamelCase, safe_serialization=__lowerCamelCase ) shutil.rmtree(__lowerCamelCase ) def __a ( __lowerCamelCase, __lowerCamelCase ): # Initialize the tokenizer based on the `spm` model UpperCAmelCase_ : Any = LlamaTokenizer if LlamaTokenizerFast is None else LlamaTokenizerFast print(f"""Saving a {tokenizer_class.__name__} to {tokenizer_path}.""" ) UpperCAmelCase_ : List[Any] = tokenizer_class(__lowerCamelCase ) tokenizer.save_pretrained(__lowerCamelCase ) def __a ( ): UpperCAmelCase_ : Tuple = argparse.ArgumentParser() parser.add_argument( "--input_dir", help="Location of LLaMA weights, which contains tokenizer.model and model folders", ) parser.add_argument( "--model_size", choices=["7B", "7Bf", "13B", "13Bf", "30B", "65B", "70B", "70Bf", "tokenizer_only"], ) parser.add_argument( "--output_dir", help="Location to write HF model and tokenizer", ) parser.add_argument("--safe_serialization", type=__lowerCamelCase, help="Whether or not to save using `safetensors`." ) UpperCAmelCase_ : Dict = parser.parse_args() if args.model_size != "tokenizer_only": write_model( model_path=args.output_dir, input_base_path=os.path.join(args.input_dir, args.model_size ), model_size=args.model_size, safe_serialization=args.safe_serialization, ) UpperCAmelCase_ : str = os.path.join(args.input_dir, "tokenizer.model" ) write_tokenizer(args.output_dir, __lowerCamelCase ) if __name__ == "__main__": main()
23
"""simple docstring""" import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow _a = logging.getLogger() @unittest.skip("""Temporarily disable the doc tests.""" ) @require_torch @require_tf @slow class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = True , ): """simple docstring""" UpperCAmelCase_ : List[str] = [file for file in os.listdir(lowercase_ ) if os.path.isfile(os.path.join(lowercase_ , lowercase_ ) )] if identifier is not None: UpperCAmelCase_ : Dict = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(lowercase_ , lowercase_ ): for n_ in n_identifier: UpperCAmelCase_ : str = [file for file in files if n_ not in file] else: UpperCAmelCase_ : Any = [file for file in files if n_identifier not in file] UpperCAmelCase_ : Union[str, Any] = ignore_files or [] ignore_files.append("__init__.py" ) UpperCAmelCase_ : Optional[int] = [file for file in files if file not in ignore_files] for file in files: # Open all files print("Testing" , lowercase_ ) if only_modules: UpperCAmelCase_ : str = file.split("." )[0] try: UpperCAmelCase_ : str = getattr(lowercase_ , lowercase_ ) UpperCAmelCase_ : Tuple = doctest.DocTestSuite(lowercase_ ) UpperCAmelCase_ : int = unittest.TextTestRunner().run(lowercase_ ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: UpperCAmelCase_ : Optional[Any] = doctest.testfile(str(".." / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = Path("src/transformers" ) UpperCAmelCase_ : str = "modeling" UpperCAmelCase_ : Optional[Any] = [ "modeling_ctrl.py", "modeling_tf_ctrl.py", ] self.analyze_directory(lowercase_ , identifier=lowercase_ , ignore_files=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = Path("src/transformers" ) UpperCAmelCase_ : Any = "tokenization" self.analyze_directory(lowercase_ , identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = Path("src/transformers" ) UpperCAmelCase_ : List[Any] = "configuration" self.analyze_directory(lowercase_ , identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = Path("src/transformers" ) UpperCAmelCase_ : List[Any] = ["configuration", "modeling", "tokenization"] self.analyze_directory(lowercase_ , n_identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = Path("docs/source" ) UpperCAmelCase_ : Union[str, Any] = ["favicon.ico"] self.analyze_directory(lowercase_ , ignore_files=lowercase_ , only_modules=lowercase_ )
23
1
"""simple docstring""" import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->UnCLIP class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : torch.FloatTensor SCREAMING_SNAKE_CASE__ : Optional[torch.FloatTensor] = None def __a ( __lowerCamelCase, __lowerCamelCase=0.999, __lowerCamelCase="cosine", ): if alpha_transform_type == "cosine": def alpha_bar_fn(__lowerCamelCase ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(__lowerCamelCase ): return math.exp(t * -12.0 ) else: raise ValueError(f"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) UpperCAmelCase_ : Optional[int] = [] for i in range(__lowerCamelCase ): UpperCAmelCase_ : Dict = i / num_diffusion_timesteps UpperCAmelCase_ : Any = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(__lowerCamelCase ) / alpha_bar_fn(__lowerCamelCase ), __lowerCamelCase ) ) return torch.tensor(__lowerCamelCase, dtype=torch.floataa ) class A_ (lowercase__ ,lowercase__ ): '''simple docstring''' @register_to_config def __init__( self , lowercase_ = 1000 , lowercase_ = "fixed_small_log" , lowercase_ = True , lowercase_ = 1.0 , lowercase_ = "epsilon" , lowercase_ = "squaredcos_cap_v2" , ): """simple docstring""" if beta_schedule != "squaredcos_cap_v2": raise ValueError("UnCLIPScheduler only supports `beta_schedule`: 'squaredcos_cap_v2'" ) UpperCAmelCase_ : Optional[int] = betas_for_alpha_bar(lowercase_ ) UpperCAmelCase_ : Union[str, Any] = 1.0 - self.betas UpperCAmelCase_ : Tuple = torch.cumprod(self.alphas , dim=0 ) UpperCAmelCase_ : Union[str, Any] = torch.tensor(1.0 ) # standard deviation of the initial noise distribution UpperCAmelCase_ : List[str] = 1.0 # setable values UpperCAmelCase_ : List[str] = None UpperCAmelCase_ : str = torch.from_numpy(np.arange(0 , lowercase_ )[::-1].copy() ) UpperCAmelCase_ : str = variance_type def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" return sample def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = num_inference_steps UpperCAmelCase_ : Optional[Any] = (self.config.num_train_timesteps - 1) / (self.num_inference_steps - 1) UpperCAmelCase_ : Dict = (np.arange(0 , lowercase_ ) * step_ratio).round()[::-1].copy().astype(np.intaa ) UpperCAmelCase_ : Union[str, Any] = torch.from_numpy(lowercase_ ).to(lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_=None , lowercase_=None , lowercase_=None ): """simple docstring""" if prev_timestep is None: UpperCAmelCase_ : str = t - 1 UpperCAmelCase_ : List[Any] = self.alphas_cumprod[t] UpperCAmelCase_ : Dict = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one UpperCAmelCase_ : Tuple = 1 - alpha_prod_t UpperCAmelCase_ : List[str] = 1 - alpha_prod_t_prev if prev_timestep == t - 1: UpperCAmelCase_ : List[Any] = self.betas[t] else: UpperCAmelCase_ : List[Any] = 1 - alpha_prod_t / alpha_prod_t_prev # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample UpperCAmelCase_ : Optional[int] = beta_prod_t_prev / beta_prod_t * beta if variance_type is None: UpperCAmelCase_ : str = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small_log": UpperCAmelCase_ : str = torch.log(torch.clamp(lowercase_ , min=1E-2_0 ) ) UpperCAmelCase_ : str = torch.exp(0.5 * variance ) elif variance_type == "learned_range": # NOTE difference with DDPM scheduler UpperCAmelCase_ : Tuple = variance.log() UpperCAmelCase_ : Optional[Any] = beta.log() UpperCAmelCase_ : Optional[int] = (predicted_variance + 1) / 2 UpperCAmelCase_ : List[str] = frac * max_log + (1 - frac) * min_log return variance def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ = None , lowercase_=None , lowercase_ = True , ): """simple docstring""" UpperCAmelCase_ : Dict = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type == "learned_range": UpperCAmelCase_ , UpperCAmelCase_ : Dict = torch.split(lowercase_ , sample.shape[1] , dim=1 ) else: UpperCAmelCase_ : Any = None # 1. compute alphas, betas if prev_timestep is None: UpperCAmelCase_ : Union[str, Any] = t - 1 UpperCAmelCase_ : List[str] = self.alphas_cumprod[t] UpperCAmelCase_ : List[Any] = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one UpperCAmelCase_ : str = 1 - alpha_prod_t UpperCAmelCase_ : Tuple = 1 - alpha_prod_t_prev if prev_timestep == t - 1: UpperCAmelCase_ : Tuple = self.betas[t] UpperCAmelCase_ : Union[str, Any] = self.alphas[t] else: UpperCAmelCase_ : Any = 1 - alpha_prod_t / alpha_prod_t_prev UpperCAmelCase_ : List[str] = 1 - beta # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": UpperCAmelCase_ : Optional[int] = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": UpperCAmelCase_ : List[Any] = model_output else: raise ValueError( F"""prediction_type given as {self.config.prediction_type} must be one of `epsilon` or `sample`""" " for the UnCLIPScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: UpperCAmelCase_ : int = torch.clamp( lowercase_ , -self.config.clip_sample_range , self.config.clip_sample_range ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase_ : Union[str, Any] = (alpha_prod_t_prev ** 0.5 * beta) / beta_prod_t UpperCAmelCase_ : Optional[int] = alpha ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase_ : Any = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise UpperCAmelCase_ : Union[str, Any] = 0 if t > 0: UpperCAmelCase_ : Dict = randn_tensor( model_output.shape , dtype=model_output.dtype , generator=lowercase_ , device=model_output.device ) UpperCAmelCase_ : Any = self._get_variance( lowercase_ , predicted_variance=lowercase_ , prev_timestep=lowercase_ , ) if self.variance_type == "fixed_small_log": UpperCAmelCase_ : Any = variance elif self.variance_type == "learned_range": UpperCAmelCase_ : Dict = (0.5 * variance).exp() else: raise ValueError( F"""variance_type given as {self.variance_type} must be one of `fixed_small_log` or `learned_range`""" " for the UnCLIPScheduler." ) UpperCAmelCase_ : List[Any] = variance * variance_noise UpperCAmelCase_ : Optional[Any] = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return UnCLIPSchedulerOutput(prev_sample=lowercase_ , pred_original_sample=lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , ): """simple docstring""" # Make sure alphas_cumprod and timestep have same device and dtype as original_samples UpperCAmelCase_ : Union[str, Any] = self.alphas_cumprod.to(device=original_samples.device , dtype=original_samples.dtype ) UpperCAmelCase_ : Union[str, Any] = timesteps.to(original_samples.device ) UpperCAmelCase_ : List[str] = alphas_cumprod[timesteps] ** 0.5 UpperCAmelCase_ : List[Any] = sqrt_alpha_prod.flatten() while len(sqrt_alpha_prod.shape ) < len(original_samples.shape ): UpperCAmelCase_ : Dict = sqrt_alpha_prod.unsqueeze(-1 ) UpperCAmelCase_ : List[str] = (1 - alphas_cumprod[timesteps]) ** 0.5 UpperCAmelCase_ : str = sqrt_one_minus_alpha_prod.flatten() while len(sqrt_one_minus_alpha_prod.shape ) < len(original_samples.shape ): UpperCAmelCase_ : Optional[int] = sqrt_one_minus_alpha_prod.unsqueeze(-1 ) UpperCAmelCase_ : str = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples
23
"""simple docstring""" import warnings from ...utils import is_sklearn_available, requires_backends if is_sklearn_available(): from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef _a = ( 'This metric will be removed from the library soon, metrics should be handled with the 🤗 Evaluate ' 'library. You can have a look at this example script for pointers: ' 'https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py' ) def __a ( __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) return (preds == labels).mean() def __a ( __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) UpperCAmelCase_ : Optional[Any] = simple_accuracy(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : List[Any] = fa_score(y_true=__lowerCamelCase, y_pred=__lowerCamelCase ) return { "acc": acc, "f1": fa, "acc_and_f1": (acc + fa) / 2, } def __a ( __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) UpperCAmelCase_ : Any = pearsonr(__lowerCamelCase, __lowerCamelCase )[0] UpperCAmelCase_ : Optional[Any] = spearmanr(__lowerCamelCase, __lowerCamelCase )[0] return { "pearson": pearson_corr, "spearmanr": spearman_corr, "corr": (pearson_corr + spearman_corr) / 2, } def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) assert len(__lowerCamelCase ) == len(__lowerCamelCase ), f"""Predictions and labels have mismatched lengths {len(__lowerCamelCase )} and {len(__lowerCamelCase )}""" if task_name == "cola": return {"mcc": matthews_corrcoef(__lowerCamelCase, __lowerCamelCase )} elif task_name == "sst-2": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "mrpc": return acc_and_fa(__lowerCamelCase, __lowerCamelCase ) elif task_name == "sts-b": return pearson_and_spearman(__lowerCamelCase, __lowerCamelCase ) elif task_name == "qqp": return acc_and_fa(__lowerCamelCase, __lowerCamelCase ) elif task_name == "mnli": return {"mnli/acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "mnli-mm": return {"mnli-mm/acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "qnli": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "rte": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "wnli": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "hans": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} else: raise KeyError(__lowerCamelCase ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) if len(__lowerCamelCase ) != len(__lowerCamelCase ): raise ValueError(f"""Predictions and labels have mismatched lengths {len(__lowerCamelCase )} and {len(__lowerCamelCase )}""" ) if task_name == "xnli": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} else: raise KeyError(__lowerCamelCase )
23
1
"""simple docstring""" from math import factorial _a = {str(d): factorial(d) for d in range(10)} def __a ( __lowerCamelCase ): return sum(DIGIT_FACTORIAL[d] for d in str(__lowerCamelCase ) ) def __a ( ): UpperCAmelCase_ : Tuple = 7 * factorial(9 ) + 1 return sum(i for i in range(3, __lowerCamelCase ) if sum_of_digit_factorial(__lowerCamelCase ) == i ) if __name__ == "__main__": print(f"""{solution() = }""")
23
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a = logging.get_logger(__name__) _a = {'vocab_file': 'vocab.json'} _a = { 'vocab_file': { 'mgp-str': 'https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json', } } _a = {'mgp-str': 27} class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : List[str] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , lowercase_ , lowercase_="[GO]" , lowercase_="[GO]" , lowercase_="[s]" , lowercase_="[GO]" , **lowercase_ ): """simple docstring""" super().__init__( unk_token=lowercase_ , bos_token=lowercase_ , eos_token=lowercase_ , pad_token=lowercase_ , **lowercase_ , ) with open(lowercase_ , encoding="utf-8" ) as vocab_handle: UpperCAmelCase_ : Dict = json.load(lowercase_ ) UpperCAmelCase_ : Dict = {v: k for k, v in self.vocab.items()} @property def UpperCamelCase__ ( self ): """simple docstring""" return len(self.vocab ) def UpperCamelCase__ ( self ): """simple docstring""" return dict(self.vocab , **self.added_tokens_encoder ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = [] for s in text: char_tokens.extend(lowercase_ ) return char_tokens def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.vocab.get(lowercase_ , self.vocab.get(self.unk_token ) ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.decoder.get(lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if not os.path.isdir(lowercase_ ): logger.error("Vocabulary path ({}) should be a directory".format(lowercase_ ) ) return UpperCAmelCase_ : Optional[int] = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) with open(lowercase_ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=lowercase_ , ensure_ascii=lowercase_ ) + "\n" ) return (vocab_file,)
23
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class A_ (metaclass=SCREAMING_SNAKE_CASE_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = ["""torch""", """transformers""", """onnx"""] def __init__( self , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(self , ["torch", "transformers", "onnx"] ) @classmethod def UpperCamelCase__ ( cls , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"] ) @classmethod def UpperCamelCase__ ( cls , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"] ) class A_ (metaclass=SCREAMING_SNAKE_CASE_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = ["""torch""", """transformers""", """onnx"""] def __init__( self , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(self , ["torch", "transformers", "onnx"] ) @classmethod def UpperCamelCase__ ( cls , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"] ) @classmethod def UpperCamelCase__ ( cls , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"] ) class A_ (metaclass=SCREAMING_SNAKE_CASE_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = ["""torch""", """transformers""", """onnx"""] def __init__( self , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(self , ["torch", "transformers", "onnx"] ) @classmethod def UpperCamelCase__ ( cls , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"] ) @classmethod def UpperCamelCase__ ( cls , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"] ) class A_ (metaclass=SCREAMING_SNAKE_CASE_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[Any] = ["""torch""", """transformers""", """onnx"""] def __init__( self , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(self , ["torch", "transformers", "onnx"] ) @classmethod def UpperCamelCase__ ( cls , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"] ) @classmethod def UpperCamelCase__ ( cls , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"] ) class A_ (metaclass=SCREAMING_SNAKE_CASE_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = ["""torch""", """transformers""", """onnx"""] def __init__( self , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(self , ["torch", "transformers", "onnx"] ) @classmethod def UpperCamelCase__ ( cls , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"] ) @classmethod def UpperCamelCase__ ( cls , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"] ) class A_ (metaclass=SCREAMING_SNAKE_CASE_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = ["""torch""", """transformers""", """onnx"""] def __init__( self , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(self , ["torch", "transformers", "onnx"] ) @classmethod def UpperCamelCase__ ( cls , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"] ) @classmethod def UpperCamelCase__ ( cls , *lowercase_ , **lowercase_ ): """simple docstring""" requires_backends(cls , ["torch", "transformers", "onnx"] )
350
"""simple docstring""" import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency _a = { 'E': 12.70, 'T': 9.06, 'A': 8.17, 'O': 7.51, 'I': 6.97, 'N': 6.75, 'S': 6.33, 'H': 6.09, 'R': 5.99, 'D': 4.25, 'L': 4.03, 'C': 2.78, 'U': 2.76, 'M': 2.41, 'W': 2.36, 'F': 2.23, 'G': 2.02, 'Y': 1.97, 'P': 1.93, 'B': 1.29, 'V': 0.98, 'K': 0.77, 'J': 0.15, 'X': 0.15, 'Q': 0.10, 'Z': 0.07, } _a = 'ETAOINSHRDLCUMWFGYPBVKJXQZ' _a = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def __a ( __lowerCamelCase ): UpperCAmelCase_ : Any = {letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def __a ( __lowerCamelCase ): return x[0] def __a ( __lowerCamelCase ): UpperCAmelCase_ : Any = get_letter_count(__lowerCamelCase ) UpperCAmelCase_ : dict[int, list[str]] = { freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(__lowerCamelCase ) UpperCAmelCase_ : dict[int, str] = {} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find, reverse=__lowerCamelCase ) UpperCAmelCase_ : Any = "".join(freq_to_letter[freq] ) UpperCAmelCase_ : str = list(freq_to_letter_str.items() ) freq_pairs.sort(key=__lowerCamelCase, reverse=__lowerCamelCase ) UpperCAmelCase_ : list[str] = [freq_pair[1] for freq_pair in freq_pairs] return "".join(__lowerCamelCase ) def __a ( __lowerCamelCase ): UpperCAmelCase_ : Any = get_frequency_order(__lowerCamelCase ) UpperCAmelCase_ : int = 0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
23
0
"""simple docstring""" def __a ( __lowerCamelCase, __lowerCamelCase ): return int((input_a, input_a).count(0 ) == 0 ) def __a ( ): assert and_gate(0, 0 ) == 0 assert and_gate(0, 1 ) == 0 assert and_gate(1, 0 ) == 0 assert and_gate(1, 1 ) == 1 if __name__ == "__main__": test_and_gate() print(and_gate(1, 0)) print(and_gate(0, 0)) print(and_gate(0, 1)) print(and_gate(1, 1))
351
"""simple docstring""" import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) _a = logging.getLogger() def __a ( ): UpperCAmelCase_ : Tuple = argparse.ArgumentParser() parser.add_argument("-f" ) UpperCAmelCase_ : Dict = parser.parse_args() return args.f class A_ (lowercase__ ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = logging.StreamHandler(sys.stdout ) logger.addHandler(lowercase_ ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , "run_glue_deebert.py" ) with patch.object(lowercase_ , "argv" , lowercase_ ): UpperCAmelCase_ : List[str] = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(lowercase_ , 0.6_66 ) @slow @require_torch_non_multi_gpu def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = "\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n ".split() self.run_and_check(lowercase_ ) UpperCAmelCase_ : Optional[Any] = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowercase_ ) UpperCAmelCase_ : Dict = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowercase_ )
23
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) _a = { 'configuration_longformer': [ 'LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LongformerConfig', 'LongformerOnnxConfig', ], 'tokenization_longformer': ['LongformerTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ['LongformerTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'LongformerForMaskedLM', 'LongformerForMultipleChoice', 'LongformerForQuestionAnswering', 'LongformerForSequenceClassification', 'LongformerForTokenClassification', 'LongformerModel', 'LongformerPreTrainedModel', 'LongformerSelfAttention', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFLongformerForMaskedLM', 'TFLongformerForMultipleChoice', 'TFLongformerForQuestionAnswering', 'TFLongformerForSequenceClassification', 'TFLongformerForTokenClassification', 'TFLongformerModel', 'TFLongformerPreTrainedModel', 'TFLongformerSelfAttention', ] if TYPE_CHECKING: from .configuration_longformer import ( LONGFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, LongformerConfig, LongformerOnnxConfig, ) from .tokenization_longformer import LongformerTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_longformer_fast import LongformerTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_longformer import ( LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, LongformerForMaskedLM, LongformerForMultipleChoice, LongformerForQuestionAnswering, LongformerForSequenceClassification, LongformerForTokenClassification, LongformerModel, LongformerPreTrainedModel, LongformerSelfAttention, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_longformer import ( TF_LONGFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFLongformerForMaskedLM, TFLongformerForMultipleChoice, TFLongformerForQuestionAnswering, TFLongformerForSequenceClassification, TFLongformerForTokenClassification, TFLongformerModel, TFLongformerPreTrainedModel, TFLongformerSelfAttention, ) else: import sys _a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
352
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _a = {'configuration_unispeech': ['UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP', 'UniSpeechConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST', 'UniSpeechForCTC', 'UniSpeechForPreTraining', 'UniSpeechForSequenceClassification', 'UniSpeechModel', 'UniSpeechPreTrainedModel', ] if TYPE_CHECKING: from .configuration_unispeech import UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP, UniSpeechConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_unispeech import ( UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST, UniSpeechForCTC, UniSpeechForPreTraining, UniSpeechForSequenceClassification, UniSpeechModel, UniSpeechPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
23
0
"""simple docstring""" import warnings from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...onnx.utils import compute_effective_axis_dimension from ...utils import TensorType, is_torch_available, logging _a = logging.get_logger(__name__) _a = { 'facebook/bart-large': 'https://huggingface.co/facebook/bart-large/resolve/main/config.json', # See all BART models at https://huggingface.co/models?filter=bart } class A_ (A__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Union[str, Any] = """bart""" SCREAMING_SNAKE_CASE__ : List[Any] = ["""past_key_values"""] SCREAMING_SNAKE_CASE__ : Optional[Any] = {"""num_attention_heads""": """encoder_attention_heads""", """hidden_size""": """d_model"""} def __init__( self , lowercase_=5_0265 , lowercase_=1024 , lowercase_=12 , lowercase_=4096 , lowercase_=16 , lowercase_=12 , lowercase_=4096 , lowercase_=16 , lowercase_=0.0 , lowercase_=0.0 , lowercase_="gelu" , lowercase_=1024 , lowercase_=0.1 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=0.02 , lowercase_=0.0 , lowercase_=False , lowercase_=True , lowercase_=3 , lowercase_=1 , lowercase_=0 , lowercase_=2 , lowercase_=True , lowercase_=2 , lowercase_=2 , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = vocab_size UpperCAmelCase_ : Any = max_position_embeddings UpperCAmelCase_ : List[Any] = d_model UpperCAmelCase_ : Optional[int] = encoder_ffn_dim UpperCAmelCase_ : Optional[Any] = encoder_layers UpperCAmelCase_ : Optional[Any] = encoder_attention_heads UpperCAmelCase_ : str = decoder_ffn_dim UpperCAmelCase_ : Tuple = decoder_layers UpperCAmelCase_ : str = decoder_attention_heads UpperCAmelCase_ : Optional[Any] = dropout UpperCAmelCase_ : Optional[Any] = attention_dropout UpperCAmelCase_ : str = activation_dropout UpperCAmelCase_ : Optional[Any] = activation_function UpperCAmelCase_ : Tuple = init_std UpperCAmelCase_ : List[str] = encoder_layerdrop UpperCAmelCase_ : List[Any] = decoder_layerdrop UpperCAmelCase_ : str = classifier_dropout UpperCAmelCase_ : Dict = use_cache UpperCAmelCase_ : List[Any] = encoder_layers UpperCAmelCase_ : int = scale_embedding # scale factor will be sqrt(d_model) if True super().__init__( num_labels=__snake_case , pad_token_id=__snake_case , bos_token_id=__snake_case , eos_token_id=__snake_case , is_encoder_decoder=__snake_case , decoder_start_token_id=__snake_case , forced_eos_token_id=__snake_case , **__snake_case , ) # ensure backward compatibility for BART CNN models if self.forced_bos_token_id is None and kwargs.get("force_bos_token_to_be_generated" , __snake_case ): UpperCAmelCase_ : Tuple = self.bos_token_id warnings.warn( F"""Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. """ "The config can simply be saved and uploaded again to be fixed." ) class A_ (A__ ): '''simple docstring''' @property def UpperCamelCase__ ( self ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase_ : int = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: UpperCAmelCase_ : Union[str, Any] = {0: "batch"} UpperCAmelCase_ : List[str] = {0: "batch", 1: "past_decoder_sequence + sequence"} else: UpperCAmelCase_ : Any = {0: "batch", 1: "decoder_sequence"} UpperCAmelCase_ : Optional[Any] = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(__snake_case , direction="inputs" ) elif self.task == "causal-lm": # TODO: figure this case out. UpperCAmelCase_ : List[str] = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ] ) if self.use_past: UpperCAmelCase_ , UpperCAmelCase_ : List[str] = self.num_layers for i in range(__snake_case ): UpperCAmelCase_ : Optional[int] = {0: "batch", 2: "past_sequence + sequence"} UpperCAmelCase_ : List[Any] = {0: "batch", 2: "past_sequence + sequence"} else: UpperCAmelCase_ : str = OrderedDict( [ ("input_ids", {0: "batch", 1: "encoder_sequence"}), ("attention_mask", {0: "batch", 1: "encoder_sequence"}), ("decoder_input_ids", {0: "batch", 1: "decoder_sequence"}), ("decoder_attention_mask", {0: "batch", 1: "decoder_sequence"}), ] ) return common_inputs @property def UpperCamelCase__ ( self ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase_ : List[Any] = super().outputs else: UpperCAmelCase_ : Union[str, Any] = super(__snake_case , self ).outputs if self.use_past: UpperCAmelCase_ , UpperCAmelCase_ : str = self.num_layers for i in range(__snake_case ): UpperCAmelCase_ : Union[str, Any] = {0: "batch", 2: "past_sequence + sequence"} UpperCAmelCase_ : List[Any] = {0: "batch", 2: "past_sequence + sequence"} return common_outputs def UpperCamelCase__ ( self , lowercase_ , lowercase_ = -1 , lowercase_ = -1 , lowercase_ = False , lowercase_ = None , ): """simple docstring""" UpperCAmelCase_ : int = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) # Generate decoder inputs UpperCAmelCase_ : Any = seq_length if not self.use_past else 1 UpperCAmelCase_ : Union[str, Any] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) UpperCAmelCase_ : Optional[Any] = {F"""decoder_{name}""": tensor for name, tensor in decoder_inputs.items()} UpperCAmelCase_ : Optional[int] = dict(**__snake_case , **__snake_case ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch UpperCAmelCase_ , UpperCAmelCase_ : Dict = common_inputs["input_ids"].shape UpperCAmelCase_ : Tuple = common_inputs["decoder_input_ids"].shape[1] UpperCAmelCase_ , UpperCAmelCase_ : List[str] = self.num_attention_heads UpperCAmelCase_ : Union[str, Any] = ( batch, num_encoder_attention_heads, encoder_seq_length, self._config.hidden_size // num_encoder_attention_heads, ) UpperCAmelCase_ : Any = decoder_seq_length + 3 UpperCAmelCase_ : str = ( batch, num_decoder_attention_heads, decoder_past_length, self._config.hidden_size // num_decoder_attention_heads, ) UpperCAmelCase_ : int = torch.cat( [common_inputs["decoder_attention_mask"], torch.ones(__snake_case , __snake_case )] , dim=1 ) UpperCAmelCase_ : str = [] # If the number of encoder and decoder layers are present in the model configuration, both are considered UpperCAmelCase_ , UpperCAmelCase_ : Dict = self.num_layers UpperCAmelCase_ : Union[str, Any] = min(__snake_case , __snake_case ) UpperCAmelCase_ : Optional[int] = max(__snake_case , __snake_case ) - min_num_layers UpperCAmelCase_ : Optional[int] = "encoder" if num_encoder_layers > num_decoder_layers else "decoder" for _ in range(__snake_case ): common_inputs["past_key_values"].append( ( torch.zeros(__snake_case ), torch.zeros(__snake_case ), torch.zeros(__snake_case ), torch.zeros(__snake_case ), ) ) # TODO: test this. UpperCAmelCase_ : Any = encoder_shape if remaining_side_name == "encoder" else decoder_shape for _ in range(__snake_case , __snake_case ): common_inputs["past_key_values"].append((torch.zeros(__snake_case ), torch.zeros(__snake_case )) ) return common_inputs def UpperCamelCase__ ( self , lowercase_ , lowercase_ = -1 , lowercase_ = -1 , lowercase_ = False , lowercase_ = None , ): """simple docstring""" UpperCAmelCase_ : List[str] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch UpperCAmelCase_ , UpperCAmelCase_ : int = common_inputs["input_ids"].shape # Not using the same length for past_key_values UpperCAmelCase_ : Optional[int] = seqlen + 2 UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = self.num_layers UpperCAmelCase_ , UpperCAmelCase_ : List[str] = self.num_attention_heads UpperCAmelCase_ : Dict = ( batch, num_encoder_attention_heads, past_key_values_length, self._config.hidden_size // num_encoder_attention_heads, ) UpperCAmelCase_ : List[Any] = common_inputs["attention_mask"].dtype UpperCAmelCase_ : int = torch.cat( [common_inputs["attention_mask"], torch.ones(__snake_case , __snake_case , dtype=__snake_case )] , dim=1 ) UpperCAmelCase_ : List[str] = [ (torch.zeros(__snake_case ), torch.zeros(__snake_case )) for _ in range(__snake_case ) ] return common_inputs def UpperCamelCase__ ( self , lowercase_ , lowercase_ = -1 , lowercase_ = -1 , lowercase_ = False , lowercase_ = None , ): """simple docstring""" # Copied from OnnxConfig.generate_dummy_inputs # Did not use super(OnnxConfigWithPast, self).generate_dummy_inputs for code clarity. # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX UpperCAmelCase_ : Dict = compute_effective_axis_dimension( __snake_case , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX UpperCAmelCase_ : str = tokenizer.num_special_tokens_to_add(__snake_case ) UpperCAmelCase_ : Optional[int] = compute_effective_axis_dimension( __snake_case , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=__snake_case ) # Generate dummy inputs according to compute batch and sequence UpperCAmelCase_ : List[str] = [" ".join([tokenizer.unk_token] ) * seq_length] * batch_size UpperCAmelCase_ : List[Any] = dict(tokenizer(__snake_case , return_tensors=__snake_case ) ) return common_inputs def UpperCamelCase__ ( self , lowercase_ , lowercase_ = -1 , lowercase_ = -1 , lowercase_ = False , lowercase_ = None , ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase_ : List[Any] = self._generate_dummy_inputs_for_default_and_seqaseq_lm( __snake_case , batch_size=__snake_case , seq_length=__snake_case , is_pair=__snake_case , framework=__snake_case ) elif self.task == "causal-lm": UpperCAmelCase_ : Optional[Any] = self._generate_dummy_inputs_for_causal_lm( __snake_case , batch_size=__snake_case , seq_length=__snake_case , is_pair=__snake_case , framework=__snake_case ) else: UpperCAmelCase_ : List[str] = self._generate_dummy_inputs_for_sequence_classification_and_question_answering( __snake_case , batch_size=__snake_case , seq_length=__snake_case , is_pair=__snake_case , framework=__snake_case ) return common_inputs def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" if self.task in ["default", "seq2seq-lm"]: UpperCAmelCase_ : Optional[int] = super()._flatten_past_key_values_(__snake_case , __snake_case , __snake_case , __snake_case ) else: UpperCAmelCase_ : int = super(__snake_case , self )._flatten_past_key_values_( __snake_case , __snake_case , __snake_case , __snake_case )
353
"""simple docstring""" from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _a = logging.get_logger(__name__) # pylint: disable=invalid-name _a = '\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained("kandinsky-community/kandinsky-2-2-prior")\n >>> pipe_prior.to("cuda")\n >>> prompt = "red cat, 4k photo"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained("kandinsky-community/kandinsky-2-2-decoder")\n >>> pipe.to("cuda")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save("cat.png")\n ```\n' def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=8 ): UpperCAmelCase_ : List[str] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 UpperCAmelCase_ : Tuple = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class A_ (lowercase__ ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , ): """simple docstring""" super().__init__() self.register_modules( unet=lowercase_ , scheduler=lowercase_ , movq=lowercase_ , ) UpperCAmelCase_ : int = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" if latents is None: UpperCAmelCase_ : Dict = randn_tensor(lowercase_ , generator=lowercase_ , device=lowercase_ , dtype=lowercase_ ) else: if latents.shape != shape: raise ValueError(F"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) UpperCAmelCase_ : str = latents.to(lowercase_ ) UpperCAmelCase_ : Dict = latents * scheduler.init_noise_sigma return latents def UpperCamelCase__ ( self , lowercase_=0 ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) UpperCAmelCase_ : Any = torch.device(F"""cuda:{gpu_id}""" ) UpperCAmelCase_ : int = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowercase_ , lowercase_ ) def UpperCamelCase__ ( self , lowercase_=0 ): """simple docstring""" if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) UpperCAmelCase_ : Any = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=lowercase_ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) UpperCAmelCase_ : List[Any] = None for cpu_offloaded_model in [self.unet, self.movq]: UpperCAmelCase_ , UpperCAmelCase_ : str = cpu_offload_with_hook(lowercase_ , lowercase_ , prev_module_hook=lowercase_ ) # We'll offload the last model manually. UpperCAmelCase_ : Tuple = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCamelCase__ ( self ): """simple docstring""" if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowercase_ , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(lowercase_ ) def __call__( self , lowercase_ , lowercase_ , lowercase_ = 512 , lowercase_ = 512 , lowercase_ = 100 , lowercase_ = 4.0 , lowercase_ = 1 , lowercase_ = None , lowercase_ = None , lowercase_ = "pil" , lowercase_ = True , ): """simple docstring""" UpperCAmelCase_ : str = self._execution_device UpperCAmelCase_ : List[Any] = guidance_scale > 1.0 if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = torch.cat(lowercase_ , dim=0 ) UpperCAmelCase_ : Any = image_embeds.shape[0] * num_images_per_prompt if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : List[Any] = torch.cat(lowercase_ , dim=0 ) if do_classifier_free_guidance: UpperCAmelCase_ : Tuple = image_embeds.repeat_interleave(lowercase_ , dim=0 ) UpperCAmelCase_ : List[str] = negative_image_embeds.repeat_interleave(lowercase_ , dim=0 ) UpperCAmelCase_ : Optional[Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=lowercase_ ) self.scheduler.set_timesteps(lowercase_ , device=lowercase_ ) UpperCAmelCase_ : List[Any] = self.scheduler.timesteps UpperCAmelCase_ : List[str] = self.unet.config.in_channels UpperCAmelCase_ , UpperCAmelCase_ : Dict = downscale_height_and_width(lowercase_ , lowercase_ , self.movq_scale_factor ) # create initial latent UpperCAmelCase_ : int = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , lowercase_ , lowercase_ , lowercase_ , self.scheduler , ) for i, t in enumerate(self.progress_bar(lowercase_ ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase_ : Tuple = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase_ : Union[str, Any] = {"image_embeds": image_embeds} UpperCAmelCase_ : Optional[Any] = self.unet( sample=lowercase_ , timestep=lowercase_ , encoder_hidden_states=lowercase_ , added_cond_kwargs=lowercase_ , return_dict=lowercase_ , )[0] if do_classifier_free_guidance: UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = noise_pred.split(latents.shape[1] , dim=1 ) UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = noise_pred.chunk(2 ) UpperCAmelCase_ , UpperCAmelCase_ : Tuple = variance_pred.chunk(2 ) UpperCAmelCase_ : int = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) UpperCAmelCase_ : str = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): UpperCAmelCase_ , UpperCAmelCase_ : Dict = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase_ : List[str] = self.scheduler.step( lowercase_ , lowercase_ , lowercase_ , generator=lowercase_ , )[0] # post-processing UpperCAmelCase_ : Tuple = self.movq.decode(lowercase_ , force_not_quantize=lowercase_ )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: UpperCAmelCase_ : List[Any] = image * 0.5 + 0.5 UpperCAmelCase_ : int = image.clamp(0 , 1 ) UpperCAmelCase_ : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": UpperCAmelCase_ : Dict = self.numpy_to_pil(lowercase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowercase_ )
23
0
"""simple docstring""" import logging import os import quant_trainer import torch from torch.utils.data import DataLoader from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput _a = logging.getLogger(__name__) if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class A_ (a__ ): '''simple docstring''' def __init__( self , *lowercase_ , lowercase_=None , lowercase_=None , lowercase_=None , **lowercase_ ): """simple docstring""" super().__init__(*_lowerCamelCase , **_lowerCamelCase ) UpperCAmelCase_ : Optional[int] = eval_examples UpperCAmelCase_ : Any = post_process_function UpperCAmelCase_ : int = quant_trainer_args UpperCAmelCase_ : str = 128 # default number of calibration samples def UpperCamelCase__ ( self , lowercase_=None ): """simple docstring""" if calib_dataset is None and self.calib_dataset is None: raise ValueError("Trainer: calibration requires an calib_dataset." ) UpperCAmelCase_ : List[str] = calib_dataset if calib_dataset is not None else self.calib_dataset UpperCAmelCase_ : Optional[int] = self._remove_unused_columns(_lowerCamelCase , description="Calibration" ) return DataLoader( _lowerCamelCase , batch_size=self.args.eval_batch_size , collate_fn=self.data_collator , drop_last=self.args.dataloader_drop_last , num_workers=self.args.dataloader_num_workers , pin_memory=self.args.dataloader_pin_memory , shuffle=_lowerCamelCase , ) def UpperCamelCase__ ( self , lowercase_=None ): """simple docstring""" UpperCAmelCase_ : Dict = self.train_dataset if calib_dataset is None else calib_dataset UpperCAmelCase_ : Optional[Any] = self.get_calib_dataloader(_lowerCamelCase ) UpperCAmelCase_ : str = self.model quant_trainer.configure_model(_lowerCamelCase , self.quant_trainer_args , calib=_lowerCamelCase ) model.eval() quant_trainer.enable_calibration(_lowerCamelCase ) logger.info("***** Running calibration *****" ) logger.info(F""" Num examples = {self.calib_num}""" ) logger.info(F""" Batch size = {calib_dataloader.batch_size}""" ) for step, inputs in enumerate(_lowerCamelCase ): # Prediction step UpperCAmelCase_ : Union[str, Any] = self.prediction_step(_lowerCamelCase , _lowerCamelCase , prediction_loss_only=_lowerCamelCase ) if (step + 1) * calib_dataloader.batch_size >= self.calib_num: break quant_trainer.finish_calibration(_lowerCamelCase , self.quant_trainer_args ) UpperCAmelCase_ : int = model def UpperCamelCase__ ( self , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_ = "eval" ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.eval_dataset if eval_dataset is None else eval_dataset UpperCAmelCase_ : Optional[int] = self.get_eval_dataloader(_lowerCamelCase ) UpperCAmelCase_ : List[Any] = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. UpperCAmelCase_ : Tuple = self.compute_metrics UpperCAmelCase_ : Optional[Any] = None UpperCAmelCase_ : str = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: UpperCAmelCase_ : List[Any] = eval_loop( _lowerCamelCase , description="Evaluation" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_lowerCamelCase , ) finally: UpperCAmelCase_ : List[str] = compute_metrics if self.post_process_function is not None and self.compute_metrics is not None: UpperCAmelCase_ : Dict = self.post_process_function(_lowerCamelCase , _lowerCamelCase , output.predictions ) UpperCAmelCase_ : Tuple = self.compute_metrics(_lowerCamelCase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"""{metric_key_prefix}_""" ): UpperCAmelCase_ : str = metrics.pop(_lowerCamelCase ) self.log(_lowerCamelCase ) else: UpperCAmelCase_ : Tuple = {} if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) UpperCAmelCase_ : Optional[Any] = self.callback_handler.on_evaluate(self.args , self.state , self.control , _lowerCamelCase ) return metrics def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_=None , lowercase_ = "test" ): """simple docstring""" UpperCAmelCase_ : int = self.get_test_dataloader(_lowerCamelCase ) # Temporarily disable metric computation, we will do it in the loop here. UpperCAmelCase_ : Optional[int] = self.compute_metrics UpperCAmelCase_ : int = None UpperCAmelCase_ : Union[str, Any] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: UpperCAmelCase_ : Any = eval_loop( _lowerCamelCase , description="Prediction" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=_lowerCamelCase , ) finally: UpperCAmelCase_ : Optional[Any] = compute_metrics if self.post_process_function is None or self.compute_metrics is None: return output UpperCAmelCase_ : List[Any] = self.post_process_function(_lowerCamelCase , _lowerCamelCase , output.predictions , "predict" ) UpperCAmelCase_ : List[str] = self.compute_metrics(_lowerCamelCase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"""{metric_key_prefix}_""" ): UpperCAmelCase_ : Any = metrics.pop(_lowerCamelCase ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=_lowerCamelCase ) def UpperCamelCase__ ( self , lowercase_="./" ): """simple docstring""" UpperCAmelCase_ : Optional[int] = self.eval_dataset UpperCAmelCase_ : Any = self.get_eval_dataloader(_lowerCamelCase ) UpperCAmelCase_ : Any = next(iter(_lowerCamelCase ) ) # saving device - to make it consistent UpperCAmelCase_ : Optional[int] = torch.device("cuda" if torch.cuda.is_available() else "cpu" ) # convert to tuple UpperCAmelCase_ : List[Any] = tuple(v.to(_lowerCamelCase ) for k, v in batch.items() ) logger.info("Converting model to be onnx compatible" ) from pytorch_quantization.nn import TensorQuantizer UpperCAmelCase_ : int = True UpperCAmelCase_ : List[Any] = self.model.to(_lowerCamelCase ) model.eval() model.float() UpperCAmelCase_ : Dict = model.module if hasattr(_lowerCamelCase , "module" ) else model quant_trainer.configure_model(_lowerCamelCase , self.quant_trainer_args ) UpperCAmelCase_ : Any = os.path.join(_lowerCamelCase , "model.onnx" ) logger.info(F"""exporting model to {output_model_file}""" ) UpperCAmelCase_ : str = {0: '''batch_size''', 1: '''seq_len'''} torch.onnx.export( _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , export_params=_lowerCamelCase , opset_version=13 , do_constant_folding=_lowerCamelCase , input_names=["input_ids", "attention_mask", "token_type_ids"] , output_names=["output_start_logits", "output_end_logits"] , dynamic_axes={ "input_ids": axes, "attention_mask": axes, "token_type_ids": axes, "output_start_logits": axes, "output_end_logits": axes, } , verbose=_lowerCamelCase , ) logger.info("onnx export finished" )
354
"""simple docstring""" import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _a = logging.get_logger(__name__) _a = { 'facebook/detr-resnet-50': 'https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json', # See all DETR models at https://huggingface.co/models?filter=detr } class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """detr""" SCREAMING_SNAKE_CASE__ : str = ["""past_key_values"""] SCREAMING_SNAKE_CASE__ : Union[str, Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , lowercase_=True , lowercase_=None , lowercase_=3 , lowercase_=100 , lowercase_=6 , lowercase_=2048 , lowercase_=8 , lowercase_=6 , lowercase_=2048 , lowercase_=8 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=True , lowercase_="relu" , lowercase_=256 , lowercase_=0.1 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=0.02 , lowercase_=1.0 , lowercase_=False , lowercase_="sine" , lowercase_="resnet50" , lowercase_=True , lowercase_=False , lowercase_=1 , lowercase_=5 , lowercase_=2 , lowercase_=1 , lowercase_=1 , lowercase_=5 , lowercase_=2 , lowercase_=0.1 , **lowercase_ , ): """simple docstring""" if backbone_config is not None and use_timm_backbone: raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`." ) if not use_timm_backbone: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) UpperCAmelCase_ : Union[str, Any] = CONFIG_MAPPING["resnet"](out_features=["stage4"] ) elif isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = backbone_config.get("model_type" ) UpperCAmelCase_ : int = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase_ : Any = config_class.from_dict(lowercase_ ) # set timm attributes to None UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = None, None, None UpperCAmelCase_ : int = use_timm_backbone UpperCAmelCase_ : int = backbone_config UpperCAmelCase_ : List[Any] = num_channels UpperCAmelCase_ : int = num_queries UpperCAmelCase_ : Union[str, Any] = d_model UpperCAmelCase_ : str = encoder_ffn_dim UpperCAmelCase_ : Tuple = encoder_layers UpperCAmelCase_ : List[Any] = encoder_attention_heads UpperCAmelCase_ : Union[str, Any] = decoder_ffn_dim UpperCAmelCase_ : Optional[Any] = decoder_layers UpperCAmelCase_ : Union[str, Any] = decoder_attention_heads UpperCAmelCase_ : Optional[int] = dropout UpperCAmelCase_ : List[str] = attention_dropout UpperCAmelCase_ : Any = activation_dropout UpperCAmelCase_ : str = activation_function UpperCAmelCase_ : Tuple = init_std UpperCAmelCase_ : Optional[Any] = init_xavier_std UpperCAmelCase_ : Optional[Any] = encoder_layerdrop UpperCAmelCase_ : Optional[int] = decoder_layerdrop UpperCAmelCase_ : Tuple = encoder_layers UpperCAmelCase_ : int = auxiliary_loss UpperCAmelCase_ : Optional[Any] = position_embedding_type UpperCAmelCase_ : Tuple = backbone UpperCAmelCase_ : Optional[int] = use_pretrained_backbone UpperCAmelCase_ : Dict = dilation # Hungarian matcher UpperCAmelCase_ : Union[str, Any] = class_cost UpperCAmelCase_ : Any = bbox_cost UpperCAmelCase_ : int = giou_cost # Loss coefficients UpperCAmelCase_ : str = mask_loss_coefficient UpperCAmelCase_ : Any = dice_loss_coefficient UpperCAmelCase_ : Optional[Any] = bbox_loss_coefficient UpperCAmelCase_ : List[str] = giou_loss_coefficient UpperCAmelCase_ : List[Any] = eos_coefficient super().__init__(is_encoder_decoder=lowercase_ , **lowercase_ ) @property def UpperCamelCase__ ( self ): """simple docstring""" return self.encoder_attention_heads @property def UpperCamelCase__ ( self ): """simple docstring""" return self.d_model @classmethod def UpperCamelCase__ ( cls , lowercase_ , **lowercase_ ): """simple docstring""" return cls(backbone_config=lowercase_ , **lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: UpperCAmelCase_ : Union[str, Any] = self.backbone_config.to_dict() UpperCAmelCase_ : str = self.__class__.model_type return output class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = version.parse("""1.11""" ) @property def UpperCamelCase__ ( self ): """simple docstring""" return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("pixel_mask", {0: "batch"}), ] ) @property def UpperCamelCase__ ( self ): """simple docstring""" return 1E-5 @property def UpperCamelCase__ ( self ): """simple docstring""" return 12
23
0
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a = logging.get_logger(__name__) _a = '▁' _a = {'vocab_file': 'spiece.model'} _a = { 'vocab_file': { 'google/reformer-crime-and-punishment': ( 'https://huggingface.co/google/reformer-crime-and-punishment/resolve/main/spiece.model' ) } } _a = { 'google/reformer-crime-and-punishment': 524_288, } class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : str = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : int = ['''input_ids''', '''attention_mask'''] def __init__( self , lowercase_ , lowercase_="</s>" , lowercase_="<unk>" , lowercase_=[] , lowercase_ = None , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : List[str] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( eos_token=_a , unk_token=_a , additional_special_tokens=_a , sp_model_kwargs=self.sp_model_kwargs , **_a , ) UpperCAmelCase_ : Dict = vocab_file UpperCAmelCase_ : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(_a ) @property def UpperCamelCase__ ( self ): """simple docstring""" return self.sp_model.get_piece_size() def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = self.__dict__.copy() UpperCAmelCase_ : Dict = None return state def __setstate__( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Any = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): UpperCAmelCase_ : List[Any] = {} UpperCAmelCase_ : int = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.sp_model.encode(_a , out_type=_a ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.sp_model.piece_to_id(_a ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" if index < self.sp_model.get_piece_size(): UpperCAmelCase_ : int = self.sp_model.IdToPiece(_a ) return token def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Tuple = [] UpperCAmelCase_ : Dict = '' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(_a ) + token UpperCAmelCase_ : str = [] else: current_sub_tokens.append(_a ) out_string += self.sp_model.decode(_a ) return out_string.strip() def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if not os.path.isdir(_a ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase_ : Dict = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _a ) elif not os.path.isfile(self.vocab_file ): with open(_a , "wb" ) as fi: UpperCAmelCase_ : str = self.sp_model.serialized_model_proto() fi.write(_a ) return (out_vocab_file,)
355
"""simple docstring""" _a = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(100_000)] def __a ( __lowerCamelCase ): UpperCAmelCase_ : Optional[int] = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 10_0000] number //= 10_0000 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution _a = [None] * 10_000_000 _a = True _a = False def __a ( __lowerCamelCase ): if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore UpperCAmelCase_ : Dict = chain(next_number(__lowerCamelCase ) ) UpperCAmelCase_ : List[str] = number_chain while number < 1000_0000: UpperCAmelCase_ : List[Any] = number_chain number *= 10 return number_chain def __a ( __lowerCamelCase = 1000_0000 ): for i in range(1, __lowerCamelCase ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(__lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() print(f"""{solution() = }""")
23
0
"""simple docstring""" import argparse from pathlib import Path import fairseq import torch from fairseq.models.xmod import XMODModel as FairseqXmodModel from packaging import version from transformers import XmodConfig, XmodForMaskedLM, XmodForSequenceClassification from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse('0.12.2'): raise Exception('requires fairseq >= 0.12.2') if version.parse(fairseq.__version__) > version.parse('2'): raise Exception('requires fairseq < v2') logging.set_verbosity_info() _a = logging.get_logger(__name__) _a = """Hello, World!""" _a = """en_XX""" def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Union[str, Any] = Path("data_bin" ) UpperCAmelCase_ : int = FairseqXmodModel.from_pretrained( model_name_or_path=str(Path(__snake_case ).parent ), checkpoint_file=Path(__snake_case ).name, _name="xmod_base", arch="xmod_base", task="multilingual_masked_lm", data_name_or_path=str(__snake_case ), bpe="sentencepiece", sentencepiece_model=str(Path(__snake_case ).parent / "sentencepiece.bpe.model" ), src_dict=str(data_dir / "dict.txt" ), ) xmod.eval() # disable dropout print(__snake_case ) UpperCAmelCase_ : Any = xmod.model.encoder.sentence_encoder UpperCAmelCase_ : Optional[Any] = XmodConfig( vocab_size=xmod_sent_encoder.embed_tokens.num_embeddings, hidden_size=xmod.cfg.model.encoder_embed_dim, num_hidden_layers=xmod.cfg.model.encoder_layers, num_attention_heads=xmod.cfg.model.encoder_attention_heads, intermediate_size=xmod.cfg.model.encoder_ffn_embed_dim, max_position_embeddings=514, type_vocab_size=1, layer_norm_eps=1E-5, pre_norm=xmod.cfg.model.encoder_normalize_before, adapter_reduction_factor=getattr(xmod.cfg.model, "bottleneck", 2 ), adapter_layer_norm=xmod.cfg.model.adapter_layer_norm, adapter_reuse_layer_norm=xmod.cfg.model.adapter_reuse_layer_norm, ln_before_adapter=xmod.cfg.model.ln_before_adapter, languages=xmod.cfg.model.languages, ) if classification_head: UpperCAmelCase_ : int = xmod.model.classification_heads["mnli"].out_proj.weight.shape[0] print("Our X-MOD config:", __snake_case ) UpperCAmelCase_ : int = XmodForSequenceClassification(__snake_case ) if classification_head else XmodForMaskedLM(__snake_case ) model.eval() # Now let's copy all the weights. # Embeddings UpperCAmelCase_ : List[Any] = xmod_sent_encoder.embed_tokens.weight UpperCAmelCase_ : Optional[Any] = xmod_sent_encoder.embed_positions.weight UpperCAmelCase_ : str = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c xmod doesn't use them. UpperCAmelCase_ : Any = xmod_sent_encoder.layernorm_embedding.weight UpperCAmelCase_ : Any = xmod_sent_encoder.layernorm_embedding.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer UpperCAmelCase_ : Any = model.roberta.encoder.layer[i] UpperCAmelCase_ : List[str] = xmod_sent_encoder.layers[i] # self attention UpperCAmelCase_ : List[Any] = layer.attention.self if not ( xmod_layer.self_attn.k_proj.weight.data.shape == xmod_layer.self_attn.q_proj.weight.data.shape == xmod_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ): raise AssertionError("Dimensions of self-attention weights do not match." ) UpperCAmelCase_ : List[str] = xmod_layer.self_attn.q_proj.weight UpperCAmelCase_ : Union[str, Any] = xmod_layer.self_attn.q_proj.bias UpperCAmelCase_ : int = xmod_layer.self_attn.k_proj.weight UpperCAmelCase_ : int = xmod_layer.self_attn.k_proj.bias UpperCAmelCase_ : List[Any] = xmod_layer.self_attn.v_proj.weight UpperCAmelCase_ : Tuple = xmod_layer.self_attn.v_proj.bias # self-attention output UpperCAmelCase_ : Union[str, Any] = layer.attention.output if self_output.dense.weight.shape != xmod_layer.self_attn.out_proj.weight.shape: raise AssertionError("Dimensions of self-attention output weights do not match." ) UpperCAmelCase_ : Union[str, Any] = xmod_layer.self_attn.out_proj.weight UpperCAmelCase_ : List[Any] = xmod_layer.self_attn.out_proj.bias UpperCAmelCase_ : Optional[Any] = xmod_layer.self_attn_layer_norm.weight UpperCAmelCase_ : Any = xmod_layer.self_attn_layer_norm.bias # intermediate UpperCAmelCase_ : Dict = layer.intermediate if intermediate.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError("Dimensions of intermediate weights do not match." ) UpperCAmelCase_ : Tuple = xmod_layer.fca.weight UpperCAmelCase_ : Optional[int] = xmod_layer.fca.bias # output UpperCAmelCase_ : Union[str, Any] = layer.output if bert_output.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError("Dimensions of feed-forward weights do not match." ) UpperCAmelCase_ : Dict = xmod_layer.fca.weight UpperCAmelCase_ : str = xmod_layer.fca.bias UpperCAmelCase_ : Dict = xmod_layer.final_layer_norm.weight UpperCAmelCase_ : Optional[int] = xmod_layer.final_layer_norm.bias if bert_output.adapter_layer_norm is not None: UpperCAmelCase_ : Optional[Any] = xmod_layer.adapter_layer_norm.weight UpperCAmelCase_ : Tuple = xmod_layer.adapter_layer_norm.bias if sorted(bert_output.adapter_modules.keys() ) != sorted(xmod_layer.adapter_modules.keys() ): raise AssertionError("Lists of language adapters do not match." ) for lang_code, adapter in xmod_layer.adapter_modules.items(): UpperCAmelCase_ : Dict = bert_output.adapter_modules[lang_code] UpperCAmelCase_ : List[str] = xmod_layer.adapter_modules[lang_code] UpperCAmelCase_ : int = from_adapter.fca.weight UpperCAmelCase_ : Dict = from_adapter.fca.bias UpperCAmelCase_ : Any = from_adapter.fca.weight UpperCAmelCase_ : List[Any] = from_adapter.fca.bias # end of layer if xmod_sent_encoder.layer_norm is not None: UpperCAmelCase_ : Union[str, Any] = xmod_sent_encoder.layer_norm.weight UpperCAmelCase_ : Any = xmod_sent_encoder.layer_norm.bias if classification_head: UpperCAmelCase_ : int = xmod.model.classification_heads["mnli"].dense.weight UpperCAmelCase_ : Tuple = xmod.model.classification_heads["mnli"].dense.bias UpperCAmelCase_ : Union[str, Any] = xmod.model.classification_heads["mnli"].out_proj.weight UpperCAmelCase_ : int = xmod.model.classification_heads["mnli"].out_proj.bias else: # LM Head UpperCAmelCase_ : List[str] = xmod.model.encoder.lm_head.dense.weight UpperCAmelCase_ : Union[str, Any] = xmod.model.encoder.lm_head.dense.bias UpperCAmelCase_ : Optional[Any] = xmod.model.encoder.lm_head.layer_norm.weight UpperCAmelCase_ : int = xmod.model.encoder.lm_head.layer_norm.bias UpperCAmelCase_ : Dict = xmod.model.encoder.lm_head.weight UpperCAmelCase_ : List[Any] = xmod.model.encoder.lm_head.bias # Let's check that we get the same results. UpperCAmelCase_ : str = xmod.encode(__snake_case ).unsqueeze(0 ) # batch of size 1 model.roberta.set_default_language(__snake_case ) UpperCAmelCase_ : str = model(__snake_case )[0] if classification_head: UpperCAmelCase_ : str = xmod.model.classification_heads["mnli"](xmod.extract_features(__snake_case ) ) else: UpperCAmelCase_ : Any = xmod.model(__snake_case, lang_id=[SAMPLE_LANGUAGE] )[0] print(our_output.shape, their_output.shape ) UpperCAmelCase_ : Optional[int] = torch.max(torch.abs(our_output - their_output ) ).item() print(f"""max_absolute_diff = {max_absolute_diff}""" ) # ~ 1e-7 UpperCAmelCase_ : Optional[int] = torch.allclose(__snake_case, __snake_case, atol=1E-3 ) print("Do both models output the same tensors?", "🔥" if success else "💩" ) if not success: raise Exception("Something went wRoNg" ) Path(__snake_case ).mkdir(parents=__snake_case, exist_ok=__snake_case ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(__snake_case ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--xmod_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--classification_head', action='store_true', help='Whether to convert a final classification head.' ) _a = parser.parse_args() convert_xmod_checkpoint_to_pytorch( args.xmod_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
356
"""simple docstring""" def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # Return True if there is node that has not iterated. UpperCAmelCase_ : List[Any] = [False] * len(__lowerCamelCase ) UpperCAmelCase_ : Any = [] queue.append(__lowerCamelCase ) UpperCAmelCase_ : Tuple = True while queue: UpperCAmelCase_ : str = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(__lowerCamelCase ) UpperCAmelCase_ : Any = True UpperCAmelCase_ : Union[str, Any] = u return visited[t] def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # This array is filled by BFS and to store path UpperCAmelCase_ : List[str] = [-1] * (len(__lowerCamelCase )) UpperCAmelCase_ : Any = 0 while bfs(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : int = float("Inf" ) UpperCAmelCase_ : Tuple = sink while s != source: # Find the minimum value in select path UpperCAmelCase_ : Tuple = min(__lowerCamelCase, graph[parent[s]][s] ) UpperCAmelCase_ : Dict = parent[s] max_flow += path_flow UpperCAmelCase_ : Optional[Any] = sink while v != source: UpperCAmelCase_ : List[str] = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow UpperCAmelCase_ : Optional[int] = parent[v] return max_flow _a = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] _a , _a = 0, 5 print(ford_fulkerson(graph, source, sink))
23
0
"""simple docstring""" def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): global f # a global dp table for knapsack if f[i][j] < 0: if j < wt[i - 1]: UpperCAmelCase_ : List[str] = mf_knapsack(i - 1, __A, __A, __A ) else: UpperCAmelCase_ : Union[str, Any] = max( mf_knapsack(i - 1, __A, __A, __A ), mf_knapsack(i - 1, __A, __A, j - wt[i - 1] ) + val[i - 1], ) UpperCAmelCase_ : Union[str, Any] = val return f[i][j] def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : str = [[0] * (w + 1) for _ in range(n + 1 )] for i in range(1, n + 1 ): for w_ in range(1, w + 1 ): if wt[i - 1] <= w_: UpperCAmelCase_ : Optional[Any] = max(val[i - 1] + dp[i - 1][w_ - wt[i - 1]], dp[i - 1][w_] ) else: UpperCAmelCase_ : Any = dp[i - 1][w_] return dp[n][w_], dp def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): if not (isinstance(__A, (list, tuple) ) and isinstance(__A, (list, tuple) )): raise ValueError( "Both the weights and values vectors must be either lists or tuples" ) UpperCAmelCase_ : List[str] = len(__A ) if num_items != len(__A ): UpperCAmelCase_ : Optional[int] = ( "The number of weights must be the same as the number of values.\n" f"""But got {num_items} weights and {len(__A )} values""" ) raise ValueError(__A ) for i in range(__A ): if not isinstance(wt[i], __A ): UpperCAmelCase_ : Dict = ( "All weights must be integers but got weight of " f"""type {type(wt[i] )} at index {i}""" ) raise TypeError(__A ) UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = knapsack(__A, __A, __A, __A ) UpperCAmelCase_ : Union[str, Any] = set() _construct_solution(__A, __A, __A, __A, __A ) return optimal_val, example_optional_set def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): if i > 0 and j > 0: if dp[i - 1][j] == dp[i][j]: _construct_solution(__A, __A, i - 1, __A, __A ) else: optimal_set.add(__A ) _construct_solution(__A, __A, i - 1, j - wt[i - 1], __A ) if __name__ == "__main__": _a = [3, 2, 4, 4] _a = [4, 3, 2, 3] _a = 4 _a = 6 _a = [[0] * (w + 1)] + [[0] + [-1] * (w + 1) for _ in range(n + 1)] _a = knapsack(w, wt, val, n) print(optimal_solution) print(mf_knapsack(n, wt, val, w)) # switched the n and w # testing the dynamic programming problem with example # the optimal subset for the above example are items 3 and 4 _a = knapsack_with_example_solution(w, wt, val) assert optimal_solution == 8 assert optimal_subset == {3, 4} print('optimal_value = ', optimal_solution) print('An optimal subset corresponding to the optimal value', optimal_subset)
357
"""simple docstring""" import datasets _a = '\\n@InProceedings{conneau2018xnli,\n author = "Conneau, Alexis\n and Rinott, Ruty\n and Lample, Guillaume\n and Williams, Adina\n and Bowman, Samuel R.\n and Schwenk, Holger\n and Stoyanov, Veselin",\n title = "XNLI: Evaluating Cross-lingual Sentence Representations",\n booktitle = "Proceedings of the 2018 Conference on Empirical Methods\n in Natural Language Processing",\n year = "2018",\n publisher = "Association for Computational Linguistics",\n location = "Brussels, Belgium",\n}\n' _a = '\\nXNLI is a subset of a few thousand examples from MNLI which has been translated\ninto a 14 different languages (some low-ish resource). As with MNLI, the goal is\nto predict textual entailment (does sentence A imply/contradict/neither sentence\nB) and is a classification task (given two sentences, predict one of three\nlabels).\n' _a = '\nComputes XNLI score which is just simple accuracy.\nArgs:\n predictions: Predicted labels.\n references: Ground truth labels.\nReturns:\n \'accuracy\': accuracy\nExamples:\n\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> xnli_metric = datasets.load_metric("xnli")\n >>> results = xnli_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n' def __a ( __lowerCamelCase, __lowerCamelCase ): return (preds == labels).mean() @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64" if self.config_name != "sts-b" else "float32" ), "references": datasets.Value("int64" if self.config_name != "sts-b" else "float32" ), } ) , codebase_urls=[] , reference_urls=[] , format="numpy" , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" return {"accuracy": simple_accuracy(lowercase_ , lowercase_ )}
23
0
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING _a = logging.get_logger(__name__) _a = Dict[str, Any] _a = List[Prediction] @add_end_docstrings(lowercase__ ) class A_ (lowercase__ ): '''simple docstring''' def __init__( self , *lowercase_ , **lowercase_ ): """simple docstring""" super().__init__(*_UpperCamelCase , **_UpperCamelCase ) if self.framework == "tf": raise ValueError(F"""The {self.__class__} is only available in PyTorch.""" ) requires_backends(self , "vision" ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = {} if "threshold" in kwargs: UpperCAmelCase_ : Optional[int] = kwargs["""threshold"""] return {}, {}, postprocess_kwargs def __call__( self , *lowercase_ , **lowercase_ ): """simple docstring""" return super().__call__(*_UpperCamelCase , **_UpperCamelCase ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = load_image(_UpperCamelCase ) UpperCAmelCase_ : Optional[int] = torch.IntTensor([[image.height, image.width]] ) UpperCAmelCase_ : int = self.image_processor(images=[image] , return_tensors="pt" ) if self.tokenizer is not None: UpperCAmelCase_ : List[Any] = self.tokenizer(text=inputs["words"] , boxes=inputs["boxes"] , return_tensors="pt" ) UpperCAmelCase_ : int = target_size return inputs def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = model_inputs.pop("target_size" ) UpperCAmelCase_ : Dict = self.model(**_UpperCamelCase ) UpperCAmelCase_ : Dict = outputs.__class__({"target_size": target_size, **outputs} ) if self.tokenizer is not None: UpperCAmelCase_ : Optional[int] = model_inputs["""bbox"""] return model_outputs def UpperCamelCase__ ( self , lowercase_ , lowercase_=0.9 ): """simple docstring""" UpperCAmelCase_ : Any = model_outputs["""target_size"""] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. UpperCAmelCase_ : Any = target_size[0].tolist() def unnormalize(lowercase_ ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1000), (height * bbox[1] / 1000), (width * bbox[2] / 1000), (height * bbox[3] / 1000), ] ) ) UpperCAmelCase_ : str = model_outputs["""logits"""].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) UpperCAmelCase_ : Union[str, Any] = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] UpperCAmelCase_ : int = [unnormalize(_UpperCamelCase ) for bbox in model_outputs["""bbox"""].squeeze(0 )] UpperCAmelCase_ : List[Any] = ["""score""", """label""", """box"""] UpperCAmelCase_ : Optional[int] = [dict(zip(_UpperCamelCase , _UpperCamelCase ) ) for vals in zip(scores.tolist() , _UpperCamelCase , _UpperCamelCase ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel UpperCAmelCase_ : int = self.image_processor.post_process_object_detection(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) UpperCAmelCase_ : Union[str, Any] = raw_annotations[0] UpperCAmelCase_ : List[Any] = raw_annotation["""scores"""] UpperCAmelCase_ : Any = raw_annotation["""labels"""] UpperCAmelCase_ : Any = raw_annotation["""boxes"""] UpperCAmelCase_ : Any = scores.tolist() UpperCAmelCase_ : str = [self.model.config.idalabel[label.item()] for label in labels] UpperCAmelCase_ : int = [self._get_bounding_box(_UpperCamelCase ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] UpperCAmelCase_ : Any = ["""score""", """label""", """box"""] UpperCAmelCase_ : str = [ dict(zip(_UpperCamelCase , _UpperCamelCase ) ) for vals in zip(raw_annotation["scores"] , raw_annotation["labels"] , raw_annotation["boxes"] ) ] return annotation def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" if self.framework != "pt": raise ValueError("The ObjectDetectionPipeline is only available in PyTorch." ) UpperCAmelCase_ : Dict = box.int().tolist() UpperCAmelCase_ : Union[str, Any] = { """xmin""": xmin, """ymin""": ymin, """xmax""": xmax, """ymax""": ymax, } return bbox
358
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy _a = logging.get_logger(__name__) class A_ (lowercase__ ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = feature_size UpperCAmelCase_ : Any = sampling_rate UpperCAmelCase_ : Any = padding_value UpperCAmelCase_ : str = kwargs.pop("padding_side" , "right" ) UpperCAmelCase_ : List[str] = kwargs.pop("return_attention_mask" , lowercase_ ) super().__init__(**lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = True , lowercase_ = None , lowercase_ = False , lowercase_ = None , lowercase_ = None , lowercase_ = None , ): """simple docstring""" # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(lowercase_ , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): UpperCAmelCase_ : Dict = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( "You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`" F""" to this method that includes {self.model_input_names[0]}, but you provided""" F""" {list(processed_features.keys() )}""" ) UpperCAmelCase_ : Tuple = processed_features[self.model_input_names[0]] UpperCAmelCase_ : List[str] = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(lowercase_ ) == 0: if return_attention_mask: UpperCAmelCase_ : Union[str, Any] = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch UpperCAmelCase_ : List[str] = required_input[0] if isinstance(lowercase_ , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. UpperCAmelCase_ : Any = 0 while len(required_input[index] ) == 0: index += 1 if index < len(lowercase_ ): UpperCAmelCase_ : Optional[Any] = required_input[index][0] if return_tensors is None: if is_tf_tensor(lowercase_ ): UpperCAmelCase_ : Dict = "tf" elif is_torch_tensor(lowercase_ ): UpperCAmelCase_ : Any = "pt" elif isinstance(lowercase_ , (int, float, list, tuple, np.ndarray) ): UpperCAmelCase_ : str = "np" else: raise ValueError( F"""type of {first_element} unknown: {type(lowercase_ )}. """ "Should be one of a python, numpy, pytorch or tensorflow object." ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): UpperCAmelCase_ : Optional[int] = to_numpy(lowercase_ ) else: UpperCAmelCase_ : List[str] = [to_numpy(lowercase_ ) for v in value] # Convert padding_strategy in PaddingStrategy UpperCAmelCase_ : Dict = self._get_padding_strategies(padding=lowercase_ , max_length=lowercase_ ) UpperCAmelCase_ : str = processed_features[self.model_input_names[0]] UpperCAmelCase_ : int = len(lowercase_ ) if not all(len(lowercase_ ) == batch_size for v in processed_features.values() ): raise ValueError("Some items in the output dictionary have a different batch size than others." ) UpperCAmelCase_ : int = [] for i in range(lowercase_ ): UpperCAmelCase_ : str = {k: v[i] for k, v in processed_features.items()} # truncation UpperCAmelCase_ : List[str] = self._truncate( lowercase_ , max_length=lowercase_ , pad_to_multiple_of=lowercase_ , truncation=lowercase_ , ) truncated_inputs.append(lowercase_ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length UpperCAmelCase_ : str = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) UpperCAmelCase_ : Dict = PaddingStrategy.MAX_LENGTH UpperCAmelCase_ : List[str] = {} for i in range(lowercase_ ): # padding UpperCAmelCase_ : int = self._pad( truncated_inputs[i] , max_length=lowercase_ , padding_strategy=lowercase_ , pad_to_multiple_of=lowercase_ , return_attention_mask=lowercase_ , ) for key, value in outputs.items(): if key not in batch_outputs: UpperCAmelCase_ : Any = [] if value.dtype is np.dtype(np.floataa ): UpperCAmelCase_ : List[Any] = value.astype(np.floataa ) batch_outputs[key].append(lowercase_ ) return BatchFeature(lowercase_ , tensor_type=lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = PaddingStrategy.DO_NOT_PAD , lowercase_ = None , lowercase_ = None , ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: UpperCAmelCase_ : Tuple = len(lowercase_ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCAmelCase_ : Tuple = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCAmelCase_ : Dict = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(lowercase_ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: UpperCAmelCase_ : Optional[int] = np.ones(len(lowercase_ ) , dtype=np.intaa ) if needs_to_be_padded: UpperCAmelCase_ : Dict = max_length - len(lowercase_ ) if self.padding_side == "right": if return_attention_mask: UpperCAmelCase_ : List[Any] = np.pad( processed_features["attention_mask"] , (0, difference) ) UpperCAmelCase_ : Dict = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) UpperCAmelCase_ : Optional[Any] = np.pad( lowercase_ , lowercase_ , "constant" , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: UpperCAmelCase_ : Optional[Any] = np.pad( processed_features["attention_mask"] , (difference, 0) ) UpperCAmelCase_ : Dict = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) UpperCAmelCase_ : str = np.pad( lowercase_ , lowercase_ , "constant" , constant_values=self.padding_value ) else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return processed_features def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , ): """simple docstring""" if not truncation: return processed_features elif truncation and max_length is None: raise ValueError("When setting ``truncation=True``, make sure that ``max_length`` is defined." ) UpperCAmelCase_ : Optional[int] = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCAmelCase_ : Union[str, Any] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCAmelCase_ : Optional[Any] = len(lowercase_ ) > max_length if needs_to_be_truncated: UpperCAmelCase_ : int = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: UpperCAmelCase_ : Dict = processed_features["attention_mask"][:max_length] return processed_features def UpperCamelCase__ ( self , lowercase_=False , lowercase_=None ): """simple docstring""" # Get padding strategy if padding is not False: if padding is True: UpperCAmelCase_ : Dict = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : Optional[Any] = PaddingStrategy(lowercase_ ) elif isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = padding else: UpperCAmelCase_ : str = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F"""When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined""" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( "Asking to pad but the feature_extractor does not have a padding value. Please select a value to use" " as `padding_value`. For example: `feature_extractor.padding_value = 0.0`." ) return padding_strategy
23
0
import importlib.metadata import operator import re import sys from typing import Optional from packaging import version _a = { """<""": operator.lt, """<=""": operator.le, """==""": operator.eq, """!=""": operator.ne, """>=""": operator.ge, """>""": operator.gt, } def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): if got_ver is None or want_ver is None: raise ValueError( f"""Unable to compare versions for {requirement}: need={want_ver} found={got_ver}. This is unusual. Consider""" f""" reinstalling {pkg}.""" ) if not ops[op](version.parse(__lowerCamelCase ), version.parse(__lowerCamelCase ) ): raise ImportError( f"""{requirement} is required for a normal functioning of this module, but found {pkg}=={got_ver}.{hint}""" ) def __a ( __lowerCamelCase, __lowerCamelCase = None ): UpperCAmelCase_ : List[Any] = f"""\n{hint}""" if hint is not None else "" # non-versioned check if re.match(r"^[\w_\-\d]+$", __lowerCamelCase ): UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = requirement, None, None else: UpperCAmelCase_ : int = re.findall(r"^([^!=<>\s]+)([\s!=<>]{1,2}.+)", __lowerCamelCase ) if not match: raise ValueError( "requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23, but" f""" got {requirement}""" ) UpperCAmelCase_ , UpperCAmelCase_ : Any = match[0] UpperCAmelCase_ : Tuple = want_full.split("," ) # there could be multiple requirements UpperCAmelCase_ : List[Any] = {} for w in want_range: UpperCAmelCase_ : List[str] = re.findall(r"^([\s!=<>]{1,2})(.+)", __lowerCamelCase ) if not match: raise ValueError( "requirement needs to be in the pip package format, .e.g., package_a==1.23, or package_b>=1.23," f""" but got {requirement}""" ) UpperCAmelCase_ , UpperCAmelCase_ : int = match[0] UpperCAmelCase_ : str = want_ver if op not in ops: raise ValueError(f"""{requirement}: need one of {list(ops.keys() )}, but got {op}""" ) # special case if pkg == "python": UpperCAmelCase_ : str = ".".join([str(__lowerCamelCase ) for x in sys.version_info[:3]] ) for op, want_ver in wanted.items(): _compare_versions(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) return # check if any version is installed try: UpperCAmelCase_ : Union[str, Any] = importlib.metadata.version(__lowerCamelCase ) except importlib.metadata.PackageNotFoundError: raise importlib.metadata.PackageNotFoundError( f"""The '{requirement}' distribution was not found and is required by this application. {hint}""" ) # check that the right version is installed if version number or a range was provided if want_ver is not None: for op, want_ver in wanted.items(): _compare_versions(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) def __a ( __lowerCamelCase ): UpperCAmelCase_ : str = "Try: pip install transformers -U or pip install -e \'.[dev]\' if you\'re working with git main" return require_version(__lowerCamelCase, __lowerCamelCase )
359
"""simple docstring""" import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = torch.nn.Linear(10 , 10 ) UpperCAmelCase_ : List[str] = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase_ : Optional[Any] = Accelerator() UpperCAmelCase_ : Tuple = accelerator.prepare(lowercase_ ) try: pickle.loads(pickle.dumps(lowercase_ ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
23
0
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: _a = None _a = logging.get_logger(__name__) _a = '▁' _a = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} _a = { 'vocab_file': {'google/pegasus-xsum': 'https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'}, 'tokenizer_file': { 'google/pegasus-xsum': 'https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json' }, } _a = { 'google/pegasus-xsum': 512, } class A_ (A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[int] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : Optional[int] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : Union[str, Any] = PegasusTokenizer SCREAMING_SNAKE_CASE__ : Optional[Any] = ["input_ids", "attention_mask"] def __init__( self , lowercase_=None , lowercase_=None , lowercase_="<pad>" , lowercase_="</s>" , lowercase_="<unk>" , lowercase_="<mask_2>" , lowercase_="<mask_1>" , lowercase_=None , lowercase_=103 , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Any = offset if additional_special_tokens is not None: if not isinstance(snake_case__ , snake_case__ ): raise TypeError( F"""additional_special_tokens should be of type {type(snake_case__ )}, but is""" F""" {type(snake_case__ )}""" ) UpperCAmelCase_ : List[Any] = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ F"""<unk_{i}>""" for i in range(len(snake_case__ ) , self.offset - 1 ) ] if len(set(snake_case__ ) ) != len(snake_case__ ): raise ValueError( "Please make sure that the provided additional_special_tokens do not contain an incorrectly" F""" shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.""" ) UpperCAmelCase_ : Union[str, Any] = additional_special_tokens_extended else: UpperCAmelCase_ : Union[str, Any] = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [F"""<unk_{i}>""" for i in range(2 , self.offset )] super().__init__( snake_case__ , tokenizer_file=snake_case__ , pad_token=snake_case__ , eos_token=snake_case__ , unk_token=snake_case__ , mask_token=snake_case__ , mask_token_sent=snake_case__ , offset=snake_case__ , additional_special_tokens=snake_case__ , **snake_case__ , ) UpperCAmelCase_ : Union[str, Any] = vocab_file UpperCAmelCase_ : Optional[int] = False if not self.vocab_file else True def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens ) + 3 ) ): raise ValueError( "There should be 3 special tokens: mask_token, pad_token, and eos_token +" F""" {len(self.additional_special_tokens )} additional_special_tokens, but got {all_special_ids}""" ) return [1 if x in all_special_ids else 0 for x in seq] def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = False ): """simple docstring""" if already_has_special_tokens: return self._special_token_mask(snake_case__ ) elif token_ids_a is None: return self._special_token_mask(snake_case__ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def UpperCamelCase__ ( self , lowercase_ , lowercase_=None ): """simple docstring""" if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(snake_case__ ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase_ : str = os.path.join( snake_case__ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case__ ): copyfile(self.vocab_file , snake_case__ ) return (out_vocab_file,)
360
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = {'ctrl': 'https://huggingface.co/ctrl/resolve/main/config.json'} class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """ctrl""" SCREAMING_SNAKE_CASE__ : Optional[int] = ["""past_key_values"""] SCREAMING_SNAKE_CASE__ : List[str] = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase_=24_6534 , lowercase_=256 , lowercase_=1280 , lowercase_=8192 , lowercase_=48 , lowercase_=16 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=1E-6 , lowercase_=0.02 , lowercase_=True , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Tuple = vocab_size UpperCAmelCase_ : Union[str, Any] = n_positions UpperCAmelCase_ : List[str] = n_embd UpperCAmelCase_ : Dict = n_layer UpperCAmelCase_ : Optional[int] = n_head UpperCAmelCase_ : List[str] = dff UpperCAmelCase_ : Tuple = resid_pdrop UpperCAmelCase_ : Optional[Any] = embd_pdrop UpperCAmelCase_ : str = layer_norm_epsilon UpperCAmelCase_ : List[str] = initializer_range UpperCAmelCase_ : List[str] = use_cache super().__init__(**lowercase_ )
23
0
"""simple docstring""" def __a ( __lowerCamelCase, __lowerCamelCase = False ): if not isinstance(lowerCAmelCase__, lowerCAmelCase__ ): UpperCAmelCase_ : Any = f"""Expected string as input, found {type(lowerCAmelCase__ )}""" raise ValueError(lowerCAmelCase__ ) if not isinstance(lowerCAmelCase__, lowerCAmelCase__ ): UpperCAmelCase_ : Union[str, Any] = f"""Expected boolean as use_pascal parameter, found {type(lowerCAmelCase__ )}""" raise ValueError(lowerCAmelCase__ ) UpperCAmelCase_ : Optional[int] = input_str.split("_" ) UpperCAmelCase_ : int = 0 if use_pascal else 1 UpperCAmelCase_ : Dict = words[start_index:] UpperCAmelCase_ : Tuple = [word[0].upper() + word[1:] for word in words_to_capitalize] UpperCAmelCase_ : Union[str, Any] = "" if use_pascal else words[0] return "".join([initial_word, *capitalized_words] ) if __name__ == "__main__": from doctest import testmod testmod()
361
"""simple docstring""" def __a ( __lowerCamelCase ): assert isinstance(__lowerCamelCase, __lowerCamelCase ), f"""The input value of [n={number}] is not an integer""" if number == 1: return 2 elif number < 1: UpperCAmelCase_ : str = f"""The input value of [n={number}] has to be > 0""" raise ValueError(__lowerCamelCase ) else: UpperCAmelCase_ : List[str] = sylvester(number - 1 ) UpperCAmelCase_ : List[str] = num - 1 UpperCAmelCase_ : List[str] = num return lower * upper + 1 if __name__ == "__main__": print(f"""The 8th number in Sylvester's sequence: {sylvester(8)}""")
23
0
import re def __a ( __lowerCamelCase ): UpperCAmelCase_ : Optional[int] = re.compile(r"^(\+91[\-\s]?)?[0]?(91)?[789]\d{9}$" ) if match := re.search(lowercase__, lowercase__ ): return match.string == phone return False if __name__ == "__main__": print(indian_phone_validator('+918827897895'))
362
"""simple docstring""" import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class A_ (lowercase__ ,lowercase__ ,unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = IFImgaImgSuperResolutionPipeline SCREAMING_SNAKE_CASE__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""width""", """height"""} SCREAMING_SNAKE_CASE__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""original_image"""} ) SCREAMING_SNAKE_CASE__ : List[Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} def UpperCamelCase__ ( self ): """simple docstring""" return self._get_superresolution_dummy_components() def UpperCamelCase__ ( self , lowercase_ , lowercase_=0 ): """simple docstring""" if str(lowercase_ ).startswith("mps" ): UpperCAmelCase_ : Optional[Any] = torch.manual_seed(lowercase_ ) else: UpperCAmelCase_ : Union[str, Any] = torch.Generator(device=lowercase_ ).manual_seed(lowercase_ ) UpperCAmelCase_ : Any = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) UpperCAmelCase_ : Optional[int] = floats_tensor((1, 3, 16, 16) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) UpperCAmelCase_ : int = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def UpperCamelCase__ ( self ): """simple docstring""" # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_save_load_local() def UpperCamelCase__ ( self ): """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
23
0
"""simple docstring""" import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class A_ (enum.Enum ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = 0 SCREAMING_SNAKE_CASE__ : Dict = 1 SCREAMING_SNAKE_CASE__ : str = 2 @add_end_docstrings(__lowercase ) class A_ (__lowercase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = ''' In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The voice of Nicholas\'s young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision and denounces one of the men as a horse thief. Although his father initially slaps him for making such an accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop, begging for his blessing. <eod> </s> <eos> ''' def __init__( self , *lowercase_ , **lowercase_ ): """simple docstring""" super().__init__(*lowercase_ , **lowercase_ ) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_CAUSAL_LM_MAPPING ) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. UpperCAmelCase_ : Dict = None if self.model.config.prefix is not None: UpperCAmelCase_ : List[str] = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. UpperCAmelCase_ : int = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. UpperCAmelCase_ : Optional[int] = self._sanitize_parameters(prefix=lowercase_ , **self._forward_params ) UpperCAmelCase_ : Any = {**self._preprocess_params, **preprocess_params} UpperCAmelCase_ : Dict = {**self._forward_params, **forward_params} def UpperCamelCase__ ( self , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_=None , lowercase_=None , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : int = {} if prefix is not None: UpperCAmelCase_ : Union[str, Any] = prefix if prefix: UpperCAmelCase_ : Optional[Any] = self.tokenizer( lowercase_ , padding=lowercase_ , add_special_tokens=lowercase_ , return_tensors=self.framework ) UpperCAmelCase_ : Any = prefix_inputs["""input_ids"""].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( F"""{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected""" " [None, 'hole']" ) UpperCAmelCase_ : Any = handle_long_generation preprocess_params.update(lowercase_ ) UpperCAmelCase_ : str = generate_kwargs UpperCAmelCase_ : Optional[Any] = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_full_text`" ) if return_tensors is not None: raise ValueError("`return_full_text` is mutually exclusive with `return_tensors`" ) UpperCAmelCase_ : List[str] = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_tensors`" ) UpperCAmelCase_ : Optional[Any] = ReturnType.TENSORS if return_type is not None: UpperCAmelCase_ : Dict = return_type if clean_up_tokenization_spaces is not None: UpperCAmelCase_ : str = clean_up_tokenization_spaces if stop_sequence is not None: UpperCAmelCase_ : List[str] = self.tokenizer.encode(lowercase_ , add_special_tokens=lowercase_ ) if len(lowercase_ ) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim." ) UpperCAmelCase_ : str = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def UpperCamelCase__ ( self , *lowercase_ , **lowercase_ ): """simple docstring""" if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({"add_space_before_punct_symbol": True} ) return super()._parse_and_tokenize(*lowercase_ , **lowercase_ ) def __call__( self , lowercase_ , **lowercase_ ): """simple docstring""" return super().__call__(lowercase_ , **lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_="" , lowercase_=None , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.tokenizer( prefix + prompt_text , padding=lowercase_ , add_special_tokens=lowercase_ , return_tensors=self.framework ) UpperCAmelCase_ : int = prompt_text if handle_long_generation == "hole": UpperCAmelCase_ : int = inputs["""input_ids"""].shape[-1] if "max_new_tokens" in generate_kwargs: UpperCAmelCase_ : Any = generate_kwargs["""max_new_tokens"""] else: UpperCAmelCase_ : int = generate_kwargs.get("max_length" , self.model.config.max_length ) - cur_len if new_tokens < 0: raise ValueError("We cannot infer how many new tokens are expected" ) if cur_len + new_tokens > self.tokenizer.model_max_length: UpperCAmelCase_ : Tuple = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( "We cannot use `hole` to handle this generation the number of desired tokens exceeds the" " models max length" ) UpperCAmelCase_ : Any = inputs["""input_ids"""][:, -keep_length:] if "attention_mask" in inputs: UpperCAmelCase_ : Any = inputs["""attention_mask"""][:, -keep_length:] return inputs def UpperCamelCase__ ( self , lowercase_ , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = model_inputs["""input_ids"""] UpperCAmelCase_ : Any = model_inputs.get("attention_mask" , lowercase_ ) # Allow empty prompts if input_ids.shape[1] == 0: UpperCAmelCase_ : Tuple = None UpperCAmelCase_ : Optional[Any] = None UpperCAmelCase_ : str = 1 else: UpperCAmelCase_ : List[Any] = input_ids.shape[0] UpperCAmelCase_ : Dict = model_inputs.pop("prompt_text" ) # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. UpperCAmelCase_ : Union[str, Any] = generate_kwargs.pop("prefix_length" , 0 ) if prefix_length > 0: UpperCAmelCase_ : List[Any] = """max_new_tokens""" in generate_kwargs or ( """generation_config""" in generate_kwargs and generate_kwargs["""generation_config"""].max_new_tokens is not None ) if not has_max_new_tokens: UpperCAmelCase_ : Optional[Any] = generate_kwargs.get("max_length" ) or self.model.config.max_length generate_kwargs["max_length"] += prefix_length UpperCAmelCase_ : Any = """min_new_tokens""" in generate_kwargs or ( """generation_config""" in generate_kwargs and generate_kwargs["""generation_config"""].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL UpperCAmelCase_ : List[Any] = self.model.generate(input_ids=lowercase_ , attention_mask=lowercase_ , **lowercase_ ) UpperCAmelCase_ : Optional[int] = generated_sequence.shape[0] if self.framework == "pt": UpperCAmelCase_ : Dict = generated_sequence.reshape(lowercase_ , out_b // in_b , *generated_sequence.shape[1:] ) elif self.framework == "tf": UpperCAmelCase_ : int = tf.reshape(lowercase_ , (in_b, out_b // in_b, *generated_sequence.shape[1:]) ) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def UpperCamelCase__ ( self , lowercase_ , lowercase_=ReturnType.FULL_TEXT , lowercase_=True ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = model_outputs["""generated_sequence"""][0] UpperCAmelCase_ : Any = model_outputs["""input_ids"""] UpperCAmelCase_ : List[Any] = model_outputs["""prompt_text"""] UpperCAmelCase_ : str = generated_sequence.numpy().tolist() UpperCAmelCase_ : Union[str, Any] = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: UpperCAmelCase_ : Optional[Any] = {"""generated_token_ids""": sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text UpperCAmelCase_ : str = self.tokenizer.decode( lowercase_ , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: UpperCAmelCase_ : Tuple = 0 else: UpperCAmelCase_ : Any = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=lowercase_ , clean_up_tokenization_spaces=lowercase_ , ) ) if return_type == ReturnType.FULL_TEXT: UpperCAmelCase_ : Union[str, Any] = prompt_text + text[prompt_length:] else: UpperCAmelCase_ : int = text[prompt_length:] UpperCAmelCase_ : List[str] = {"""generated_text""": all_text} records.append(lowercase_ ) return records
363
"""simple docstring""" import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = "ylacombe/bark-small" UpperCAmelCase_ : Union[str, Any] = tempfile.mkdtemp() UpperCAmelCase_ : List[str] = "en_speaker_1" UpperCAmelCase_ : Tuple = "This is a test string" UpperCAmelCase_ : List[Any] = "speaker_embeddings_path.json" UpperCAmelCase_ : Any = "speaker_embeddings" def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" return AutoTokenizer.from_pretrained(self.checkpoint , **lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.get_tokenizer() UpperCAmelCase_ : Union[str, Any] = BarkProcessor(tokenizer=lowercase_ ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase_ : Optional[int] = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) UpperCAmelCase_ : Dict = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) UpperCAmelCase_ : Union[str, Any] = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token="(BOS)" , eos_token="(EOS)" , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) UpperCAmelCase_ : int = 35 UpperCAmelCase_ : Optional[Any] = 2 UpperCAmelCase_ : List[Any] = 8 UpperCAmelCase_ : Optional[Any] = { "semantic_prompt": np.ones(lowercase_ ), "coarse_prompt": np.ones((nb_codebooks_coarse, seq_len) ), "fine_prompt": np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset UpperCAmelCase_ : Dict = processor(text=self.input_string , voice_preset=lowercase_ ) UpperCAmelCase_ : List[str] = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowercase_ , np.array([] ) ).tolist() ) # test loading voice preset from npz file UpperCAmelCase_ : Tuple = os.path.join(self.tmpdirname , "file.npz" ) np.savez(lowercase_ , **lowercase_ ) UpperCAmelCase_ : Optional[int] = processor(text=self.input_string , voice_preset=lowercase_ ) UpperCAmelCase_ : List[str] = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowercase_ , np.array([] ) ).tolist() ) # test loading voice preset from the hub UpperCAmelCase_ : Tuple = processor(text=self.input_string , voice_preset=self.voice_preset ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = self.get_tokenizer() UpperCAmelCase_ : Optional[Any] = BarkProcessor(tokenizer=lowercase_ ) UpperCAmelCase_ : Tuple = processor(text=self.input_string ) UpperCAmelCase_ : Union[str, Any] = tokenizer( self.input_string , padding="max_length" , max_length=256 , add_special_tokens=lowercase_ , return_attention_mask=lowercase_ , return_token_type_ids=lowercase_ , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
23
0
"""simple docstring""" # # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # You may need to add --master_addr $MASTER_ADDR --master_port $MASTER_PORT if using a custom addr:port # # You can also use the rdzv API: --rdzv_endpoint $MASTER_ADDR:$MASTER_PORT --rdzv_backend c10d # # use torch.distributed.launch instead of torch.distributed.run for torch < 1.9 # # If you get a hanging in `barrier` calls you have some network issues, you may try to debug this with: # # NCCL_DEBUG=INFO python -m torch.distributed.run --nproc_per_node 2 --nnodes 1 torch-distributed-gpu-test.py # # which should tell you what's going on behind the scenes. # # # This script can be run via `srun` in the SLURM environment as well. Here is a SLURM script that # runs on 2 nodes of 4 gpus per node: # # #SBATCH --job-name=test-nodes # name # #SBATCH --nodes=2 # nodes # #SBATCH --ntasks-per-node=1 # crucial - only 1 task per dist per node! # #SBATCH --cpus-per-task=10 # number of cores per tasks # #SBATCH --gres=gpu:4 # number of gpus # #SBATCH --time 0:05:00 # maximum execution time (HH:MM:SS) # #SBATCH --output=%x-%j.out # output file name # # GPUS_PER_NODE=4 # MASTER_ADDR=$(scontrol show hostnames $SLURM_JOB_NODELIST | head -n 1) # MASTER_PORT=6000 # # srun --jobid $SLURM_JOBID bash -c 'python -m torch.distributed.run \ # --nproc_per_node $GPUS_PER_NODE --nnodes $SLURM_NNODES --node_rank $SLURM_PROCID \ # --master_addr $MASTER_ADDR --master_port $MASTER_PORT \ # torch-distributed-gpu-test.py' # import fcntl import os import socket import torch import torch.distributed as dist def __a ( *__lowerCamelCase ): with open(a_, "r" ) as fh: fcntl.flock(a_, fcntl.LOCK_EX ) try: print(*a_ ) finally: fcntl.flock(a_, fcntl.LOCK_UN ) _a = int(os.environ['LOCAL_RANK']) torch.cuda.set_device(local_rank) _a = torch.device('cuda', local_rank) _a = socket.gethostname() _a = f"""[{hostname}-{local_rank}]""" try: # test distributed dist.init_process_group('nccl') dist.all_reduce(torch.ones(1).to(device), op=dist.ReduceOp.SUM) dist.barrier() # test cuda is available and can allocate memory torch.cuda.is_available() torch.ones(1).cuda(local_rank) # global rank _a = dist.get_rank() _a = dist.get_world_size() printflock(f"""{gpu} is OK (global rank: {rank}/{world_size})""") dist.barrier() if rank == 0: printflock(f"""pt={torch.__version__}, cuda={torch.version.cuda}, nccl={torch.cuda.nccl.version()}""") except Exception: printflock(f"""{gpu} is broken""") raise
364
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def __a ( __lowerCamelCase, __lowerCamelCase=False ): UpperCAmelCase_ : Optional[int] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""deit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""deit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""deit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""deit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""deit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""deit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""deit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""deit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""deit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""deit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "deit.embeddings.cls_token"), ("dist_token", "deit.embeddings.distillation_token"), ("patch_embed.proj.weight", "deit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "deit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "deit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" UpperCAmelCase_ : Dict = [(pair[0], pair[1][4:]) if pair[1].startswith("deit" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("norm.weight", "deit.layernorm.weight"), ("norm.bias", "deit.layernorm.bias"), ("head.weight", "cls_classifier.weight"), ("head.bias", "cls_classifier.bias"), ("head_dist.weight", "distillation_classifier.weight"), ("head_dist.bias", "distillation_classifier.bias"), ] ) return rename_keys def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=False ): for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase_ : int = "" else: UpperCAmelCase_ : Union[str, Any] = "deit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase_ : Tuple = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) UpperCAmelCase_ : Dict = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ : Union[str, Any] = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase_ : Any = in_proj_bias[: config.hidden_size] UpperCAmelCase_ : Optional[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase_ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase_ : List[Any] = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase_ : str = in_proj_bias[-config.hidden_size :] def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Tuple = dct.pop(__lowerCamelCase ) UpperCAmelCase_ : Tuple = val def __a ( ): UpperCAmelCase_ : Union[str, Any] = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ : str = Image.open(requests.get(__lowerCamelCase, stream=__lowerCamelCase ).raw ) return im @torch.no_grad() def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : List[str] = DeiTConfig() # all deit models have fine-tuned heads UpperCAmelCase_ : Optional[int] = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase_ : Tuple = 1000 UpperCAmelCase_ : str = "huggingface/label-files" UpperCAmelCase_ : str = "imagenet-1k-id2label.json" UpperCAmelCase_ : List[Any] = json.load(open(hf_hub_download(__lowerCamelCase, __lowerCamelCase, repo_type="dataset" ), "r" ) ) UpperCAmelCase_ : List[str] = {int(__lowerCamelCase ): v for k, v in idalabel.items()} UpperCAmelCase_ : Any = idalabel UpperCAmelCase_ : int = {v: k for k, v in idalabel.items()} UpperCAmelCase_ : Any = int(deit_name[-6:-4] ) UpperCAmelCase_ : Dict = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("tiny" ): UpperCAmelCase_ : Any = 192 UpperCAmelCase_ : Union[str, Any] = 768 UpperCAmelCase_ : Union[str, Any] = 12 UpperCAmelCase_ : int = 3 elif deit_name[9:].startswith("small" ): UpperCAmelCase_ : List[str] = 384 UpperCAmelCase_ : List[str] = 1536 UpperCAmelCase_ : Dict = 12 UpperCAmelCase_ : Any = 6 if deit_name[9:].startswith("base" ): pass elif deit_name[4:].startswith("large" ): UpperCAmelCase_ : int = 1024 UpperCAmelCase_ : List[Any] = 4096 UpperCAmelCase_ : Optional[int] = 24 UpperCAmelCase_ : int = 16 # load original model from timm UpperCAmelCase_ : Union[str, Any] = timm.create_model(__lowerCamelCase, pretrained=__lowerCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCAmelCase_ : Optional[Any] = timm_model.state_dict() UpperCAmelCase_ : Tuple = create_rename_keys(__lowerCamelCase, __lowerCamelCase ) for src, dest in rename_keys: rename_key(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) read_in_q_k_v(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) # load HuggingFace model UpperCAmelCase_ : str = DeiTForImageClassificationWithTeacher(__lowerCamelCase ).eval() model.load_state_dict(__lowerCamelCase ) # Check outputs on an image, prepared by DeiTImageProcessor UpperCAmelCase_ : Union[str, Any] = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 UpperCAmelCase_ : Optional[Any] = DeiTImageProcessor(size=__lowerCamelCase, crop_size=config.image_size ) UpperCAmelCase_ : Any = image_processor(images=prepare_img(), return_tensors="pt" ) UpperCAmelCase_ : int = encoding["pixel_values"] UpperCAmelCase_ : Optional[Any] = model(__lowerCamelCase ) UpperCAmelCase_ : Any = timm_model(__lowerCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__lowerCamelCase, outputs.logits, atol=1E-3 ) Path(__lowerCamelCase ).mkdir(exist_ok=__lowerCamelCase ) print(f"""Saving model {deit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCamelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--deit_name', default='vit_deit_base_distilled_patch16_224', type=str, help='Name of the DeiT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _a = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
23
0
"""simple docstring""" import argparse import os import re import packaging.version _a = 'examples/' _a = { 'examples': (re.compile(R'^check_min_version\("[^"]+"\)\s*$', re.MULTILINE), 'check_min_version("VERSION")\n'), 'init': (re.compile(R'^__version__\s+=\s+"([^"]+)"\s*$', re.MULTILINE), '__version__ = "VERSION"\n'), 'setup': (re.compile(R'^(\s*)version\s*=\s*"[^"]+",', re.MULTILINE), R'\1version="VERSION",'), 'doc': (re.compile(R'^(\s*)release\s*=\s*"[^"]+"$', re.MULTILINE), 'release = "VERSION"\n'), } _a = { 'init': 'src/transformers/__init__.py', 'setup': 'setup.py', } _a = 'README.md' def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): with open(__a, "r", encoding="utf-8", newline="\n" ) as f: UpperCAmelCase_ : Tuple = f.read() UpperCAmelCase_ : str = REPLACE_PATTERNS[pattern] UpperCAmelCase_ : List[str] = replace.replace("VERSION", __a ) UpperCAmelCase_ : List[Any] = re_pattern.sub(__a, __a ) with open(__a, "w", encoding="utf-8", newline="\n" ) as f: f.write(__a ) def __a ( __lowerCamelCase ): for folder, directories, fnames in os.walk(__a ): # Removing some of the folders with non-actively maintained examples from the walk if "research_projects" in directories: directories.remove("research_projects" ) if "legacy" in directories: directories.remove("legacy" ) for fname in fnames: if fname.endswith(".py" ): update_version_in_file(os.path.join(__a, __a ), __a, pattern="examples" ) def __a ( __lowerCamelCase, __lowerCamelCase=False ): for pattern, fname in REPLACE_FILES.items(): update_version_in_file(__a, __a, __a ) if not patch: update_version_in_examples(__a ) def __a ( ): UpperCAmelCase_ : Optional[Any] = '''🤗 Transformers currently provides the following architectures''' UpperCAmelCase_ : str = '''1. Want to contribute a new model?''' with open(__a, "r", encoding="utf-8", newline="\n" ) as f: UpperCAmelCase_ : Optional[int] = f.readlines() # Find the start of the list. UpperCAmelCase_ : Optional[int] = 0 while not lines[start_index].startswith(_start_prompt ): start_index += 1 start_index += 1 UpperCAmelCase_ : List[Any] = start_index # Update the lines in the model list. while not lines[index].startswith(_end_prompt ): if lines[index].startswith("1." ): UpperCAmelCase_ : Tuple = lines[index].replace( "https://huggingface.co/docs/transformers/main/model_doc", "https://huggingface.co/docs/transformers/model_doc", ) index += 1 with open(__a, "w", encoding="utf-8", newline="\n" ) as f: f.writelines(__a ) def __a ( ): with open(REPLACE_FILES["init"], "r" ) as f: UpperCAmelCase_ : Optional[Any] = f.read() UpperCAmelCase_ : Optional[Any] = REPLACE_PATTERNS['''init'''][0].search(__a ).groups()[0] return packaging.version.parse(__a ) def __a ( __lowerCamelCase=False ): UpperCAmelCase_ : Optional[Any] = get_version() if patch and default_version.is_devrelease: raise ValueError("Can\'t create a patch version from the dev branch, checkout a released version!" ) if default_version.is_devrelease: UpperCAmelCase_ : List[Any] = default_version.base_version elif patch: UpperCAmelCase_ : str = f"""{default_version.major}.{default_version.minor}.{default_version.micro + 1}""" else: UpperCAmelCase_ : List[str] = f"""{default_version.major}.{default_version.minor + 1}.0""" # Now let's ask nicely if that's the right one. UpperCAmelCase_ : Dict = input(f"""Which version are you releasing? [{default_version}]""" ) if len(__a ) == 0: UpperCAmelCase_ : int = default_version print(f"""Updating version to {version}.""" ) global_version_update(__a, patch=__a ) if not patch: print("Cleaning main README, don\'t forget to run `make fix-copies`." ) clean_main_ref_in_model_list() def __a ( ): UpperCAmelCase_ : str = get_version() UpperCAmelCase_ : int = f"""{current_version.major}.{current_version.minor + 1}.0.dev0""" UpperCAmelCase_ : List[Any] = current_version.base_version # Check with the user we got that right. UpperCAmelCase_ : Union[str, Any] = input(f"""Which version are we developing now? [{dev_version}]""" ) if len(__a ) == 0: UpperCAmelCase_ : List[str] = dev_version print(f"""Updating version to {version}.""" ) global_version_update(__a ) print("Cleaning main README, don\'t forget to run `make fix-copies`." ) clean_main_ref_in_model_list() if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument('--post_release', action='store_true', help='Whether this is pre or post release.') parser.add_argument('--patch', action='store_true', help='Whether or not this is a patch release.') _a = parser.parse_args() if not args.post_release: pre_release_work(patch=args.patch) elif args.patch: print('Nothing to do after a patch :-)') else: post_release_work()
365
"""simple docstring""" import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights UpperCAmelCase_ : Optional[Any] = FlaxDiffusionPipeline.from_pretrained( "hf-internal-testing/tiny-stable-diffusion-pipe" , safety_checker=lowercase_ , cache_dir=lowercase_ ) UpperCAmelCase_ : List[Any] = [t[-1] for t in os.walk(os.path.join(lowercase_ , os.listdir(lowercase_ )[0] , "snapshots" ) )] UpperCAmelCase_ : Dict = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith(".bin" ) for f in files ) @slow @require_flax class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( "hf-internal-testing/tiny-stable-diffusion-pipe" , safety_checker=lowercase_ ) UpperCAmelCase_ : Tuple = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : List[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase_ : List[str] = 4 UpperCAmelCase_ : Tuple = jax.device_count() UpperCAmelCase_ : Optional[int] = num_samples * [prompt] UpperCAmelCase_ : List[Any] = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : int = replicate(lowercase_ ) UpperCAmelCase_ : str = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : Dict = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_51_47_45 ) < 1E-3 assert np.abs(np.abs(lowercase_ , dtype=np.floataa ).sum() - 4_99_47.8_75 ) < 5E-1 UpperCAmelCase_ : List[Any] = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(lowercase_ ) == num_samples def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="flax" , safety_checker=lowercase_ ) UpperCAmelCase_ : Optional[int] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : str = jax.random.PRNGKey(0 ) UpperCAmelCase_ : Union[str, Any] = 50 UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[str] = num_samples * [prompt] UpperCAmelCase_ : Union[str, Any] = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Any = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : int = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_65_24_01) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_38_38_08.2) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ ) UpperCAmelCase_ : Any = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : str = jax.random.PRNGKey(0 ) UpperCAmelCase_ : str = 50 UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : Any = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Dict = replicate(lowercase_ ) UpperCAmelCase_ : str = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Union[str, Any] = shard(lowercase_ ) UpperCAmelCase_ : List[Any] = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : str = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa ) UpperCAmelCase_ : List[Any] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : Dict = jax.random.PRNGKey(0 ) UpperCAmelCase_ : Optional[int] = 50 UpperCAmelCase_ : Optional[int] = jax.device_count() UpperCAmelCase_ : str = num_samples * [prompt] UpperCAmelCase_ : int = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Union[str, Any] = replicate(lowercase_ ) UpperCAmelCase_ : Union[str, Any] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[int] = shard(lowercase_ ) UpperCAmelCase_ : Any = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = FlaxDDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="scaled_linear" , set_alpha_to_one=lowercase_ , steps_offset=1 , ) UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , scheduler=lowercase_ , safety_checker=lowercase_ , ) UpperCAmelCase_ : List[Any] = scheduler.create_state() UpperCAmelCase_ : int = scheduler_state UpperCAmelCase_ : Union[str, Any] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : Optional[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase_ : int = 50 UpperCAmelCase_ : str = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : int = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : int = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = shard(lowercase_ ) UpperCAmelCase_ : Any = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_45_04_39_45) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_34_76_93.5) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : Union[str, Any] = jax.random.split(jax.random.PRNGKey(0 ) , lowercase_ ) UpperCAmelCase_ , UpperCAmelCase_ : str = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ , ) UpperCAmelCase_ : Any = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = pipeline.prepare_inputs(lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : List[Any] = pipeline(lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase_ : int = images[2, 0, 256, 10:17, 1] # With memory efficient attention UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ , use_memory_efficient_attention=lowercase_ , ) UpperCAmelCase_ : str = replicate(lowercase_ ) UpperCAmelCase_ : str = pipeline.prepare_inputs(lowercase_ ) UpperCAmelCase_ : Optional[int] = shard(lowercase_ ) UpperCAmelCase_ : str = pipeline(lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase_ : Optional[int] = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1E-2
23
0
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _a = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f"""transformer.encoder.layers.{i}.self_attn.out_proj.weight""", f"""encoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (f"""transformer.encoder.layers.{i}.self_attn.out_proj.bias""", f"""encoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.linear1.weight""", f"""encoder.layers.{i}.fc1.weight""")) rename_keys.append((f"""transformer.encoder.layers.{i}.linear1.bias""", f"""encoder.layers.{i}.fc1.bias""")) rename_keys.append((f"""transformer.encoder.layers.{i}.linear2.weight""", f"""encoder.layers.{i}.fc2.weight""")) rename_keys.append((f"""transformer.encoder.layers.{i}.linear2.bias""", f"""encoder.layers.{i}.fc2.bias""")) rename_keys.append( (f"""transformer.encoder.layers.{i}.norm1.weight""", f"""encoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((f"""transformer.encoder.layers.{i}.norm1.bias""", f"""encoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append((f"""transformer.encoder.layers.{i}.norm2.weight""", f"""encoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((f"""transformer.encoder.layers.{i}.norm2.bias""", f"""encoder.layers.{i}.final_layer_norm.bias""")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f"""transformer.decoder.layers.{i}.self_attn.out_proj.weight""", f"""decoder.layers.{i}.self_attn.out_proj.weight""") ) rename_keys.append( (f"""transformer.decoder.layers.{i}.self_attn.out_proj.bias""", f"""decoder.layers.{i}.self_attn.out_proj.bias""") ) rename_keys.append( ( f"""transformer.decoder.layers.{i}.multihead_attn.out_proj.weight""", f"""decoder.layers.{i}.encoder_attn.out_proj.weight""", ) ) rename_keys.append( ( f"""transformer.decoder.layers.{i}.multihead_attn.out_proj.bias""", f"""decoder.layers.{i}.encoder_attn.out_proj.bias""", ) ) rename_keys.append((f"""transformer.decoder.layers.{i}.linear1.weight""", f"""decoder.layers.{i}.fc1.weight""")) rename_keys.append((f"""transformer.decoder.layers.{i}.linear1.bias""", f"""decoder.layers.{i}.fc1.bias""")) rename_keys.append((f"""transformer.decoder.layers.{i}.linear2.weight""", f"""decoder.layers.{i}.fc2.weight""")) rename_keys.append((f"""transformer.decoder.layers.{i}.linear2.bias""", f"""decoder.layers.{i}.fc2.bias""")) rename_keys.append( (f"""transformer.decoder.layers.{i}.norm1.weight""", f"""decoder.layers.{i}.self_attn_layer_norm.weight""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.norm1.bias""", f"""decoder.layers.{i}.self_attn_layer_norm.bias""")) rename_keys.append( (f"""transformer.decoder.layers.{i}.norm2.weight""", f"""decoder.layers.{i}.encoder_attn_layer_norm.weight""") ) rename_keys.append( (f"""transformer.decoder.layers.{i}.norm2.bias""", f"""decoder.layers.{i}.encoder_attn_layer_norm.bias""") ) rename_keys.append((f"""transformer.decoder.layers.{i}.norm3.weight""", f"""decoder.layers.{i}.final_layer_norm.weight""")) rename_keys.append((f"""transformer.decoder.layers.{i}.norm3.bias""", f"""decoder.layers.{i}.final_layer_norm.bias""")) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ('input_proj.weight', 'input_projection.weight'), ('input_proj.bias', 'input_projection.bias'), ('query_embed.weight', 'query_position_embeddings.weight'), ('transformer.encoder.norm.weight', 'encoder.layernorm.weight'), ('transformer.encoder.norm.bias', 'encoder.layernorm.bias'), ('transformer.decoder.norm.weight', 'decoder.layernorm.weight'), ('transformer.decoder.norm.bias', 'decoder.layernorm.bias'), ('class_embed.weight', 'class_labels_classifier.weight'), ('class_embed.bias', 'class_labels_classifier.bias'), ('bbox_embed.layers.0.weight', 'bbox_predictor.layers.0.weight'), ('bbox_embed.layers.0.bias', 'bbox_predictor.layers.0.bias'), ('bbox_embed.layers.1.weight', 'bbox_predictor.layers.1.weight'), ('bbox_embed.layers.1.bias', 'bbox_predictor.layers.1.bias'), ('bbox_embed.layers.2.weight', 'bbox_predictor.layers.2.weight'), ('bbox_embed.layers.2.bias', 'bbox_predictor.layers.2.bias'), ] ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : int = state_dict.pop(lowerCAmelCase__ ) UpperCAmelCase_ : List[Any] = val def __a ( __lowerCamelCase ): UpperCAmelCase_ : Union[str, Any] = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: UpperCAmelCase_ : List[Any] = key.replace("backbone.0.body", "backbone.conv_encoder.model" ) UpperCAmelCase_ : str = value else: UpperCAmelCase_ : List[str] = value return new_state_dict def __a ( __lowerCamelCase ): UpperCAmelCase_ : str = """""" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) UpperCAmelCase_ : Tuple = state_dict.pop(f"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight""" ) UpperCAmelCase_ : int = state_dict.pop(f"""{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ : Union[str, Any] = in_proj_weight[:256, :] UpperCAmelCase_ : Optional[Any] = in_proj_bias[:256] UpperCAmelCase_ : Optional[int] = in_proj_weight[256:512, :] UpperCAmelCase_ : Optional[Any] = in_proj_bias[256:512] UpperCAmelCase_ : List[Any] = in_proj_weight[-256:, :] UpperCAmelCase_ : Optional[Any] = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention UpperCAmelCase_ : Tuple = state_dict.pop(f"""{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight""" ) UpperCAmelCase_ : Optional[int] = state_dict.pop(f"""{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ : Tuple = in_proj_weight[:256, :] UpperCAmelCase_ : Optional[int] = in_proj_bias[:256] UpperCAmelCase_ : str = in_proj_weight[256:512, :] UpperCAmelCase_ : str = in_proj_bias[256:512] UpperCAmelCase_ : Optional[int] = in_proj_weight[-256:, :] UpperCAmelCase_ : List[Any] = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention UpperCAmelCase_ : List[str] = state_dict.pop( f"""{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight""" ) UpperCAmelCase_ : Optional[int] = state_dict.pop(f"""{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias""" ) # next, add query, keys and values (in that order) of cross-attention to the state dict UpperCAmelCase_ : Optional[Any] = in_proj_weight_cross_attn[:256, :] UpperCAmelCase_ : List[Any] = in_proj_bias_cross_attn[:256] UpperCAmelCase_ : List[Any] = in_proj_weight_cross_attn[256:512, :] UpperCAmelCase_ : List[Any] = in_proj_bias_cross_attn[256:512] UpperCAmelCase_ : Any = in_proj_weight_cross_attn[-256:, :] UpperCAmelCase_ : Optional[Any] = in_proj_bias_cross_attn[-256:] def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Union[str, Any] = image.size UpperCAmelCase_ : Optional[Any] = max(lowerCAmelCase__, lowerCAmelCase__ ) UpperCAmelCase_ : Dict = 800 if """detection""" in checkpoint_url else 1000 UpperCAmelCase_ : Optional[int] = target_max_size / current_max_size UpperCAmelCase_ : List[str] = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def __a ( __lowerCamelCase ): UpperCAmelCase_ : Union[str, Any] = F.to_tensor(lowerCAmelCase__ ) UpperCAmelCase_ : str = F.normalize(lowerCAmelCase__, mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): logger.info("Converting model..." ) # load original state dict UpperCAmelCase_ : Any = torch.hub.load_state_dict_from_url(lowerCAmelCase__, map_location="cpu" ) # rename keys for src, dest in rename_keys: rename_key(lowerCAmelCase__, lowerCAmelCase__, lowerCAmelCase__ ) UpperCAmelCase_ : Tuple = rename_backbone_keys(lowerCAmelCase__ ) # query, key and value matrices need special treatment read_in_q_k_v(lowerCAmelCase__ ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them UpperCAmelCase_ : Tuple = """model.""" for key in state_dict.copy().keys(): if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): UpperCAmelCase_ : Union[str, Any] = state_dict.pop(lowerCAmelCase__ ) UpperCAmelCase_ : Dict = val # create HuggingFace model and load state dict UpperCAmelCase_ : Union[str, Any] = TableTransformerConfig( backbone="resnet18", mask_loss_coefficient=1, dice_loss_coefficient=1, ce_loss_coefficient=1, bbox_loss_coefficient=5, giou_loss_coefficient=2, eos_coefficient=0.4, class_cost=1, bbox_cost=5, giou_cost=2, ) if "detection" in checkpoint_url: UpperCAmelCase_ : List[str] = 15 UpperCAmelCase_ : Union[str, Any] = 2 UpperCAmelCase_ : str = {0: """table""", 1: """table rotated"""} UpperCAmelCase_ : List[Any] = idalabel UpperCAmelCase_ : int = {v: k for k, v in idalabel.items()} else: UpperCAmelCase_ : Optional[Any] = 125 UpperCAmelCase_ : Any = 6 UpperCAmelCase_ : Dict = { 0: """table""", 1: """table column""", 2: """table row""", 3: """table column header""", 4: """table projected row header""", 5: """table spanning cell""", } UpperCAmelCase_ : Any = idalabel UpperCAmelCase_ : int = {v: k for k, v in idalabel.items()} UpperCAmelCase_ : str = DetrImageProcessor( format="coco_detection", max_size=800 if "detection" in checkpoint_url else 1000 ) UpperCAmelCase_ : Optional[Any] = TableTransformerForObjectDetection(lowerCAmelCase__ ) model.load_state_dict(lowerCAmelCase__ ) model.eval() # verify our conversion UpperCAmelCase_ : str = """example_pdf.png""" if """detection""" in checkpoint_url else """example_table.png""" UpperCAmelCase_ : List[Any] = hf_hub_download(repo_id="nielsr/example-pdf", repo_type="dataset", filename=lowerCAmelCase__ ) UpperCAmelCase_ : Optional[Any] = Image.open(lowerCAmelCase__ ).convert("RGB" ) UpperCAmelCase_ : int = normalize(resize(lowerCAmelCase__, lowerCAmelCase__ ) ).unsqueeze(0 ) UpperCAmelCase_ : Dict = model(lowerCAmelCase__ ) if "detection" in checkpoint_url: UpperCAmelCase_ : Optional[int] = (1, 15, 3) UpperCAmelCase_ : Optional[int] = torch.tensor( [[-6.7897, -16.9985, 6.7937], [-8.0186, -22.2192, 6.9677], [-7.3117, -21.0708, 7.4055]] ) UpperCAmelCase_ : str = torch.tensor([[0.4867, 0.1767, 0.6732], [0.6718, 0.4479, 0.3830], [0.4716, 0.1760, 0.6364]] ) else: UpperCAmelCase_ : List[str] = (1, 125, 7) UpperCAmelCase_ : Dict = torch.tensor( [[-18.1430, -8.3214, 4.8274], [-18.4685, -7.1361, -4.2667], [-26.3693, -9.3429, -4.9962]] ) UpperCAmelCase_ : Any = torch.tensor([[0.4983, 0.5595, 0.9440], [0.4916, 0.6315, 0.5954], [0.6108, 0.8637, 0.1135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3], lowerCAmelCase__, atol=1E-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3], lowerCAmelCase__, atol=1E-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f"""Saving PyTorch model and image processor to {pytorch_dump_folder_path}...""" ) Path(lowerCAmelCase__ ).mkdir(exist_ok=lowerCAmelCase__ ) model.save_pretrained(lowerCAmelCase__ ) image_processor.save_pretrained(lowerCAmelCase__ ) if push_to_hub: # Push model to HF hub logger.info("Pushing model to the hub..." ) UpperCAmelCase_ : Tuple = ( """microsoft/table-transformer-detection""" if """detection""" in checkpoint_url else """microsoft/table-transformer-structure-recognition""" ) model.push_to_hub(lowerCAmelCase__ ) image_processor.push_to_hub(lowerCAmelCase__ ) if __name__ == "__main__": _a = argparse.ArgumentParser() parser.add_argument( '--checkpoint_url', default='https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth', type=str, choices=[ 'https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth', 'https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth', ], help='URL of the Table Transformer checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether or not to push the converted model to the 🤗 hub.' ) _a = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
366
"""simple docstring""" from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean _a = 0 _a = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] _a = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right _a = tuple[int, int] class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ): """simple docstring""" UpperCAmelCase_ : int = pos_x UpperCAmelCase_ : List[Any] = pos_y UpperCAmelCase_ : Union[str, Any] = (pos_y, pos_x) UpperCAmelCase_ : Any = goal_x UpperCAmelCase_ : Dict = goal_y UpperCAmelCase_ : Any = g_cost UpperCAmelCase_ : List[str] = parent UpperCAmelCase_ : int = self.calculate_heuristic() UpperCAmelCase_ : Any = self.g_cost + self.h_cost def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.pos_x - self.goal_x UpperCAmelCase_ : Union[str, Any] = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowercase_ ) + abs(lowercase_ ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self , lowercase_ ): """simple docstring""" return self.f_cost < other.f_cost class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Tuple = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , lowercase_ ) UpperCAmelCase_ : List[Any] = Node(goal[1] , goal[0] , goal[1] , goal[0] , 9_9999 , lowercase_ ) UpperCAmelCase_ : str = [self.start] UpperCAmelCase_ : list[Node] = [] UpperCAmelCase_ : int = False def UpperCamelCase__ ( self ): """simple docstring""" while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() UpperCAmelCase_ : List[str] = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(lowercase_ ) self.closed_nodes.append(lowercase_ ) UpperCAmelCase_ : str = self.get_successors(lowercase_ ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowercase_ ) else: # retrieve the best current path UpperCAmelCase_ : Union[str, Any] = self.open_nodes.pop(self.open_nodes.index(lowercase_ ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowercase_ ) else: self.open_nodes.append(lowercase_ ) return [self.start.pos] def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Any = [] for action in delta: UpperCAmelCase_ : str = parent.pos_x + action[1] UpperCAmelCase_ : int = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowercase_ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowercase_ , lowercase_ , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , lowercase_ , ) ) return successors def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = node UpperCAmelCase_ : int = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) UpperCAmelCase_ : Optional[int] = current_node.parent path.reverse() return path class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Dict = AStar(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = AStar(lowercase_ , lowercase_ ) UpperCAmelCase_ : Tuple = False def UpperCamelCase__ ( self ): """simple docstring""" while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() UpperCAmelCase_ : List[str] = self.fwd_astar.open_nodes.pop(0 ) UpperCAmelCase_ : List[Any] = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowercase_ , lowercase_ ) self.fwd_astar.closed_nodes.append(lowercase_ ) self.bwd_astar.closed_nodes.append(lowercase_ ) UpperCAmelCase_ : Tuple = current_bwd_node UpperCAmelCase_ : str = current_fwd_node UpperCAmelCase_ : Dict = { self.fwd_astar: self.fwd_astar.get_successors(lowercase_ ), self.bwd_astar: self.bwd_astar.get_successors(lowercase_ ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowercase_ ) else: # retrieve the best current path UpperCAmelCase_ : List[Any] = astar.open_nodes.pop( astar.open_nodes.index(lowercase_ ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowercase_ ) else: astar.open_nodes.append(lowercase_ ) return [self.fwd_astar.start.pos] def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.fwd_astar.retrace_path(lowercase_ ) UpperCAmelCase_ : int = self.bwd_astar.retrace_path(lowercase_ ) bwd_path.pop() bwd_path.reverse() UpperCAmelCase_ : Any = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] _a = (0, 0) _a = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) _a = time.time() _a = AStar(init, goal) _a = a_star.search() _a = time.time() - start_time print(f"""AStar execution time = {end_time:f} seconds""") _a = time.time() _a = BidirectionalAStar(init, goal) _a = time.time() - bd_start_time print(f"""BidirectionalAStar execution time = {bd_end_time:f} seconds""")
23
0
"""simple docstring""" from pathlib import Path import fire from tqdm import tqdm def __a ( __lowerCamelCase="ro", __lowerCamelCase="en", __lowerCamelCase="wmt16", __lowerCamelCase=None ): try: import datasets except (ModuleNotFoundError, ImportError): raise ImportError("run pip install datasets" ) UpperCAmelCase_ : Union[str, Any] = f"""{src_lang}-{tgt_lang}""" print(f"""Converting {dataset}-{pair}""" ) UpperCAmelCase_ : Optional[int] = datasets.load_dataset(a_, a_ ) if save_dir is None: UpperCAmelCase_ : List[Any] = f"""{dataset}-{pair}""" UpperCAmelCase_ : Tuple = Path(a_ ) save_dir.mkdir(exist_ok=a_ ) for split in ds.keys(): print(f"""Splitting {split} with {ds[split].num_rows} records""" ) # to save to val.source, val.target like summary datasets UpperCAmelCase_ : Dict = "val" if split == "validation" else split UpperCAmelCase_ : Optional[Any] = save_dir.joinpath(f"""{fn}.source""" ) UpperCAmelCase_ : Any = save_dir.joinpath(f"""{fn}.target""" ) UpperCAmelCase_ : str = src_path.open("w+" ) UpperCAmelCase_ : int = tgt_path.open("w+" ) # reader is the bottleneck so writing one record at a time doesn't slow things down for x in tqdm(ds[split] ): UpperCAmelCase_ : Union[str, Any] = x["translation"] src_fp.write(ex[src_lang] + "\n" ) tgt_fp.write(ex[tgt_lang] + "\n" ) print(f"""Saved {dataset} dataset to {save_dir}""" ) if __name__ == "__main__": fire.Fire(download_wmt_dataset)
367
"""simple docstring""" import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = (PNDMScheduler,) SCREAMING_SNAKE_CASE__ : str = (("""num_inference_steps""", 50),) def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : int = { "num_train_timesteps": 1000, "beta_start": 0.00_01, "beta_end": 0.02, "beta_schedule": "linear", } config.update(**lowercase_ ) return config def UpperCamelCase__ ( self , lowercase_=0 , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = dict(self.forward_default_kwargs ) UpperCAmelCase_ : List[str] = kwargs.pop("num_inference_steps" , lowercase_ ) UpperCAmelCase_ : Union[str, Any] = self.dummy_sample UpperCAmelCase_ : Dict = 0.1 * sample UpperCAmelCase_ : Dict = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : List[Any] = self.get_scheduler_config(**lowercase_ ) UpperCAmelCase_ : Dict = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals UpperCAmelCase_ : List[Any] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) UpperCAmelCase_ : Optional[int] = scheduler_class.from_pretrained(lowercase_ ) new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals UpperCAmelCase_ : int = dummy_past_residuals[:] UpperCAmelCase_ : List[str] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : str = new_scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" UpperCAmelCase_ : Optional[int] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Dict = new_scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self , lowercase_=0 , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = dict(self.forward_default_kwargs ) UpperCAmelCase_ : str = kwargs.pop("num_inference_steps" , lowercase_ ) UpperCAmelCase_ : Optional[int] = self.dummy_sample UpperCAmelCase_ : List[str] = 0.1 * sample UpperCAmelCase_ : Tuple = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : str = self.get_scheduler_config() UpperCAmelCase_ : Dict = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals (must be after setting timesteps) UpperCAmelCase_ : List[Any] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) UpperCAmelCase_ : Dict = scheduler_class.from_pretrained(lowercase_ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residual (must be after setting timesteps) UpperCAmelCase_ : Optional[Any] = dummy_past_residuals[:] UpperCAmelCase_ : Union[str, Any] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Dict = new_scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" UpperCAmelCase_ : List[str] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : int = new_scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = self.scheduler_classes[0] UpperCAmelCase_ : Union[str, Any] = self.get_scheduler_config(**lowercase_ ) UpperCAmelCase_ : List[Any] = scheduler_class(**lowercase_ ) UpperCAmelCase_ : Tuple = 10 UpperCAmelCase_ : List[str] = self.dummy_model() UpperCAmelCase_ : str = self.dummy_sample_deter scheduler.set_timesteps(lowercase_ ) for i, t in enumerate(scheduler.prk_timesteps ): UpperCAmelCase_ : Tuple = model(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[int] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): UpperCAmelCase_ : Any = model(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ ).prev_sample return sample def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = dict(self.forward_default_kwargs ) UpperCAmelCase_ : Optional[Any] = kwargs.pop("num_inference_steps" , lowercase_ ) for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : Any = self.get_scheduler_config() UpperCAmelCase_ : Tuple = scheduler_class(**lowercase_ ) UpperCAmelCase_ : str = self.dummy_sample UpperCAmelCase_ : List[Any] = 0.1 * sample if num_inference_steps is not None and hasattr(lowercase_ , "set_timesteps" ): scheduler.set_timesteps(lowercase_ ) elif num_inference_steps is not None and not hasattr(lowercase_ , "set_timesteps" ): UpperCAmelCase_ : List[Any] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) UpperCAmelCase_ : List[str] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] UpperCAmelCase_ : List[str] = dummy_past_residuals[:] UpperCAmelCase_ : str = scheduler.step_prk(lowercase_ , 0 , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Any = scheduler.step_prk(lowercase_ , 1 , lowercase_ , **lowercase_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , 0 , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , 1 , lowercase_ , **lowercase_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def UpperCamelCase__ ( self ): """simple docstring""" for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for steps_offset in [0, 1]: self.check_over_configs(steps_offset=lowercase_ ) UpperCAmelCase_ : Optional[int] = self.scheduler_classes[0] UpperCAmelCase_ : int = self.get_scheduler_config(steps_offset=1 ) UpperCAmelCase_ : Optional[Any] = scheduler_class(**lowercase_ ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [901, 851, 851, 801, 801, 751, 751, 701, 701, 651, 651, 601, 601, 501, 401, 301, 201, 101, 1] ) , ) def UpperCamelCase__ ( self ): """simple docstring""" for beta_start, beta_end in zip([0.00_01, 0.0_01] , [0.0_02, 0.02] ): self.check_over_configs(beta_start=lowercase_ , beta_end=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for t in [1, 5, 10]: self.check_over_forward(time_step=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" # earlier version of set_timesteps() caused an error indexing alpha's with inference steps as power of 3 UpperCAmelCase_ : List[Any] = 27 for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : List[Any] = self.dummy_sample UpperCAmelCase_ : Optional[int] = 0.1 * sample UpperCAmelCase_ : List[str] = self.get_scheduler_config() UpperCAmelCase_ : List[str] = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): UpperCAmelCase_ : List[str] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ ).prev_sample def UpperCamelCase__ ( self ): """simple docstring""" with self.assertRaises(lowercase_ ): UpperCAmelCase_ : List[str] = self.scheduler_classes[0] UpperCAmelCase_ : str = self.get_scheduler_config() UpperCAmelCase_ : Tuple = scheduler_class(**lowercase_ ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.full_loop() UpperCAmelCase_ : Any = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Dict = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 1_98.13_18 ) < 1E-2 assert abs(result_mean.item() - 0.25_80 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = self.full_loop(prediction_type="v_prediction" ) UpperCAmelCase_ : str = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Tuple = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 67.39_86 ) < 1E-2 assert abs(result_mean.item() - 0.08_78 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase_ : Union[str, Any] = self.full_loop(set_alpha_to_one=lowercase_ , beta_start=0.01 ) UpperCAmelCase_ : List[Any] = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : int = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 2_30.03_99 ) < 1E-2 assert abs(result_mean.item() - 0.29_95 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase_ : Tuple = self.full_loop(set_alpha_to_one=lowercase_ , beta_start=0.01 ) UpperCAmelCase_ : int = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Tuple = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 1_86.94_82 ) < 1E-2 assert abs(result_mean.item() - 0.24_34 ) < 1E-3
23
0
import argparse import re from flax.traverse_util import flatten_dict, unflatten_dict from tax import checkpoints from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model from transformers.utils import logging logging.set_verbosity_info() # should not include what is already done by the `from_pt` argument _a = { '/attention/': '/0/SelfAttention/', '/self_attention/': '/0/SelfAttention/', '/encoder_decoder_attention/': '/1/EncDecAttention/', 'value': 'v', 'query': 'q', 'key': 'k', 'out': 'o', 'pre_self_attention_layer_norm': '0/layer_norm', 'pre_cross_attention_layer_norm': '1/layer_norm', 'pre_attention_layer_norm': '0/layer_norm', # previously 1, but seems wrong 'token_embedder': 'shared', 'encoder_norm': 'final_layer_norm', 'decoder_norm': 'final_layer_norm', 'relpos_bias/rel_embedding': 'block/0/layer/0/SelfAttention/relative_attention_bias/weight', 'router/router_weights/w/': 'router/classifier/', 'roer/roer_weights/w/': 'router/classifier/', 'logits_dense': 'lm_head', } def __a ( __lowerCamelCase ): # 1. in HF T5, we have block.{x}.layer.{y}. which corresponds to layer.{x} in # the original model UpperCAmelCase_ : Any = list(s_dict.keys() ) for key in keys: UpperCAmelCase_ : Tuple = r".*/layers_(\d+)" UpperCAmelCase_ : Union[str, Any] = key if re.match(_UpperCamelCase, _UpperCamelCase ): UpperCAmelCase_ : Union[str, Any] = re.sub(r"layers_(\d+)", r"block/\1/layer", _UpperCamelCase ) UpperCAmelCase_ : int = r"(encoder|decoder)\/" if re.match(_UpperCamelCase, _UpperCamelCase ): UpperCAmelCase_ : Dict = re.match(_UpperCamelCase, _UpperCamelCase ).groups() if groups[0] == "encoder": UpperCAmelCase_ : Tuple = re.sub(r"/mlp/", r"/1/mlp/", _UpperCamelCase ) UpperCAmelCase_ : Dict = re.sub(r"/pre_mlp_layer_norm/", r"/1/layer_norm/", _UpperCamelCase ) elif groups[0] == "decoder": UpperCAmelCase_ : Dict = re.sub(r"/mlp/", r"/2/mlp/", _UpperCamelCase ) UpperCAmelCase_ : str = re.sub(r"/pre_mlp_layer_norm/", r"/2/layer_norm/", _UpperCamelCase ) # 2. Convert other classic mappings for old_key, temp_key in MOE_LAYER_NAME_MAPPING.items(): if old_key in new_key: UpperCAmelCase_ : Optional[int] = new_key.replace(_UpperCamelCase, _UpperCamelCase ) print(f"""{key} -> {new_key}""" ) UpperCAmelCase_ : Optional[Any] = s_dict.pop(_UpperCamelCase ) if "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: UpperCAmelCase_ : Optional[int] = s_dict[ "encoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" ].T if "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" in s_dict: UpperCAmelCase_ : Optional[int] = s_dict[ "decoder/block/0/layer/0/SelfAttention/relative_attention_bias/weight" ].T # 3. Take extra care of the EXPERTS layer for key in list(s_dict.keys() ): if "expert" in key: UpperCAmelCase_ : str = s_dict[key].shape[0] UpperCAmelCase_ : Optional[int] = s_dict[key] for idx in range(_UpperCamelCase ): UpperCAmelCase_ : Any = expert_weihts[idx] print(f"""{key} -> {key.replace("expert/", "nested fstring" )}""" ) s_dict.pop(_UpperCamelCase ) return s_dict _a = { 'NUM_ENCODER_LAYERS': 'num_layers', 'NUM_DECODER_LAYERS': 'num_decoder_layers', 'NUM_HEADS': 'num_heads', 'HEAD_DIM': 'd_kv', 'EMBED_DIM': 'd_model', 'MLP_DIM': 'd_ff', 'NUM_SELECTED_EXPERTS': 'num_selected_experts', 'NUM_ENCODER_SPARSE_LAYERS': 'num_sparse_encoder_layers', 'NUM_DECODER_SPARSE_LAYERS': 'num_sparse_decoder_layers', 'dense.MlpBlock.activations': 'feed_forward_proj', } def __a ( __lowerCamelCase, __lowerCamelCase ): # Convert a google style config to the hugging face fromat import regex as re with open(_UpperCamelCase, "r" ) as f: UpperCAmelCase_ : Dict = f.read() UpperCAmelCase_ : Dict = re.findall(r"(.*) = ([0-9.]*)", _UpperCamelCase ) UpperCAmelCase_ : Union[str, Any] = {} for param, value in regex_match: if param in GIN_TO_CONFIG_MAPPING and value != "": UpperCAmelCase_ : List[Any] = float(_UpperCamelCase ) if "." in value else int(_UpperCamelCase ) UpperCAmelCase_ : Any = re.findall(r"(.*activations) = \(\'(.*)\',\)", _UpperCamelCase )[0] UpperCAmelCase_ : str = str(activation[1] ) UpperCAmelCase_ : str = num_experts UpperCAmelCase_ : Any = SwitchTransformersConfig(**_UpperCamelCase ) return config def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=None, __lowerCamelCase="./", __lowerCamelCase=8 ): # Initialise PyTorch model print(f"""Loading flax weights from : {flax_checkpoint_path}""" ) UpperCAmelCase_ : List[Any] = checkpoints.load_tax_checkpoint(_UpperCamelCase ) if gin_file is not None: UpperCAmelCase_ : Dict = convert_gin_to_config(_UpperCamelCase, _UpperCamelCase ) else: UpperCAmelCase_ : int = SwitchTransformersConfig.from_pretrained(_UpperCamelCase ) UpperCAmelCase_ : str = SwitchTransformersForConditionalGeneration(_UpperCamelCase ) UpperCAmelCase_ : List[Any] = flax_params["target"] UpperCAmelCase_ : Optional[Any] = flatten_dict(_UpperCamelCase, sep="/" ) UpperCAmelCase_ : Dict = rename_keys(_UpperCamelCase ) UpperCAmelCase_ : Any = unflatten_dict(_UpperCamelCase, sep="/" ) # Load the flax params in the PT model load_flax_weights_in_pytorch_model(_UpperCamelCase, _UpperCamelCase ) print(f"""Save PyTorch model to {pytorch_dump_path}""" ) pt_model.save_pretrained(_UpperCamelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--switch_t5x_checkpoint_path', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained SwitchTransformers model. \nThis specifies the' ' model architecture. If not provided, a `gin_file` has to be provided.' ), ) parser.add_argument( '--gin_file', default=None, type=str, required=False, help='Path to the gin config file. If not provided, a `config_file` has to be passed ', ) parser.add_argument( '--config_name', default=None, type=str, required=False, help='Config name of SwitchTransformers model.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output pytorch model.' ) parser.add_argument('--num_experts', default=8, type=int, required=False, help='Number of experts') _a = parser.parse_args() convert_flax_checkpoint_to_pytorch( args.switch_tax_checkpoint_path, args.config_name, args.gin_file, args.pytorch_dump_folder_path, args.num_experts, )
368
"""simple docstring""" import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels _a = object() # For specifying empty leaf dict `{}` _a = object() def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Any = tuple((re.compile(x + "$" ) for x in qs) ) for i in range(len(__lowerCamelCase ) - len(__lowerCamelCase ) + 1 ): UpperCAmelCase_ : List[str] = [x.match(__lowerCamelCase ) for x, y in zip(__lowerCamelCase, ks[i:] )] if matches and all(__lowerCamelCase ): return True return False def __a ( __lowerCamelCase ): def replace(__lowerCamelCase, __lowerCamelCase ): for rule, replacement in rules: if _match(__lowerCamelCase, __lowerCamelCase ): return replacement return val return replace def __a ( ): return [ # embeddings (("transformer", "wpe", "embedding"), P("mp", __lowerCamelCase )), (("transformer", "wte", "embedding"), P("mp", __lowerCamelCase )), # atention (("attention", "(q_proj|k_proj|v_proj)", "kernel"), P(__lowerCamelCase, "mp" )), (("attention", "out_proj", "kernel"), P("mp", __lowerCamelCase )), (("attention", "out_proj", "bias"), None), # mlp (("mlp", "c_fc", "kernel"), P(__lowerCamelCase, "mp" )), (("mlp", "c_fc", "bias"), P("mp" )), (("mlp", "c_proj", "kernel"), P("mp", __lowerCamelCase )), (("mlp", "c_proj", "bias"), None), # layer norms ((r"ln_\d+", "bias"), None), ((r"\d+", r"ln_\d+", "scale"), None), (("ln_f", "bias"), None), (("ln_f", "scale"), None), ] def __a ( __lowerCamelCase ): UpperCAmelCase_ : List[str] = _get_partition_rules() UpperCAmelCase_ : Any = _replacement_rules(__lowerCamelCase ) UpperCAmelCase_ : Any = {k: _unmatched for k in flatten_dict(__lowerCamelCase )} UpperCAmelCase_ : Dict = {k: replace(__lowerCamelCase, __lowerCamelCase ) for k, v in initd.items()} assert _unmatched not in result.values(), "Incomplete partition spec." return freeze(unflatten_dict(__lowerCamelCase ) )
23
0
"""simple docstring""" _a = {} def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) -> Any: if late == 3 or absent == 2: return 0 # if we have no days left, and have not failed any other rules, # we have a prize string if days == 0: return 1 # No easy solution, so now we need to do the recursive calculation # First, check if the combination is already in the cache, and # if yes, return the stored value from there since we already # know the number of possible prize strings from this point on UpperCAmelCase_ : Any = (days, absent, late) if key in cache: return cache[key] # now we calculate the three possible ways that can unfold from # this point on, depending on our attendance today # 1) if we are late (but not absent), the "absent" counter stays as # it is, but the "late" counter increases by one UpperCAmelCase_ : int = _calculate(days - 1, a__, late + 1 ) # 2) if we are absent, the "absent" counter increases by 1, and the # "late" counter resets to 0 UpperCAmelCase_ : Dict = _calculate(days - 1, absent + 1, 0 ) # 3) if we are on time, this resets the "late" counter and keeps the # absent counter UpperCAmelCase_ : Dict = _calculate(days - 1, a__, 0 ) UpperCAmelCase_ : Any = state_late + state_absent + state_ontime UpperCAmelCase_ : Optional[Any] = prizestrings return prizestrings def __a ( __lowerCamelCase = 30 ) -> int: return _calculate(a__, absent=0, late=0 ) if __name__ == "__main__": print(solution())
369
"""simple docstring""" import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow _a = logging.getLogger() @unittest.skip("""Temporarily disable the doc tests.""" ) @require_torch @require_tf @slow class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = True , ): """simple docstring""" UpperCAmelCase_ : List[str] = [file for file in os.listdir(lowercase_ ) if os.path.isfile(os.path.join(lowercase_ , lowercase_ ) )] if identifier is not None: UpperCAmelCase_ : Dict = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(lowercase_ , lowercase_ ): for n_ in n_identifier: UpperCAmelCase_ : str = [file for file in files if n_ not in file] else: UpperCAmelCase_ : Any = [file for file in files if n_identifier not in file] UpperCAmelCase_ : Union[str, Any] = ignore_files or [] ignore_files.append("__init__.py" ) UpperCAmelCase_ : Optional[int] = [file for file in files if file not in ignore_files] for file in files: # Open all files print("Testing" , lowercase_ ) if only_modules: UpperCAmelCase_ : str = file.split("." )[0] try: UpperCAmelCase_ : str = getattr(lowercase_ , lowercase_ ) UpperCAmelCase_ : Tuple = doctest.DocTestSuite(lowercase_ ) UpperCAmelCase_ : int = unittest.TextTestRunner().run(lowercase_ ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: UpperCAmelCase_ : Optional[Any] = doctest.testfile(str(".." / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = Path("src/transformers" ) UpperCAmelCase_ : str = "modeling" UpperCAmelCase_ : Optional[Any] = [ "modeling_ctrl.py", "modeling_tf_ctrl.py", ] self.analyze_directory(lowercase_ , identifier=lowercase_ , ignore_files=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = Path("src/transformers" ) UpperCAmelCase_ : Any = "tokenization" self.analyze_directory(lowercase_ , identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = Path("src/transformers" ) UpperCAmelCase_ : List[Any] = "configuration" self.analyze_directory(lowercase_ , identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = Path("src/transformers" ) UpperCAmelCase_ : List[Any] = ["configuration", "modeling", "tokenization"] self.analyze_directory(lowercase_ , n_identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = Path("docs/source" ) UpperCAmelCase_ : Union[str, Any] = ["favicon.ico"] self.analyze_directory(lowercase_ , ignore_files=lowercase_ , only_modules=lowercase_ )
23
0
"""simple docstring""" def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): if index == r: for j in range(__SCREAMING_SNAKE_CASE ): print(data[j], end=" " ) print(" " ) return # When no more elements are there to put in data[] if i >= n: return # current is included, put next at next location UpperCAmelCase_ : List[Any] = arr[i] combination_util(__SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE, index + 1, __SCREAMING_SNAKE_CASE, i + 1 ) # current is excluded, replace it with # next (Note that i+1 is passed, but # index is not changed) combination_util(__SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE, i + 1 ) # The main function that prints all combinations # of size r in arr[] of size n. This function # mainly uses combinationUtil() def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : str = [0] * r # Print all combination using temporary array 'data[]' combination_util(__SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE, __SCREAMING_SNAKE_CASE, 0, __SCREAMING_SNAKE_CASE, 0 ) if __name__ == "__main__": # Driver code to check the function above _a = [10, 20, 30, 40, 50] print_combination(arr, len(arr), 3) # This code is contributed by Ambuj sahu
370
"""simple docstring""" import warnings from ...utils import is_sklearn_available, requires_backends if is_sklearn_available(): from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef _a = ( 'This metric will be removed from the library soon, metrics should be handled with the 🤗 Evaluate ' 'library. You can have a look at this example script for pointers: ' 'https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py' ) def __a ( __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) return (preds == labels).mean() def __a ( __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) UpperCAmelCase_ : Optional[Any] = simple_accuracy(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : List[Any] = fa_score(y_true=__lowerCamelCase, y_pred=__lowerCamelCase ) return { "acc": acc, "f1": fa, "acc_and_f1": (acc + fa) / 2, } def __a ( __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) UpperCAmelCase_ : Any = pearsonr(__lowerCamelCase, __lowerCamelCase )[0] UpperCAmelCase_ : Optional[Any] = spearmanr(__lowerCamelCase, __lowerCamelCase )[0] return { "pearson": pearson_corr, "spearmanr": spearman_corr, "corr": (pearson_corr + spearman_corr) / 2, } def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) assert len(__lowerCamelCase ) == len(__lowerCamelCase ), f"""Predictions and labels have mismatched lengths {len(__lowerCamelCase )} and {len(__lowerCamelCase )}""" if task_name == "cola": return {"mcc": matthews_corrcoef(__lowerCamelCase, __lowerCamelCase )} elif task_name == "sst-2": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "mrpc": return acc_and_fa(__lowerCamelCase, __lowerCamelCase ) elif task_name == "sts-b": return pearson_and_spearman(__lowerCamelCase, __lowerCamelCase ) elif task_name == "qqp": return acc_and_fa(__lowerCamelCase, __lowerCamelCase ) elif task_name == "mnli": return {"mnli/acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "mnli-mm": return {"mnli-mm/acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "qnli": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "rte": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "wnli": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "hans": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} else: raise KeyError(__lowerCamelCase ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) if len(__lowerCamelCase ) != len(__lowerCamelCase ): raise ValueError(f"""Predictions and labels have mismatched lengths {len(__lowerCamelCase )} and {len(__lowerCamelCase )}""" ) if task_name == "xnli": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} else: raise KeyError(__lowerCamelCase )
23
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.xglm.modeling_tf_xglm import ( TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFXGLMForCausalLM, TFXGLMModel, ) @require_tf class A_ : '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = XGLMConfig SCREAMING_SNAKE_CASE__ : List[str] = {} SCREAMING_SNAKE_CASE__ : str = """gelu""" def __init__( self , lowercase_ , lowercase_=14 , lowercase_=7 , lowercase_=True , lowercase_=True , lowercase_=True , lowercase_=99 , lowercase_=32 , lowercase_=2 , lowercase_=4 , lowercase_=37 , lowercase_="gelu" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=512 , lowercase_=0.02 , ): """simple docstring""" UpperCAmelCase_ : List[Any] = parent UpperCAmelCase_ : Optional[Any] = batch_size UpperCAmelCase_ : Dict = seq_length UpperCAmelCase_ : int = is_training UpperCAmelCase_ : Dict = use_input_mask UpperCAmelCase_ : Optional[int] = use_labels UpperCAmelCase_ : str = vocab_size UpperCAmelCase_ : int = d_model UpperCAmelCase_ : Any = num_hidden_layers UpperCAmelCase_ : Tuple = num_attention_heads UpperCAmelCase_ : Optional[Any] = ffn_dim UpperCAmelCase_ : List[str] = activation_function UpperCAmelCase_ : Any = activation_dropout UpperCAmelCase_ : List[str] = attention_dropout UpperCAmelCase_ : int = max_position_embeddings UpperCAmelCase_ : Optional[int] = initializer_range UpperCAmelCase_ : int = None UpperCAmelCase_ : Any = 0 UpperCAmelCase_ : Union[str, Any] = 2 UpperCAmelCase_ : Optional[int] = 1 def UpperCamelCase__ ( self ): """simple docstring""" return XGLMConfig.from_pretrained("facebook/xglm-564M" ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = tf.clip_by_value( ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) , clip_value_min=0 , clip_value_max=3 ) UpperCAmelCase_ : Tuple = None if self.use_input_mask: UpperCAmelCase_ : Optional[int] = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase_ : str = self.get_config() UpperCAmelCase_ : Union[str, Any] = floats_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, input_mask, head_mask, ) def UpperCamelCase__ ( self ): """simple docstring""" return XGLMConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , num_layers=self.num_hidden_layers , attention_heads=self.num_attention_heads , ffn_dim=self.ffn_dim , activation_function=self.activation_function , activation_dropout=self.activation_dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , use_cache=lowercase_ , bos_token_id=self.bos_token_id , eos_token_id=self.eos_token_id , pad_token_id=self.pad_token_id , return_dict=lowercase_ , ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = self.prepare_config_and_inputs() ( UpperCAmelCase_ ) : List[str] = config_and_inputs UpperCAmelCase_ : str = { """input_ids""": input_ids, """head_mask""": head_mask, } return config, inputs_dict @require_tf class A_ (_UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = (TFXGLMModel, TFXGLMForCausalLM) if is_tf_available() else () SCREAMING_SNAKE_CASE__ : List[Any] = (TFXGLMForCausalLM,) if is_tf_available() else () SCREAMING_SNAKE_CASE__ : Tuple = ( {"""feature-extraction""": TFXGLMModel, """text-generation""": TFXGLMForCausalLM} if is_tf_available() else {} ) SCREAMING_SNAKE_CASE__ : List[str] = False SCREAMING_SNAKE_CASE__ : Optional[int] = False SCREAMING_SNAKE_CASE__ : Union[str, Any] = False def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = TFXGLMModelTester(self ) UpperCAmelCase_ : List[str] = ConfigTester(self , config_class=lowercase_ , n_embd=37 ) def UpperCamelCase__ ( self ): """simple docstring""" self.config_tester.run_common_tests() @slow def UpperCamelCase__ ( self ): """simple docstring""" for model_name in TF_XGLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase_ : Optional[int] = TFXGLMModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) @unittest.skip(reason="Currently, model embeddings are going to undergo a major refactor." ) def UpperCamelCase__ ( self ): """simple docstring""" super().test_resize_token_embeddings() @require_tf class A_ (unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase__ ( self , lowercase_=True ): """simple docstring""" UpperCAmelCase_ : Optional[int] = TFXGLMForCausalLM.from_pretrained("facebook/xglm-564M" ) UpperCAmelCase_ : Optional[Any] = tf.convert_to_tensor([[2, 268, 9865]] , dtype=tf.intaa ) # The dog # </s> The dog is a very friendly dog. He is very affectionate and loves to play with other # fmt: off UpperCAmelCase_ : Optional[Any] = [2, 268, 9865, 67, 11, 1988, 5_7252, 9865, 5, 984, 67, 1988, 21_3838, 1658, 53, 7_0446, 33, 6657, 278, 1581] # fmt: on UpperCAmelCase_ : List[str] = model.generate(lowercase_ , do_sample=lowercase_ , num_beams=1 ) if verify_outputs: self.assertListEqual(output_ids[0].numpy().tolist() , lowercase_ ) @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = XGLMTokenizer.from_pretrained("facebook/xglm-564M" ) UpperCAmelCase_ : Any = TFXGLMForCausalLM.from_pretrained("facebook/xglm-564M" ) tf.random.set_seed(0 ) UpperCAmelCase_ : Union[str, Any] = tokenizer("Today is a nice day and" , return_tensors="tf" ) UpperCAmelCase_ : Optional[Any] = tokenized.input_ids # forces the generation to happen on CPU, to avoid GPU-related quirks (and assure same output regardless of the available devices) with tf.device(":/CPU:0" ): UpperCAmelCase_ : Any = model.generate(lowercase_ , do_sample=lowercase_ , seed=[7, 0] ) UpperCAmelCase_ : List[Any] = tokenizer.decode(output_ids[0] , skip_special_tokens=lowercase_ ) UpperCAmelCase_ : Any = ( """Today is a nice day and warm evening here over Southern Alberta!! Today when they closed schools due""" ) self.assertEqual(lowercase_ , lowercase_ ) @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = TFXGLMForCausalLM.from_pretrained("facebook/xglm-564M" ) UpperCAmelCase_ : Tuple = XGLMTokenizer.from_pretrained("facebook/xglm-564M" ) UpperCAmelCase_ : int = """left""" # use different length sentences to test batching UpperCAmelCase_ : str = [ """This is an extremelly long sentence that only exists to test the ability of the model to cope with """ """left-padding, such as in batched generation. The output for the sequence below should be the same """ """regardless of whether left padding is applied or not. When""", """Hello, my dog is a little""", ] UpperCAmelCase_ : Tuple = tokenizer(lowercase_ , return_tensors="tf" , padding=lowercase_ ) UpperCAmelCase_ : Dict = inputs["""input_ids"""] UpperCAmelCase_ : Any = model.generate(input_ids=lowercase_ , attention_mask=inputs["attention_mask"] , max_new_tokens=12 ) UpperCAmelCase_ : Optional[Any] = tokenizer(sentences[0] , return_tensors="tf" ).input_ids UpperCAmelCase_ : Union[str, Any] = model.generate(input_ids=lowercase_ , max_new_tokens=12 ) UpperCAmelCase_ : Dict = tokenizer(sentences[1] , return_tensors="tf" ).input_ids UpperCAmelCase_ : Optional[Any] = model.generate(input_ids=lowercase_ , max_new_tokens=12 ) UpperCAmelCase_ : Optional[Any] = tokenizer.batch_decode(lowercase_ , skip_special_tokens=lowercase_ ) UpperCAmelCase_ : Optional[Any] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowercase_ ) UpperCAmelCase_ : List[str] = tokenizer.decode(output_padded[0] , skip_special_tokens=lowercase_ ) UpperCAmelCase_ : Dict = [ """This is an extremelly long sentence that only exists to test the ability of the model to cope with """ """left-padding, such as in batched generation. The output for the sequence below should be the same """ """regardless of whether left padding is applied or not. When left padding is applied, the sequence will be """ """a single""", """Hello, my dog is a little bit of a shy one, but he is very friendly""", ] self.assertListEqual(lowercase_ , lowercase_ ) self.assertListEqual(lowercase_ , [non_padded_sentence, padded_sentence] )
371
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a = logging.get_logger(__name__) _a = {'vocab_file': 'vocab.json'} _a = { 'vocab_file': { 'mgp-str': 'https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json', } } _a = {'mgp-str': 27} class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : List[str] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , lowercase_ , lowercase_="[GO]" , lowercase_="[GO]" , lowercase_="[s]" , lowercase_="[GO]" , **lowercase_ ): """simple docstring""" super().__init__( unk_token=lowercase_ , bos_token=lowercase_ , eos_token=lowercase_ , pad_token=lowercase_ , **lowercase_ , ) with open(lowercase_ , encoding="utf-8" ) as vocab_handle: UpperCAmelCase_ : Dict = json.load(lowercase_ ) UpperCAmelCase_ : Dict = {v: k for k, v in self.vocab.items()} @property def UpperCamelCase__ ( self ): """simple docstring""" return len(self.vocab ) def UpperCamelCase__ ( self ): """simple docstring""" return dict(self.vocab , **self.added_tokens_encoder ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = [] for s in text: char_tokens.extend(lowercase_ ) return char_tokens def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.vocab.get(lowercase_ , self.vocab.get(self.unk_token ) ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.decoder.get(lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if not os.path.isdir(lowercase_ ): logger.error("Vocabulary path ({}) should be a directory".format(lowercase_ ) ) return UpperCAmelCase_ : Optional[int] = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) with open(lowercase_ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=lowercase_ , ensure_ascii=lowercase_ ) + "\n" ) return (vocab_file,)
23
0
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a = logging.get_logger(__name__) _a = {'vocab_file': 'sentencepiece.model'} _a = { 'vocab_file': { 'google/rembert': 'https://huggingface.co/google/rembert/resolve/main/sentencepiece.model', }, } _a = { 'google/rembert': 256, } class A_ (__lowercase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : Tuple = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , lowercase_ , lowercase_=False , lowercase_=True , lowercase_=True , lowercase_="[CLS]" , lowercase_="[SEP]" , lowercase_="[UNK]" , lowercase_="[SEP]" , lowercase_="[PAD]" , lowercase_="[CLS]" , lowercase_="[MASK]" , **lowercase_ , ): """simple docstring""" super().__init__( do_lower_case=_a , remove_space=_a , keep_accents=_a , bos_token=_a , eos_token=_a , unk_token=_a , sep_token=_a , pad_token=_a , cls_token=_a , mask_token=_a , **_a , ) UpperCAmelCase_ : str = do_lower_case UpperCAmelCase_ : List[str] = remove_space UpperCAmelCase_ : Optional[int] = keep_accents UpperCAmelCase_ : Optional[Any] = vocab_file UpperCAmelCase_ : Optional[int] = spm.SentencePieceProcessor() self.sp_model.Load(_a ) @property def UpperCamelCase__ ( self ): """simple docstring""" return len(self.sp_model ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = {self.convert_ids_to_tokens(_a ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): """simple docstring""" UpperCAmelCase_ : int = self.__dict__.copy() UpperCAmelCase_ : List[str] = None return state def __setstate__( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = d UpperCAmelCase_ : List[Any] = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def UpperCamelCase__ ( self , lowercase_ , lowercase_=False ): """simple docstring""" UpperCAmelCase_ : Any = self.sp_model.EncodeAsPieces(_a ) return pieces def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.sp_model.PieceToId(_a ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.sp_model.IdToPiece(_a ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.sp_model.decode_pieces(_a ) return out_string def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" UpperCAmelCase_ : Optional[int] = [self.sep_token_id] UpperCAmelCase_ : Any = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = False ): """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(_a )) + [1] + ([0] * len(_a )) + [1] return [1] + ([0] * len(_a )) + [1] def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" UpperCAmelCase_ : List[str] = [self.sep_token_id] UpperCAmelCase_ : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if not os.path.isdir(_a ): logger.error("Vocabulary path ({}) should be a directory".format(_a ) ) return UpperCAmelCase_ : Optional[Any] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_a ): copyfile(self.vocab_file , _a ) return (out_vocab_file,)
350
"""simple docstring""" import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency _a = { 'E': 12.70, 'T': 9.06, 'A': 8.17, 'O': 7.51, 'I': 6.97, 'N': 6.75, 'S': 6.33, 'H': 6.09, 'R': 5.99, 'D': 4.25, 'L': 4.03, 'C': 2.78, 'U': 2.76, 'M': 2.41, 'W': 2.36, 'F': 2.23, 'G': 2.02, 'Y': 1.97, 'P': 1.93, 'B': 1.29, 'V': 0.98, 'K': 0.77, 'J': 0.15, 'X': 0.15, 'Q': 0.10, 'Z': 0.07, } _a = 'ETAOINSHRDLCUMWFGYPBVKJXQZ' _a = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def __a ( __lowerCamelCase ): UpperCAmelCase_ : Any = {letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def __a ( __lowerCamelCase ): return x[0] def __a ( __lowerCamelCase ): UpperCAmelCase_ : Any = get_letter_count(__lowerCamelCase ) UpperCAmelCase_ : dict[int, list[str]] = { freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(__lowerCamelCase ) UpperCAmelCase_ : dict[int, str] = {} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find, reverse=__lowerCamelCase ) UpperCAmelCase_ : Any = "".join(freq_to_letter[freq] ) UpperCAmelCase_ : str = list(freq_to_letter_str.items() ) freq_pairs.sort(key=__lowerCamelCase, reverse=__lowerCamelCase ) UpperCAmelCase_ : list[str] = [freq_pair[1] for freq_pair in freq_pairs] return "".join(__lowerCamelCase ) def __a ( __lowerCamelCase ): UpperCAmelCase_ : Any = get_frequency_order(__lowerCamelCase ) UpperCAmelCase_ : int = 0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
23
0
"""simple docstring""" import random import timeit from functools import wraps from typing import Callable, Optional from ..configuration_utils import PretrainedConfig from ..models.auto.modeling_tf_auto import TF_MODEL_MAPPING, TF_MODEL_WITH_LM_HEAD_MAPPING from ..utils import is_pyanvml_available, is_tf_available, logging from .benchmark_utils import ( Benchmark, Memory, MemorySummary, measure_peak_memory_cpu, start_memory_tracing, stop_memory_tracing, ) if is_tf_available(): import tensorflow as tf from tensorflow.python.framework.errors_impl import ResourceExhaustedError from .benchmark_args_tf import TensorFlowBenchmarkArguments if is_pyanvml_available(): import pyanvml.pyanvml as nvml _a = logging.get_logger(__name__) def __a ( __lowerCamelCase, __lowerCamelCase ): def run_func(__lowerCamelCase ): @wraps(_snake_case ) def run_in_eager_mode(*__lowerCamelCase, **__lowerCamelCase ): return func(*_snake_case, **_snake_case ) @wraps(_snake_case ) @tf.function(experimental_compile=_snake_case ) def run_in_graph_mode(*__lowerCamelCase, **__lowerCamelCase ): return func(*_snake_case, **_snake_case ) if do_eager_mode is True: if use_xla is not False: raise ValueError( "Cannot run model in XLA, if `args.eager_mode` is set to `True`. Please set `args.eager_mode=False`." ) return run_in_eager_mode else: return run_in_graph_mode return run_func def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Any = random.Random() UpperCAmelCase_ : Optional[Any] = [rng.randint(0, vocab_size - 1 ) for i in range(batch_size * sequence_length )] return tf.constant(_snake_case, shape=(batch_size, sequence_length), dtype=tf.intaa ) class A_ (__snake_case ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = 42 SCREAMING_SNAKE_CASE__ = 42 SCREAMING_SNAKE_CASE__ = """TensorFlow""" @property def UpperCamelCase__ ( self ): """simple docstring""" return tf.__version__ def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[str] = self.args.strategy if strategy is None: raise ValueError("A device strategy has to be initialized before using TensorFlow." ) UpperCAmelCase_ : List[str] = self._prepare_inference_func(a_ , a_ , a_ ) return self._measure_speed(_inference ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.args.strategy if strategy is None: raise ValueError("A device strategy has to be initialized before using TensorFlow." ) UpperCAmelCase_ : Optional[int] = self._prepare_train_func(a_ , a_ , a_ ) return self._measure_speed(_train ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" if self.args.is_gpu: tf.config.experimental.set_memory_growth(self.args.gpu_list[self.args.device_idx] , a_ ) UpperCAmelCase_ : List[str] = self.args.strategy if strategy is None: raise ValueError("A device strategy has to be initialized before using TensorFlow." ) UpperCAmelCase_ : Tuple = self._prepare_inference_func(a_ , a_ , a_ ) return self._measure_memory(_inference ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" if self.args.is_gpu: tf.config.experimental.set_memory_growth(self.args.gpu_list[self.args.device_idx] , a_ ) UpperCAmelCase_ : str = self.args.strategy if strategy is None: raise ValueError("A device strategy has to be initialized before using TensorFlow." ) UpperCAmelCase_ : str = self._prepare_train_func(a_ , a_ , a_ ) return self._measure_memory(_train ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : int = self.config_dict[model_name] if self.args.fpaa: raise NotImplementedError("Mixed precision is currently not supported." ) UpperCAmelCase_ : Dict = ( hasattr(a_ , "architectures" ) and isinstance(config.architectures , a_ ) and len(config.architectures ) > 0 ) if not self.args.only_pretrain_model and has_model_class_in_config: try: UpperCAmelCase_ : List[str] = '''TF''' + config.architectures[0] # prepend 'TF' for tensorflow model UpperCAmelCase_ : str = __import__("transformers" , fromlist=[model_class] ) UpperCAmelCase_ : Optional[Any] = getattr(a_ , a_ ) UpperCAmelCase_ : Dict = model_cls(a_ ) except ImportError: raise ImportError( F"""{model_class} does not exist. If you just want to test the pretrained model, you might want to""" " set `--only_pretrain_model` or `args.only_pretrain_model=True`." ) else: UpperCAmelCase_ : Optional[Any] = TF_MODEL_MAPPING[config.__class__](a_ ) # encoder-decoder has vocab size saved differently UpperCAmelCase_ : Any = config.vocab_size if hasattr(a_ , "vocab_size" ) else config.encoder.vocab_size UpperCAmelCase_ : Union[str, Any] = random_input_ids(a_ , a_ , a_ ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_decoder_forward(): return model(a_ , decoder_input_ids=a_ , training=a_ ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_forward(): return model(a_ , training=a_ ) UpperCAmelCase_ : List[Any] = encoder_decoder_forward if config.is_encoder_decoder else encoder_forward return _inference def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Tuple = self.config_dict[model_name] if self.args.eager_mode is not False: raise ValueError("Training cannot be done in eager mode. Please make sure that `args.eager_mode = False`." ) if self.args.fpaa: raise NotImplementedError("Mixed precision is currently not supported." ) UpperCAmelCase_ : Optional[Any] = ( hasattr(a_ , "architectures" ) and isinstance(config.architectures , a_ ) and len(config.architectures ) > 0 ) if not self.args.only_pretrain_model and has_model_class_in_config: try: UpperCAmelCase_ : Dict = '''TF''' + config.architectures[0] # prepend 'TF' for tensorflow model UpperCAmelCase_ : Any = __import__("transformers" , fromlist=[model_class] ) UpperCAmelCase_ : List[str] = getattr(a_ , a_ ) UpperCAmelCase_ : int = model_cls(a_ ) except ImportError: raise ImportError( F"""{model_class} does not exist. If you just want to test the pretrained model, you might want to""" " set `--only_pretrain_model` or `args.only_pretrain_model=True`." ) else: UpperCAmelCase_ : Any = TF_MODEL_WITH_LM_HEAD_MAPPING[config.__class__](a_ ) # encoder-decoder has vocab size saved differently UpperCAmelCase_ : Optional[Any] = config.vocab_size if hasattr(a_ , "vocab_size" ) else config.encoder.vocab_size UpperCAmelCase_ : Union[str, Any] = random_input_ids(a_ , a_ , a_ ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_decoder_train(): UpperCAmelCase_ : str = model(a_ , decoder_input_ids=a_ , labels=a_ , training=a_ )[0] UpperCAmelCase_ : Any = tf.gradients(a_ , model.trainable_variables ) return gradients @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_train(): UpperCAmelCase_ : Union[str, Any] = model(a_ , labels=a_ , training=a_ )[0] UpperCAmelCase_ : List[Any] = tf.gradients(a_ , model.trainable_variables ) return gradients UpperCAmelCase_ : Tuple = encoder_decoder_train if config.is_encoder_decoder else encoder_train return _train def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" with self.args.strategy.scope(): try: if self.args.is_tpu or self.args.use_xla: # run additional 10 times to stabilize compilation for tpu logger.info("Do inference on TPU. Running model 5 times to stabilize compilation" ) timeit.repeat(a_ , repeat=1 , number=5 ) # as written in https://docs.python.org/2/library/timeit.html#timeit.Timer.repeat, min should be taken rather than the average UpperCAmelCase_ : Union[str, Any] = timeit.repeat( a_ , repeat=self.args.repeat , number=10 , ) return min(a_ ) / 10.0 except ResourceExhaustedError as e: self.print_fn(F"""Doesn't fit on GPU. {e}""" ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" logger.info( "Note that TensorFlow allocates more memory than " "it might need to speed up computation. " "The memory reported here corresponds to the memory " "reported by `nvidia-smi`, which can vary depending " "on total available memory on the GPU that is used." ) with self.args.strategy.scope(): try: if self.args.trace_memory_line_by_line: if not self.args.eager_mode: raise ValueError( "`args.eager_mode` is set to `False`. Make sure to run model in eager mode to measure memory" " consumption line by line." ) UpperCAmelCase_ : List[Any] = start_memory_tracing("transformers" ) if self.args.is_tpu: # tpu raise NotImplementedError( "Memory Benchmarking is currently not implemented for TPU. Please disable memory benchmarking" " with `args.memory=False`" ) elif self.args.is_gpu: # gpu if not is_pyanvml_available(): logger.warning( "py3nvml not installed, we won\'t log GPU memory usage. " "Install py3nvml (pip install py3nvml) to log information about GPU." ) UpperCAmelCase_ : Any = '''N/A''' else: logger.info( "Measuring total GPU usage on GPU device. Make sure to not have additional processes" " running on the same GPU." ) # init nvml nvml.nvmlInit() func() UpperCAmelCase_ : int = nvml.nvmlDeviceGetHandleByIndex(self.args.device_idx ) UpperCAmelCase_ : Any = nvml.nvmlDeviceGetMemoryInfo(a_ ) UpperCAmelCase_ : List[str] = meminfo.used UpperCAmelCase_ : Any = Memory(a_ ) # shutdown nvml nvml.nvmlShutdown() else: # cpu if self.args.trace_memory_line_by_line: logger.info( "When enabling line by line tracing, the max peak memory for CPU is inaccurate in" " TensorFlow." ) UpperCAmelCase_ : int = None else: UpperCAmelCase_ : Optional[Any] = measure_peak_memory_cpu(a_ ) UpperCAmelCase_ : str = Memory(a_ ) if isinstance(a_ , a_ ) else memory_bytes if self.args.trace_memory_line_by_line: UpperCAmelCase_ : str = stop_memory_tracing(a_ ) if memory is None: UpperCAmelCase_ : str = summary.total else: UpperCAmelCase_ : List[str] = None return memory, summary except ResourceExhaustedError as e: self.print_fn(F"""Doesn't fit on GPU. {e}""" ) return "N/A", None
351
"""simple docstring""" import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) _a = logging.getLogger() def __a ( ): UpperCAmelCase_ : Tuple = argparse.ArgumentParser() parser.add_argument("-f" ) UpperCAmelCase_ : Dict = parser.parse_args() return args.f class A_ (lowercase__ ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = logging.StreamHandler(sys.stdout ) logger.addHandler(lowercase_ ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , "run_glue_deebert.py" ) with patch.object(lowercase_ , "argv" , lowercase_ ): UpperCAmelCase_ : List[str] = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(lowercase_ , 0.6_66 ) @slow @require_torch_non_multi_gpu def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = "\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n ".split() self.run_and_check(lowercase_ ) UpperCAmelCase_ : Optional[Any] = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowercase_ ) UpperCAmelCase_ : Dict = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowercase_ )
23
0
"""simple docstring""" import requests _a = """YOUR API KEY""" def __a ( __lowerCamelCase, __lowerCamelCase = giphy_api_key ): UpperCAmelCase_ : Optional[Any] = '+'.join(query.split() ) UpperCAmelCase_ : Any = f"""https://api.giphy.com/v1/gifs/search?q={formatted_query}&api_key={api_key}""" UpperCAmelCase_ : Dict = requests.get(_A ).json()['data'] return [gif["url"] for gif in gifs] if __name__ == "__main__": print('\n'.join(get_gifs('space ship')))
352
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _a = {'configuration_unispeech': ['UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP', 'UniSpeechConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST', 'UniSpeechForCTC', 'UniSpeechForPreTraining', 'UniSpeechForSequenceClassification', 'UniSpeechModel', 'UniSpeechPreTrainedModel', ] if TYPE_CHECKING: from .configuration_unispeech import UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP, UniSpeechConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_unispeech import ( UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST, UniSpeechForCTC, UniSpeechForPreTraining, UniSpeechForSequenceClassification, UniSpeechModel, UniSpeechPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
23
0
"""simple docstring""" import unittest from transformers import CamembertTokenizer, CamembertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import is_torch_available from ...test_tokenization_common import TokenizerTesterMixin _a = get_tests_dir('fixtures/test_sentencepiece.model') _a = get_tests_dir('fixtures/test_sentencepiece_bpe.model') _a = 'pt' if is_torch_available() else 'tf' @require_sentencepiece @require_tokenizers class A_ (lowercase__ ,unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Optional[int] = CamembertTokenizer SCREAMING_SNAKE_CASE__ : Optional[Any] = CamembertTokenizerFast SCREAMING_SNAKE_CASE__ : Optional[Any] = True SCREAMING_SNAKE_CASE__ : str = True def UpperCamelCase__ ( self ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing UpperCAmelCase_ : Optional[int] = CamembertTokenizer(__UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = "<pad>" UpperCAmelCase_ : Optional[int] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__UpperCAmelCase ) , __UpperCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__UpperCAmelCase ) , __UpperCAmelCase ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<s>NOTUSED" ) self.assertEqual(vocab_keys[1] , "<pad>" ) self.assertEqual(vocab_keys[-1] , "<mask>" ) self.assertEqual(len(__UpperCAmelCase ) , 1004 ) def UpperCamelCase__ ( self ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1005 ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[str] = CamembertTokenizer(__UpperCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) UpperCAmelCase_ : Tuple = CamembertTokenizerFast.from_pretrained(self.tmpdirname ) UpperCAmelCase_ : Tuple = "I was born in 92000, and this is falsé." UpperCAmelCase_ : Union[str, Any] = tokenizer.encode(__UpperCAmelCase ) UpperCAmelCase_ : Optional[int] = rust_tokenizer.encode(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase_ : List[str] = tokenizer.encode(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) UpperCAmelCase_ : Optional[Any] = rust_tokenizer.encode(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) # <unk> tokens are not the same for `rust` than for `slow`. # Because spm gives back raw token instead of `unk` in EncodeAsPieces # tokens = tokenizer.tokenize(sequence) UpperCAmelCase_ : Dict = tokenizer.convert_ids_to_tokens(__UpperCAmelCase ) UpperCAmelCase_ : Optional[int] = rust_tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) def UpperCamelCase__ ( self ): """simple docstring""" if not self.test_rust_tokenizer: return UpperCAmelCase_ : Optional[Any] = self.get_tokenizer() UpperCAmelCase_ : Union[str, Any] = self.get_rust_tokenizer() UpperCAmelCase_ : Any = "I was born in 92000, and this is falsé." UpperCAmelCase_ : Optional[int] = tokenizer.tokenize(__UpperCAmelCase ) UpperCAmelCase_ : Optional[Any] = rust_tokenizer.tokenize(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase_ : Optional[int] = tokenizer.encode(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) UpperCAmelCase_ : List[Any] = rust_tokenizer.encode(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase_ : Tuple = self.get_rust_tokenizer() UpperCAmelCase_ : Any = tokenizer.encode(__UpperCAmelCase ) UpperCAmelCase_ : str = rust_tokenizer.encode(__UpperCAmelCase ) self.assertListEqual(__UpperCAmelCase , __UpperCAmelCase ) @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = {"input_ids": [[5, 54, 7196, 297, 30, 23, 776, 18, 11, 3215, 3705, 8252, 22, 3164, 1181, 2116, 29, 16, 813, 25, 791, 3314, 20, 3446, 38, 2_7575, 120, 6, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [5, 468, 17, 11, 9088, 20, 1517, 8, 2_2804, 1_8818, 10, 38, 629, 607, 607, 142, 19, 7196, 867, 56, 1_0326, 24, 2267, 20, 416, 5072, 1_5612, 233, 734, 7, 2399, 27, 16, 3015, 1649, 7, 24, 20, 4338, 2399, 27, 13, 3400, 14, 13, 6189, 8, 930, 9, 6]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # camembert is a french model. So we also use french texts. UpperCAmelCase_ : Dict = [ "Le transformeur est un modèle d\'apprentissage profond introduit en 2017, " "utilisé principalement dans le domaine du traitement automatique des langues (TAL).", "À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus " "pour gérer des données séquentielles, telles que le langage naturel, pour des tâches " "telles que la traduction et la synthèse de texte.", ] self.tokenizer_integration_test_util( expected_encoding=__UpperCAmelCase , model_name="camembert-base" , revision="3a0641d9a1aeb7e848a74299e7e4c4bca216b4cf" , sequences=__UpperCAmelCase , )
353
"""simple docstring""" from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _a = logging.get_logger(__name__) # pylint: disable=invalid-name _a = '\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained("kandinsky-community/kandinsky-2-2-prior")\n >>> pipe_prior.to("cuda")\n >>> prompt = "red cat, 4k photo"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained("kandinsky-community/kandinsky-2-2-decoder")\n >>> pipe.to("cuda")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save("cat.png")\n ```\n' def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=8 ): UpperCAmelCase_ : List[str] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 UpperCAmelCase_ : Tuple = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class A_ (lowercase__ ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , ): """simple docstring""" super().__init__() self.register_modules( unet=lowercase_ , scheduler=lowercase_ , movq=lowercase_ , ) UpperCAmelCase_ : int = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" if latents is None: UpperCAmelCase_ : Dict = randn_tensor(lowercase_ , generator=lowercase_ , device=lowercase_ , dtype=lowercase_ ) else: if latents.shape != shape: raise ValueError(F"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) UpperCAmelCase_ : str = latents.to(lowercase_ ) UpperCAmelCase_ : Dict = latents * scheduler.init_noise_sigma return latents def UpperCamelCase__ ( self , lowercase_=0 ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) UpperCAmelCase_ : Any = torch.device(F"""cuda:{gpu_id}""" ) UpperCAmelCase_ : int = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowercase_ , lowercase_ ) def UpperCamelCase__ ( self , lowercase_=0 ): """simple docstring""" if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) UpperCAmelCase_ : Any = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=lowercase_ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) UpperCAmelCase_ : List[Any] = None for cpu_offloaded_model in [self.unet, self.movq]: UpperCAmelCase_ , UpperCAmelCase_ : str = cpu_offload_with_hook(lowercase_ , lowercase_ , prev_module_hook=lowercase_ ) # We'll offload the last model manually. UpperCAmelCase_ : Tuple = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCamelCase__ ( self ): """simple docstring""" if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowercase_ , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(lowercase_ ) def __call__( self , lowercase_ , lowercase_ , lowercase_ = 512 , lowercase_ = 512 , lowercase_ = 100 , lowercase_ = 4.0 , lowercase_ = 1 , lowercase_ = None , lowercase_ = None , lowercase_ = "pil" , lowercase_ = True , ): """simple docstring""" UpperCAmelCase_ : str = self._execution_device UpperCAmelCase_ : List[Any] = guidance_scale > 1.0 if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = torch.cat(lowercase_ , dim=0 ) UpperCAmelCase_ : Any = image_embeds.shape[0] * num_images_per_prompt if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : List[Any] = torch.cat(lowercase_ , dim=0 ) if do_classifier_free_guidance: UpperCAmelCase_ : Tuple = image_embeds.repeat_interleave(lowercase_ , dim=0 ) UpperCAmelCase_ : List[str] = negative_image_embeds.repeat_interleave(lowercase_ , dim=0 ) UpperCAmelCase_ : Optional[Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=lowercase_ ) self.scheduler.set_timesteps(lowercase_ , device=lowercase_ ) UpperCAmelCase_ : List[Any] = self.scheduler.timesteps UpperCAmelCase_ : List[str] = self.unet.config.in_channels UpperCAmelCase_ , UpperCAmelCase_ : Dict = downscale_height_and_width(lowercase_ , lowercase_ , self.movq_scale_factor ) # create initial latent UpperCAmelCase_ : int = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , lowercase_ , lowercase_ , lowercase_ , self.scheduler , ) for i, t in enumerate(self.progress_bar(lowercase_ ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase_ : Tuple = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase_ : Union[str, Any] = {"image_embeds": image_embeds} UpperCAmelCase_ : Optional[Any] = self.unet( sample=lowercase_ , timestep=lowercase_ , encoder_hidden_states=lowercase_ , added_cond_kwargs=lowercase_ , return_dict=lowercase_ , )[0] if do_classifier_free_guidance: UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = noise_pred.split(latents.shape[1] , dim=1 ) UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = noise_pred.chunk(2 ) UpperCAmelCase_ , UpperCAmelCase_ : Tuple = variance_pred.chunk(2 ) UpperCAmelCase_ : int = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) UpperCAmelCase_ : str = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): UpperCAmelCase_ , UpperCAmelCase_ : Dict = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase_ : List[str] = self.scheduler.step( lowercase_ , lowercase_ , lowercase_ , generator=lowercase_ , )[0] # post-processing UpperCAmelCase_ : Tuple = self.movq.decode(lowercase_ , force_not_quantize=lowercase_ )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: UpperCAmelCase_ : List[Any] = image * 0.5 + 0.5 UpperCAmelCase_ : int = image.clamp(0 , 1 ) UpperCAmelCase_ : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": UpperCAmelCase_ : Dict = self.numpy_to_pil(lowercase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowercase_ )
23
0
"""simple docstring""" import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class A_ (_UpperCAmelCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = tempfile.mkdtemp() UpperCAmelCase_ : List[Any] = 8 # DPR tok UpperCAmelCase_ : Dict = [ """[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] UpperCAmelCase_ : Any = os.path.join(self.tmpdirname , "dpr_tokenizer" ) os.makedirs(SCREAMING_SNAKE_CASE_ , exist_ok=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : Union[str, Any] = os.path.join(SCREAMING_SNAKE_CASE_ , DPR_VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) # BART tok UpperCAmelCase_ : List[Any] = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] UpperCAmelCase_ : str = dict(zip(SCREAMING_SNAKE_CASE_ , range(len(SCREAMING_SNAKE_CASE_ ) ) ) ) UpperCAmelCase_ : Dict = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] UpperCAmelCase_ : str = {"""unk_token""": """<unk>"""} UpperCAmelCase_ : List[str] = os.path.join(self.tmpdirname , "bart_tokenizer" ) os.makedirs(SCREAMING_SNAKE_CASE_ , exist_ok=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : Optional[int] = os.path.join(SCREAMING_SNAKE_CASE_ , BART_VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase_ : List[Any] = os.path.join(SCREAMING_SNAKE_CASE_ , BART_VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as fp: fp.write(json.dumps(SCREAMING_SNAKE_CASE_ ) + "\n" ) with open(self.merges_file , "w" , encoding="utf-8" ) as fp: fp.write("\n".join(SCREAMING_SNAKE_CASE_ ) ) def UpperCamelCase__ ( self ): """simple docstring""" return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , "dpr_tokenizer" ) ) def UpperCamelCase__ ( self ): """simple docstring""" return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , "dpr_tokenizer" ) ) def UpperCamelCase__ ( self ): """simple docstring""" return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , "bart_tokenizer" ) ) def UpperCamelCase__ ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = Dataset.from_dict( { "id": ["0", "1"], "text": ["foo", "bar"], "title": ["Foo", "Bar"], "embeddings": [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index("embeddings" , string_factory="Flat" , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[str] = self.get_dummy_dataset() UpperCAmelCase_ : str = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch("transformers.models.rag.retrieval_rag.load_dataset" ) as mock_load_dataset: UpperCAmelCase_ : Any = dataset UpperCAmelCase_ : Tuple = RagRetriever( SCREAMING_SNAKE_CASE_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : int = self.get_dummy_dataset() UpperCAmelCase_ : List[str] = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name="custom" , ) if from_disk: UpperCAmelCase_ : Optional[Any] = os.path.join(self.tmpdirname , "dataset" ) UpperCAmelCase_ : Union[str, Any] = os.path.join(self.tmpdirname , "index.faiss" ) dataset.get_index("embeddings" ).save(os.path.join(self.tmpdirname , "index.faiss" ) ) dataset.drop_index("embeddings" ) dataset.save_to_disk(os.path.join(self.tmpdirname , "dataset" ) ) del dataset UpperCAmelCase_ : int = RagRetriever( SCREAMING_SNAKE_CASE_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: UpperCAmelCase_ : str = RagRetriever( SCREAMING_SNAKE_CASE_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , SCREAMING_SNAKE_CASE_ ) , ) return retriever def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = Dataset.from_dict( { "id": ["0", "1"], "text": ["foo", "bar"], "title": ["Foo", "Bar"], "embeddings": [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index("embeddings" , string_factory="Flat" , metric_type=faiss.METRIC_INNER_PRODUCT ) UpperCAmelCase_ : Optional[Any] = os.path.join(self.tmpdirname , "hf_bert_base.hnswSQ8_correct_phi_128.c_index" ) dataset.save_faiss_index("embeddings" , index_file_name + ".index.dpr" ) pickle.dump(dataset["id"] , open(index_file_name + ".index_meta.dpr" , "wb" ) ) UpperCAmelCase_ : List[Any] = os.path.join(self.tmpdirname , "psgs_w100.tsv.pkl" ) UpperCAmelCase_ : Dict = {sample["""id"""]: [sample["""text"""], sample["""title"""]] for sample in dataset} pickle.dump(SCREAMING_SNAKE_CASE_ , open(SCREAMING_SNAKE_CASE_ , "wb" ) ) UpperCAmelCase_ : Optional[Any] = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name="legacy" , index_path=self.tmpdirname , ) UpperCAmelCase_ : Optional[int] = RagRetriever( SCREAMING_SNAKE_CASE_ , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Any = 1 UpperCAmelCase_ : List[Any] = self.get_dummy_canonical_hf_index_retriever() UpperCAmelCase_ : Optional[int] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) UpperCAmelCase_ : Tuple = retriever.retrieve(SCREAMING_SNAKE_CASE_ , n_docs=SCREAMING_SNAKE_CASE_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["embeddings", "id", "text", "title"] ) self.assertEqual(len(doc_dicts[0]["id"] ) , SCREAMING_SNAKE_CASE_ ) self.assertEqual(doc_dicts[0]["id"][0] , "1" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["id"][0] , "0" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch("transformers.models.rag.retrieval_rag.load_dataset" ) as mock_load_dataset: UpperCAmelCase_ : Dict = self.get_dummy_dataset() retriever.save_pretrained(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : Optional[int] = RagRetriever.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : str = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) UpperCAmelCase_ : Optional[Any] = retriever.retrieve(SCREAMING_SNAKE_CASE_ , n_docs=1 ) self.assertTrue(out is not None ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = 1 UpperCAmelCase_ : List[str] = self.get_dummy_custom_hf_index_retriever(from_disk=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : List[str] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) UpperCAmelCase_ : Tuple = retriever.retrieve(SCREAMING_SNAKE_CASE_ , n_docs=SCREAMING_SNAKE_CASE_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["embeddings", "id", "text", "title"] ) self.assertEqual(len(doc_dicts[0]["id"] ) , SCREAMING_SNAKE_CASE_ ) self.assertEqual(doc_dicts[0]["id"][0] , "1" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["id"][0] , "0" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.get_dummy_custom_hf_index_retriever(from_disk=SCREAMING_SNAKE_CASE_ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : int = RagRetriever.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : List[str] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) UpperCAmelCase_ : str = retriever.retrieve(SCREAMING_SNAKE_CASE_ , n_docs=1 ) self.assertTrue(out is not None ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = 1 UpperCAmelCase_ : List[str] = self.get_dummy_custom_hf_index_retriever(from_disk=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : List[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) UpperCAmelCase_ : str = retriever.retrieve(SCREAMING_SNAKE_CASE_ , n_docs=SCREAMING_SNAKE_CASE_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["embeddings", "id", "text", "title"] ) self.assertEqual(len(doc_dicts[0]["id"] ) , SCREAMING_SNAKE_CASE_ ) self.assertEqual(doc_dicts[0]["id"][0] , "1" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["id"][0] , "0" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.get_dummy_custom_hf_index_retriever(from_disk=SCREAMING_SNAKE_CASE_ ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : str = RagRetriever.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : int = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) UpperCAmelCase_ : List[Any] = retriever.retrieve(SCREAMING_SNAKE_CASE_ , n_docs=1 ) self.assertTrue(out is not None ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = 1 UpperCAmelCase_ : List[str] = self.get_dummy_legacy_index_retriever() UpperCAmelCase_ : Dict = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) UpperCAmelCase_ : Union[str, Any] = retriever.retrieve(SCREAMING_SNAKE_CASE_ , n_docs=SCREAMING_SNAKE_CASE_ ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(SCREAMING_SNAKE_CASE_ ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["text", "title"] ) self.assertEqual(len(doc_dicts[0]["text"] ) , SCREAMING_SNAKE_CASE_ ) self.assertEqual(doc_dicts[0]["text"][0] , "bar" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["text"][0] , "foo" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : Tuple = RagRetriever.from_pretrained(SCREAMING_SNAKE_CASE_ ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : Optional[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) UpperCAmelCase_ : Optional[Any] = retriever.retrieve(SCREAMING_SNAKE_CASE_ , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def UpperCamelCase__ ( self ): """simple docstring""" import torch UpperCAmelCase_ : str = 1 UpperCAmelCase_ : List[Any] = self.get_dummy_canonical_hf_index_retriever() UpperCAmelCase_ : Tuple = [[5, 7], [10, 11]] UpperCAmelCase_ : str = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) UpperCAmelCase_ : Any = retriever(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , prefix=retriever.config.generator.prefix , n_docs=SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : Any = ( out["""context_input_ids"""], out["""context_attention_mask"""], out["""retrieved_doc_embeds"""], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , np.ndarray ) UpperCAmelCase_ : Dict = retriever( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , prefix=retriever.config.generator.prefix , n_docs=SCREAMING_SNAKE_CASE_ , return_tensors="pt" , ) UpperCAmelCase_ : Union[str, Any] = ( # noqa: F841 out["""context_input_ids"""], out["""context_attention_mask"""], out["""retrieved_doc_embeds"""], out["""doc_ids"""], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , torch.Tensor ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , torch.Tensor ) self.assertIsInstance(SCREAMING_SNAKE_CASE_ , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.get_dpr_ctx_encoder_tokenizer() UpperCAmelCase_ : Optional[int] = 1 UpperCAmelCase_ : int = self.get_dummy_custom_hf_index_retriever(from_disk=SCREAMING_SNAKE_CASE_ ) retriever.set_ctx_encoder_tokenizer(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : Optional[int] = [[5, 7], [10, 11]] UpperCAmelCase_ : List[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) UpperCAmelCase_ : Any = retriever(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , prefix=retriever.config.generator.prefix , n_docs=SCREAMING_SNAKE_CASE_ ) self.assertEqual( len(SCREAMING_SNAKE_CASE_ ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ("tokenized_doc_ids", "tokenized_doc_attention_mask") ) , SCREAMING_SNAKE_CASE_ ) # check for doc token related keys in dictionary.
354
"""simple docstring""" import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _a = logging.get_logger(__name__) _a = { 'facebook/detr-resnet-50': 'https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json', # See all DETR models at https://huggingface.co/models?filter=detr } class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """detr""" SCREAMING_SNAKE_CASE__ : str = ["""past_key_values"""] SCREAMING_SNAKE_CASE__ : Union[str, Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , lowercase_=True , lowercase_=None , lowercase_=3 , lowercase_=100 , lowercase_=6 , lowercase_=2048 , lowercase_=8 , lowercase_=6 , lowercase_=2048 , lowercase_=8 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=True , lowercase_="relu" , lowercase_=256 , lowercase_=0.1 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=0.02 , lowercase_=1.0 , lowercase_=False , lowercase_="sine" , lowercase_="resnet50" , lowercase_=True , lowercase_=False , lowercase_=1 , lowercase_=5 , lowercase_=2 , lowercase_=1 , lowercase_=1 , lowercase_=5 , lowercase_=2 , lowercase_=0.1 , **lowercase_ , ): """simple docstring""" if backbone_config is not None and use_timm_backbone: raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`." ) if not use_timm_backbone: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) UpperCAmelCase_ : Union[str, Any] = CONFIG_MAPPING["resnet"](out_features=["stage4"] ) elif isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = backbone_config.get("model_type" ) UpperCAmelCase_ : int = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase_ : Any = config_class.from_dict(lowercase_ ) # set timm attributes to None UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = None, None, None UpperCAmelCase_ : int = use_timm_backbone UpperCAmelCase_ : int = backbone_config UpperCAmelCase_ : List[Any] = num_channels UpperCAmelCase_ : int = num_queries UpperCAmelCase_ : Union[str, Any] = d_model UpperCAmelCase_ : str = encoder_ffn_dim UpperCAmelCase_ : Tuple = encoder_layers UpperCAmelCase_ : List[Any] = encoder_attention_heads UpperCAmelCase_ : Union[str, Any] = decoder_ffn_dim UpperCAmelCase_ : Optional[Any] = decoder_layers UpperCAmelCase_ : Union[str, Any] = decoder_attention_heads UpperCAmelCase_ : Optional[int] = dropout UpperCAmelCase_ : List[str] = attention_dropout UpperCAmelCase_ : Any = activation_dropout UpperCAmelCase_ : str = activation_function UpperCAmelCase_ : Tuple = init_std UpperCAmelCase_ : Optional[Any] = init_xavier_std UpperCAmelCase_ : Optional[Any] = encoder_layerdrop UpperCAmelCase_ : Optional[int] = decoder_layerdrop UpperCAmelCase_ : Tuple = encoder_layers UpperCAmelCase_ : int = auxiliary_loss UpperCAmelCase_ : Optional[Any] = position_embedding_type UpperCAmelCase_ : Tuple = backbone UpperCAmelCase_ : Optional[int] = use_pretrained_backbone UpperCAmelCase_ : Dict = dilation # Hungarian matcher UpperCAmelCase_ : Union[str, Any] = class_cost UpperCAmelCase_ : Any = bbox_cost UpperCAmelCase_ : int = giou_cost # Loss coefficients UpperCAmelCase_ : str = mask_loss_coefficient UpperCAmelCase_ : Any = dice_loss_coefficient UpperCAmelCase_ : Optional[Any] = bbox_loss_coefficient UpperCAmelCase_ : List[str] = giou_loss_coefficient UpperCAmelCase_ : List[Any] = eos_coefficient super().__init__(is_encoder_decoder=lowercase_ , **lowercase_ ) @property def UpperCamelCase__ ( self ): """simple docstring""" return self.encoder_attention_heads @property def UpperCamelCase__ ( self ): """simple docstring""" return self.d_model @classmethod def UpperCamelCase__ ( cls , lowercase_ , **lowercase_ ): """simple docstring""" return cls(backbone_config=lowercase_ , **lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: UpperCAmelCase_ : Union[str, Any] = self.backbone_config.to_dict() UpperCAmelCase_ : str = self.__class__.model_type return output class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = version.parse("""1.11""" ) @property def UpperCamelCase__ ( self ): """simple docstring""" return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("pixel_mask", {0: "batch"}), ] ) @property def UpperCamelCase__ ( self ): """simple docstring""" return 1E-5 @property def UpperCamelCase__ ( self ): """simple docstring""" return 12
23
0
"""simple docstring""" from __future__ import annotations import math import numpy as np from numpy.linalg import norm def __a ( __lowerCamelCase, __lowerCamelCase ): return math.sqrt(sum(pow(a - b, 2 ) for a, b in zip(lowercase__, lowercase__ ) ) ) def __a ( __lowerCamelCase, __lowerCamelCase ): if dataset.ndim != value_array.ndim: UpperCAmelCase_ : Dict = ( """Wrong input data's dimensions... """ f"""dataset : {dataset.ndim}, value_array : {value_array.ndim}""" ) raise ValueError(lowercase__ ) try: if dataset.shape[1] != value_array.shape[1]: UpperCAmelCase_ : Union[str, Any] = ( """Wrong input data's shape... """ f"""dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}""" ) raise ValueError(lowercase__ ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError("Wrong shape" ) if dataset.dtype != value_array.dtype: UpperCAmelCase_ : str = ( """Input data have different datatype... """ f"""dataset : {dataset.dtype}, value_array : {value_array.dtype}""" ) raise TypeError(lowercase__ ) UpperCAmelCase_ : str = [] for value in value_array: UpperCAmelCase_ : int = euclidean(lowercase__, dataset[0] ) UpperCAmelCase_ : Tuple = dataset[0].tolist() for dataset_value in dataset[1:]: UpperCAmelCase_ : Any = euclidean(lowercase__, lowercase__ ) if dist > temp_dist: UpperCAmelCase_ : Any = temp_dist UpperCAmelCase_ : List[Any] = dataset_value.tolist() answer.append([vector, dist] ) return answer def __a ( __lowerCamelCase, __lowerCamelCase ): return np.dot(lowercase__, lowercase__ ) / (norm(lowercase__ ) * norm(lowercase__ )) if __name__ == "__main__": import doctest doctest.testmod()
355
"""simple docstring""" _a = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(100_000)] def __a ( __lowerCamelCase ): UpperCAmelCase_ : Optional[int] = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 10_0000] number //= 10_0000 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution _a = [None] * 10_000_000 _a = True _a = False def __a ( __lowerCamelCase ): if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore UpperCAmelCase_ : Dict = chain(next_number(__lowerCamelCase ) ) UpperCAmelCase_ : List[str] = number_chain while number < 1000_0000: UpperCAmelCase_ : List[Any] = number_chain number *= 10 return number_chain def __a ( __lowerCamelCase = 1000_0000 ): for i in range(1, __lowerCamelCase ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(__lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() print(f"""{solution() = }""")
23
0
"""simple docstring""" import argparse from pathlib import Path import fairseq import torch from fairseq.models.xmod import XMODModel as FairseqXmodModel from packaging import version from transformers import XmodConfig, XmodForMaskedLM, XmodForSequenceClassification from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse('0.12.2'): raise Exception('requires fairseq >= 0.12.2') if version.parse(fairseq.__version__) > version.parse('2'): raise Exception('requires fairseq < v2') logging.set_verbosity_info() _a = logging.get_logger(__name__) _a = 'Hello, World!' _a = 'en_XX' def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : List[Any] = Path("data_bin" ) UpperCAmelCase_ : str = FairseqXmodModel.from_pretrained( model_name_or_path=str(Path(A__ ).parent ), checkpoint_file=Path(A__ ).name, _name="xmod_base", arch="xmod_base", task="multilingual_masked_lm", data_name_or_path=str(A__ ), bpe="sentencepiece", sentencepiece_model=str(Path(A__ ).parent / "sentencepiece.bpe.model" ), src_dict=str(data_dir / "dict.txt" ), ) xmod.eval() # disable dropout print(A__ ) UpperCAmelCase_ : int = xmod.model.encoder.sentence_encoder UpperCAmelCase_ : List[str] = XmodConfig( vocab_size=xmod_sent_encoder.embed_tokens.num_embeddings, hidden_size=xmod.cfg.model.encoder_embed_dim, num_hidden_layers=xmod.cfg.model.encoder_layers, num_attention_heads=xmod.cfg.model.encoder_attention_heads, intermediate_size=xmod.cfg.model.encoder_ffn_embed_dim, max_position_embeddings=514, type_vocab_size=1, layer_norm_eps=1E-5, pre_norm=xmod.cfg.model.encoder_normalize_before, adapter_reduction_factor=getattr(xmod.cfg.model, "bottleneck", 2 ), adapter_layer_norm=xmod.cfg.model.adapter_layer_norm, adapter_reuse_layer_norm=xmod.cfg.model.adapter_reuse_layer_norm, ln_before_adapter=xmod.cfg.model.ln_before_adapter, languages=xmod.cfg.model.languages, ) if classification_head: UpperCAmelCase_ : Optional[int] = xmod.model.classification_heads["mnli"].out_proj.weight.shape[0] print("Our X-MOD config:", A__ ) UpperCAmelCase_ : Optional[Any] = XmodForSequenceClassification(A__ ) if classification_head else XmodForMaskedLM(A__ ) model.eval() # Now let's copy all the weights. # Embeddings UpperCAmelCase_ : int = xmod_sent_encoder.embed_tokens.weight UpperCAmelCase_ : int = xmod_sent_encoder.embed_positions.weight UpperCAmelCase_ : Union[str, Any] = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c xmod doesn't use them. UpperCAmelCase_ : List[Any] = xmod_sent_encoder.layernorm_embedding.weight UpperCAmelCase_ : Dict = xmod_sent_encoder.layernorm_embedding.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer UpperCAmelCase_ : int = model.roberta.encoder.layer[i] UpperCAmelCase_ : str = xmod_sent_encoder.layers[i] # self attention UpperCAmelCase_ : Optional[Any] = layer.attention.self if not ( xmod_layer.self_attn.k_proj.weight.data.shape == xmod_layer.self_attn.q_proj.weight.data.shape == xmod_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ): raise AssertionError("Dimensions of self-attention weights do not match." ) UpperCAmelCase_ : Tuple = xmod_layer.self_attn.q_proj.weight UpperCAmelCase_ : Dict = xmod_layer.self_attn.q_proj.bias UpperCAmelCase_ : Dict = xmod_layer.self_attn.k_proj.weight UpperCAmelCase_ : List[Any] = xmod_layer.self_attn.k_proj.bias UpperCAmelCase_ : str = xmod_layer.self_attn.v_proj.weight UpperCAmelCase_ : int = xmod_layer.self_attn.v_proj.bias # self-attention output UpperCAmelCase_ : Tuple = layer.attention.output if self_output.dense.weight.shape != xmod_layer.self_attn.out_proj.weight.shape: raise AssertionError("Dimensions of self-attention output weights do not match." ) UpperCAmelCase_ : List[str] = xmod_layer.self_attn.out_proj.weight UpperCAmelCase_ : Dict = xmod_layer.self_attn.out_proj.bias UpperCAmelCase_ : List[str] = xmod_layer.self_attn_layer_norm.weight UpperCAmelCase_ : Optional[Any] = xmod_layer.self_attn_layer_norm.bias # intermediate UpperCAmelCase_ : List[str] = layer.intermediate if intermediate.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError("Dimensions of intermediate weights do not match." ) UpperCAmelCase_ : List[str] = xmod_layer.fca.weight UpperCAmelCase_ : int = xmod_layer.fca.bias # output UpperCAmelCase_ : Tuple = layer.output if bert_output.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError("Dimensions of feed-forward weights do not match." ) UpperCAmelCase_ : int = xmod_layer.fca.weight UpperCAmelCase_ : Union[str, Any] = xmod_layer.fca.bias UpperCAmelCase_ : List[Any] = xmod_layer.final_layer_norm.weight UpperCAmelCase_ : str = xmod_layer.final_layer_norm.bias if bert_output.adapter_layer_norm is not None: UpperCAmelCase_ : Optional[int] = xmod_layer.adapter_layer_norm.weight UpperCAmelCase_ : Tuple = xmod_layer.adapter_layer_norm.bias if sorted(bert_output.adapter_modules.keys() ) != sorted(xmod_layer.adapter_modules.keys() ): raise AssertionError("Lists of language adapters do not match." ) for lang_code, adapter in xmod_layer.adapter_modules.items(): UpperCAmelCase_ : Dict = bert_output.adapter_modules[lang_code] UpperCAmelCase_ : Dict = xmod_layer.adapter_modules[lang_code] UpperCAmelCase_ : int = from_adapter.fca.weight UpperCAmelCase_ : Optional[Any] = from_adapter.fca.bias UpperCAmelCase_ : Any = from_adapter.fca.weight UpperCAmelCase_ : Tuple = from_adapter.fca.bias # end of layer if xmod_sent_encoder.layer_norm is not None: UpperCAmelCase_ : Union[str, Any] = xmod_sent_encoder.layer_norm.weight UpperCAmelCase_ : str = xmod_sent_encoder.layer_norm.bias if classification_head: UpperCAmelCase_ : List[Any] = xmod.model.classification_heads["mnli"].dense.weight UpperCAmelCase_ : List[Any] = xmod.model.classification_heads["mnli"].dense.bias UpperCAmelCase_ : Any = xmod.model.classification_heads["mnli"].out_proj.weight UpperCAmelCase_ : int = xmod.model.classification_heads["mnli"].out_proj.bias else: # LM Head UpperCAmelCase_ : Dict = xmod.model.encoder.lm_head.dense.weight UpperCAmelCase_ : Tuple = xmod.model.encoder.lm_head.dense.bias UpperCAmelCase_ : Optional[Any] = xmod.model.encoder.lm_head.layer_norm.weight UpperCAmelCase_ : str = xmod.model.encoder.lm_head.layer_norm.bias UpperCAmelCase_ : Tuple = xmod.model.encoder.lm_head.weight UpperCAmelCase_ : str = xmod.model.encoder.lm_head.bias # Let's check that we get the same results. UpperCAmelCase_ : List[str] = xmod.encode(A__ ).unsqueeze(0 ) # batch of size 1 model.roberta.set_default_language(A__ ) UpperCAmelCase_ : str = model(A__ )[0] if classification_head: UpperCAmelCase_ : str = xmod.model.classification_heads["mnli"](xmod.extract_features(A__ ) ) else: UpperCAmelCase_ : str = xmod.model(A__, lang_id=[SAMPLE_LANGUAGE] )[0] print(our_output.shape, their_output.shape ) UpperCAmelCase_ : int = torch.max(torch.abs(our_output - their_output ) ).item() print(f"""max_absolute_diff = {max_absolute_diff}""" ) # ~ 1e-7 UpperCAmelCase_ : int = torch.allclose(A__, A__, atol=1E-3 ) print("Do both models output the same tensors?", "🔥" if success else "💩" ) if not success: raise Exception("Something went wRoNg" ) Path(A__ ).mkdir(parents=A__, exist_ok=A__ ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(A__ ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--xmod_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--classification_head', action='store_true', help='Whether to convert a final classification head.' ) _a = parser.parse_args() convert_xmod_checkpoint_to_pytorch( args.xmod_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
356
"""simple docstring""" def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # Return True if there is node that has not iterated. UpperCAmelCase_ : List[Any] = [False] * len(__lowerCamelCase ) UpperCAmelCase_ : Any = [] queue.append(__lowerCamelCase ) UpperCAmelCase_ : Tuple = True while queue: UpperCAmelCase_ : str = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(__lowerCamelCase ) UpperCAmelCase_ : Any = True UpperCAmelCase_ : Union[str, Any] = u return visited[t] def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # This array is filled by BFS and to store path UpperCAmelCase_ : List[str] = [-1] * (len(__lowerCamelCase )) UpperCAmelCase_ : Any = 0 while bfs(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : int = float("Inf" ) UpperCAmelCase_ : Tuple = sink while s != source: # Find the minimum value in select path UpperCAmelCase_ : Tuple = min(__lowerCamelCase, graph[parent[s]][s] ) UpperCAmelCase_ : Dict = parent[s] max_flow += path_flow UpperCAmelCase_ : Optional[Any] = sink while v != source: UpperCAmelCase_ : List[str] = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow UpperCAmelCase_ : Optional[int] = parent[v] return max_flow _a = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] _a , _a = 0, 5 print(ford_fulkerson(graph, source, sink))
23
0
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { 'facebook/xmod-base': 'https://huggingface.co/facebook/xmod-base/resolve/main/config.json', 'facebook/xmod-large-prenorm': 'https://huggingface.co/facebook/xmod-large-prenorm/resolve/main/config.json', 'facebook/xmod-base-13-125k': 'https://huggingface.co/facebook/xmod-base-13-125k/resolve/main/config.json', 'facebook/xmod-base-30-125k': 'https://huggingface.co/facebook/xmod-base-30-125k/resolve/main/config.json', 'facebook/xmod-base-30-195k': 'https://huggingface.co/facebook/xmod-base-30-195k/resolve/main/config.json', 'facebook/xmod-base-60-125k': 'https://huggingface.co/facebook/xmod-base-60-125k/resolve/main/config.json', 'facebook/xmod-base-60-265k': 'https://huggingface.co/facebook/xmod-base-60-265k/resolve/main/config.json', 'facebook/xmod-base-75-125k': 'https://huggingface.co/facebook/xmod-base-75-125k/resolve/main/config.json', 'facebook/xmod-base-75-269k': 'https://huggingface.co/facebook/xmod-base-75-269k/resolve/main/config.json', } class A_ (lowerCAmelCase_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = 'xmod' def __init__( self , lowercase_=3_0522 , lowercase_=768 , lowercase_=12 , lowercase_=12 , lowercase_=3072 , lowercase_="gelu" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=512 , lowercase_=2 , lowercase_=0.02 , lowercase_=1E-1_2 , lowercase_=1 , lowercase_=0 , lowercase_=2 , lowercase_="absolute" , lowercase_=True , lowercase_=None , lowercase_=False , lowercase_=2 , lowercase_=False , lowercase_=True , lowercase_=True , lowercase_=("en_XX",) , lowercase_=None , **lowercase_ , ): """simple docstring""" super().__init__(pad_token_id=__lowerCAmelCase , bos_token_id=__lowerCAmelCase , eos_token_id=__lowerCAmelCase , **__lowerCAmelCase ) UpperCAmelCase_ : int = vocab_size UpperCAmelCase_ : List[str] = hidden_size UpperCAmelCase_ : Tuple = num_hidden_layers UpperCAmelCase_ : List[str] = num_attention_heads UpperCAmelCase_ : List[str] = hidden_act UpperCAmelCase_ : int = intermediate_size UpperCAmelCase_ : Union[str, Any] = hidden_dropout_prob UpperCAmelCase_ : Union[str, Any] = attention_probs_dropout_prob UpperCAmelCase_ : Optional[Any] = max_position_embeddings UpperCAmelCase_ : Tuple = type_vocab_size UpperCAmelCase_ : int = initializer_range UpperCAmelCase_ : Optional[int] = layer_norm_eps UpperCAmelCase_ : List[Any] = position_embedding_type UpperCAmelCase_ : str = use_cache UpperCAmelCase_ : Tuple = classifier_dropout UpperCAmelCase_ : Optional[Any] = pre_norm UpperCAmelCase_ : Tuple = adapter_reduction_factor UpperCAmelCase_ : str = adapter_layer_norm UpperCAmelCase_ : Tuple = adapter_reuse_layer_norm UpperCAmelCase_ : List[str] = ln_before_adapter UpperCAmelCase_ : str = list(__lowerCAmelCase ) UpperCAmelCase_ : Union[str, Any] = default_language class A_ (lowerCAmelCase_ ): '''simple docstring''' @property def UpperCamelCase__ ( self ): """simple docstring""" if self.task == "multiple-choice": UpperCAmelCase_ : str = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCAmelCase_ : str = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
357
"""simple docstring""" import datasets _a = '\\n@InProceedings{conneau2018xnli,\n author = "Conneau, Alexis\n and Rinott, Ruty\n and Lample, Guillaume\n and Williams, Adina\n and Bowman, Samuel R.\n and Schwenk, Holger\n and Stoyanov, Veselin",\n title = "XNLI: Evaluating Cross-lingual Sentence Representations",\n booktitle = "Proceedings of the 2018 Conference on Empirical Methods\n in Natural Language Processing",\n year = "2018",\n publisher = "Association for Computational Linguistics",\n location = "Brussels, Belgium",\n}\n' _a = '\\nXNLI is a subset of a few thousand examples from MNLI which has been translated\ninto a 14 different languages (some low-ish resource). As with MNLI, the goal is\nto predict textual entailment (does sentence A imply/contradict/neither sentence\nB) and is a classification task (given two sentences, predict one of three\nlabels).\n' _a = '\nComputes XNLI score which is just simple accuracy.\nArgs:\n predictions: Predicted labels.\n references: Ground truth labels.\nReturns:\n \'accuracy\': accuracy\nExamples:\n\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> xnli_metric = datasets.load_metric("xnli")\n >>> results = xnli_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n' def __a ( __lowerCamelCase, __lowerCamelCase ): return (preds == labels).mean() @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64" if self.config_name != "sts-b" else "float32" ), "references": datasets.Value("int64" if self.config_name != "sts-b" else "float32" ), } ) , codebase_urls=[] , reference_urls=[] , format="numpy" , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" return {"accuracy": simple_accuracy(lowercase_ , lowercase_ )}
23
0
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class A_ (unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = FlaxMTaForConditionalGeneration.from_pretrained("google/mt5-small" ) UpperCAmelCase_ : Optional[Any] = AutoTokenizer.from_pretrained("google/mt5-small" ) UpperCAmelCase_ : List[Any] = tokenizer("Hello there" , return_tensors="np" ).input_ids UpperCAmelCase_ : List[str] = tokenizer("Hi I am" , return_tensors="np" ).input_ids UpperCAmelCase_ : Dict = shift_tokens_right(a_ , model.config.pad_token_id , model.config.decoder_start_token_id ) UpperCAmelCase_ : Optional[int] = model(a_ , decoder_input_ids=a_ ).logits UpperCAmelCase_ : Dict = optax.softmax_cross_entropy(a_ , onehot(a_ , logits.shape[-1] ) ).mean() UpperCAmelCase_ : List[Any] = -(labels.shape[-1] * loss.item()) UpperCAmelCase_ : str = -84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
358
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy _a = logging.get_logger(__name__) class A_ (lowercase__ ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = feature_size UpperCAmelCase_ : Any = sampling_rate UpperCAmelCase_ : Any = padding_value UpperCAmelCase_ : str = kwargs.pop("padding_side" , "right" ) UpperCAmelCase_ : List[str] = kwargs.pop("return_attention_mask" , lowercase_ ) super().__init__(**lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = True , lowercase_ = None , lowercase_ = False , lowercase_ = None , lowercase_ = None , lowercase_ = None , ): """simple docstring""" # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(lowercase_ , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): UpperCAmelCase_ : Dict = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( "You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`" F""" to this method that includes {self.model_input_names[0]}, but you provided""" F""" {list(processed_features.keys() )}""" ) UpperCAmelCase_ : Tuple = processed_features[self.model_input_names[0]] UpperCAmelCase_ : List[str] = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(lowercase_ ) == 0: if return_attention_mask: UpperCAmelCase_ : Union[str, Any] = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch UpperCAmelCase_ : List[str] = required_input[0] if isinstance(lowercase_ , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. UpperCAmelCase_ : Any = 0 while len(required_input[index] ) == 0: index += 1 if index < len(lowercase_ ): UpperCAmelCase_ : Optional[Any] = required_input[index][0] if return_tensors is None: if is_tf_tensor(lowercase_ ): UpperCAmelCase_ : Dict = "tf" elif is_torch_tensor(lowercase_ ): UpperCAmelCase_ : Any = "pt" elif isinstance(lowercase_ , (int, float, list, tuple, np.ndarray) ): UpperCAmelCase_ : str = "np" else: raise ValueError( F"""type of {first_element} unknown: {type(lowercase_ )}. """ "Should be one of a python, numpy, pytorch or tensorflow object." ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): UpperCAmelCase_ : Optional[int] = to_numpy(lowercase_ ) else: UpperCAmelCase_ : List[str] = [to_numpy(lowercase_ ) for v in value] # Convert padding_strategy in PaddingStrategy UpperCAmelCase_ : Dict = self._get_padding_strategies(padding=lowercase_ , max_length=lowercase_ ) UpperCAmelCase_ : str = processed_features[self.model_input_names[0]] UpperCAmelCase_ : int = len(lowercase_ ) if not all(len(lowercase_ ) == batch_size for v in processed_features.values() ): raise ValueError("Some items in the output dictionary have a different batch size than others." ) UpperCAmelCase_ : int = [] for i in range(lowercase_ ): UpperCAmelCase_ : str = {k: v[i] for k, v in processed_features.items()} # truncation UpperCAmelCase_ : List[str] = self._truncate( lowercase_ , max_length=lowercase_ , pad_to_multiple_of=lowercase_ , truncation=lowercase_ , ) truncated_inputs.append(lowercase_ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length UpperCAmelCase_ : str = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) UpperCAmelCase_ : Dict = PaddingStrategy.MAX_LENGTH UpperCAmelCase_ : List[str] = {} for i in range(lowercase_ ): # padding UpperCAmelCase_ : int = self._pad( truncated_inputs[i] , max_length=lowercase_ , padding_strategy=lowercase_ , pad_to_multiple_of=lowercase_ , return_attention_mask=lowercase_ , ) for key, value in outputs.items(): if key not in batch_outputs: UpperCAmelCase_ : Any = [] if value.dtype is np.dtype(np.floataa ): UpperCAmelCase_ : List[Any] = value.astype(np.floataa ) batch_outputs[key].append(lowercase_ ) return BatchFeature(lowercase_ , tensor_type=lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = PaddingStrategy.DO_NOT_PAD , lowercase_ = None , lowercase_ = None , ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: UpperCAmelCase_ : Tuple = len(lowercase_ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCAmelCase_ : Tuple = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCAmelCase_ : Dict = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(lowercase_ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: UpperCAmelCase_ : Optional[int] = np.ones(len(lowercase_ ) , dtype=np.intaa ) if needs_to_be_padded: UpperCAmelCase_ : Dict = max_length - len(lowercase_ ) if self.padding_side == "right": if return_attention_mask: UpperCAmelCase_ : List[Any] = np.pad( processed_features["attention_mask"] , (0, difference) ) UpperCAmelCase_ : Dict = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) UpperCAmelCase_ : Optional[Any] = np.pad( lowercase_ , lowercase_ , "constant" , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: UpperCAmelCase_ : Optional[Any] = np.pad( processed_features["attention_mask"] , (difference, 0) ) UpperCAmelCase_ : Dict = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) UpperCAmelCase_ : str = np.pad( lowercase_ , lowercase_ , "constant" , constant_values=self.padding_value ) else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return processed_features def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , ): """simple docstring""" if not truncation: return processed_features elif truncation and max_length is None: raise ValueError("When setting ``truncation=True``, make sure that ``max_length`` is defined." ) UpperCAmelCase_ : Optional[int] = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCAmelCase_ : Union[str, Any] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCAmelCase_ : Optional[Any] = len(lowercase_ ) > max_length if needs_to_be_truncated: UpperCAmelCase_ : int = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: UpperCAmelCase_ : Dict = processed_features["attention_mask"][:max_length] return processed_features def UpperCamelCase__ ( self , lowercase_=False , lowercase_=None ): """simple docstring""" # Get padding strategy if padding is not False: if padding is True: UpperCAmelCase_ : Dict = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : Optional[Any] = PaddingStrategy(lowercase_ ) elif isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = padding else: UpperCAmelCase_ : str = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F"""When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined""" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( "Asking to pad but the feature_extractor does not have a padding value. Please select a value to use" " as `padding_value`. For example: `feature_extractor.padding_value = 0.0`." ) return padding_strategy
23
0
import datasets from .evaluate import evaluate _a = """\ @article{hendrycks2021cuad, title={CUAD: An Expert-Annotated NLP Dataset for Legal Contract Review}, author={Dan Hendrycks and Collin Burns and Anya Chen and Spencer Ball}, journal={arXiv preprint arXiv:2103.06268}, year={2021} } """ _a = """ This metric wrap the official scoring script for version 1 of the Contract Understanding Atticus Dataset (CUAD). Contract Understanding Atticus Dataset (CUAD) v1 is a corpus of more than 13,000 labels in 510 commercial legal contracts that have been manually labeled to identify 41 categories of important clauses that lawyers look for when reviewing contracts in connection with corporate transactions. """ _a = """ Computes CUAD scores (EM, F1, AUPR, Precision@80%Recall, and Precision@90%Recall). Args: predictions: List of question-answers dictionaries with the following key-values: - 'id': id of the question-answer pair as given in the references (see below) - 'prediction_text': list of possible texts for the answer, as a list of strings depending on a threshold on the confidence probability of each prediction. references: List of question-answers dictionaries with the following key-values: - 'id': id of the question-answer pair (see above), - 'answers': a Dict in the CUAD dataset format { 'text': list of possible texts for the answer, as a list of strings 'answer_start': list of start positions for the answer, as a list of ints } Note that answer_start values are not taken into account to compute the metric. Returns: 'exact_match': Exact match (the normalized answer exactly match the gold answer) 'f1': The F-score of predicted tokens versus the gold answer 'aupr': Area Under the Precision-Recall curve 'prec_at_80_recall': Precision at 80% recall 'prec_at_90_recall': Precision at 90% recall Examples: >>> predictions = [{'prediction_text': ['The seller:', 'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.'], 'id': 'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties'}] >>> references = [{'answers': {'answer_start': [143, 49], 'text': ['The seller:', 'The buyer/End-User: Shenzhen LOHAS Supply Chain Management Co., Ltd.']}, 'id': 'LohaCompanyltd_20191209_F-1_EX-10.16_11917878_EX-10.16_Supply Agreement__Parties'}] >>> cuad_metric = datasets.load_metric(\"cuad\") >>> results = cuad_metric.compute(predictions=predictions, references=references) >>> print(results) {'exact_match': 100.0, 'f1': 100.0, 'aupr': 0.0, 'prec_at_80_recall': 1.0, 'prec_at_90_recall': 1.0} """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": { "id": datasets.Value("string" ), "prediction_text": datasets.features.Sequence(datasets.Value("string" ) ), }, "references": { "id": datasets.Value("string" ), "answers": datasets.features.Sequence( { "text": datasets.Value("string" ), "answer_start": datasets.Value("int32" ), } ), }, } ) , codebase_urls=["https://www.atticusprojectai.org/cuad"] , reference_urls=["https://www.atticusprojectai.org/cuad"] , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = {prediction["id"]: prediction["prediction_text"] for prediction in predictions} UpperCAmelCase_ : Tuple = [ { "paragraphs": [ { "qas": [ { "answers": [{"text": answer_text} for answer_text in ref["answers"]["text"]], "id": ref["id"], } for ref in references ] } ] } ] UpperCAmelCase_ : Dict = evaluate(dataset=__lowerCAmelCase , predictions=__lowerCAmelCase ) return score
359
"""simple docstring""" import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = torch.nn.Linear(10 , 10 ) UpperCAmelCase_ : List[str] = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase_ : Optional[Any] = Accelerator() UpperCAmelCase_ : Tuple = accelerator.prepare(lowercase_ ) try: pickle.loads(pickle.dumps(lowercase_ ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
23
0
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=None, __lowerCamelCase=None ): if attention_mask is None: UpperCAmelCase_ : List[str] = tf.cast(tf.math.not_equal(a__, config.pad_token_id ), tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class A_ : '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = OPTConfig SCREAMING_SNAKE_CASE__ : Tuple = {} SCREAMING_SNAKE_CASE__ : int = """gelu""" def __init__( self , lowercase_ , lowercase_=13 , lowercase_=7 , lowercase_=True , lowercase_=False , lowercase_=99 , lowercase_=16 , lowercase_=2 , lowercase_=4 , lowercase_=4 , lowercase_="gelu" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=20 , lowercase_=2 , lowercase_=1 , lowercase_=0 , lowercase_=16 , lowercase_=16 , ): """simple docstring""" UpperCAmelCase_ : Tuple = parent UpperCAmelCase_ : Dict = batch_size UpperCAmelCase_ : Union[str, Any] = seq_length UpperCAmelCase_ : Optional[int] = is_training UpperCAmelCase_ : Union[str, Any] = use_labels UpperCAmelCase_ : str = vocab_size UpperCAmelCase_ : List[Any] = hidden_size UpperCAmelCase_ : Union[str, Any] = num_hidden_layers UpperCAmelCase_ : str = num_attention_heads UpperCAmelCase_ : List[Any] = intermediate_size UpperCAmelCase_ : List[Any] = hidden_act UpperCAmelCase_ : Any = hidden_dropout_prob UpperCAmelCase_ : str = attention_probs_dropout_prob UpperCAmelCase_ : Union[str, Any] = max_position_embeddings UpperCAmelCase_ : Optional[int] = eos_token_id UpperCAmelCase_ : Optional[Any] = pad_token_id UpperCAmelCase_ : Optional[Any] = bos_token_id UpperCAmelCase_ : List[Any] = embed_dim UpperCAmelCase_ : List[Any] = word_embed_proj_dim UpperCAmelCase_ : Any = False def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) UpperCAmelCase_ : Optional[int] = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) UpperCAmelCase_ : Union[str, Any] = tf.concat([input_ids, eos_tensor] , axis=1 ) UpperCAmelCase_ : List[Any] = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=_snake_case , **self.config_updates , ) UpperCAmelCase_ : Any = prepare_opt_inputs_dict(_snake_case , _snake_case ) return config, inputs_dict def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = TFOPTModel(config=_snake_case ) UpperCAmelCase_ : Union[str, Any] = inputs_dict["input_ids"] UpperCAmelCase_ : Dict = input_ids[:1, :] UpperCAmelCase_ : int = inputs_dict["attention_mask"][:1, :] UpperCAmelCase_ : Optional[Any] = 1 # first forward pass UpperCAmelCase_ : Tuple = model(_snake_case , attention_mask=_snake_case , use_cache=_snake_case ) UpperCAmelCase_ , UpperCAmelCase_ : Any = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids UpperCAmelCase_ : List[Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) UpperCAmelCase_ : Dict = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and UpperCAmelCase_ : List[str] = tf.concat([input_ids, next_tokens] , axis=-1 ) UpperCAmelCase_ : str = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) UpperCAmelCase_ : str = model(_snake_case , attention_mask=_snake_case )[0] UpperCAmelCase_ : Dict = model(_snake_case , attention_mask=_snake_case , past_key_values=_snake_case )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice UpperCAmelCase_ : Optional[Any] = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) UpperCAmelCase_ : Any = output_from_no_past[:, -3:, random_slice_idx] UpperCAmelCase_ : List[str] = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_snake_case , _snake_case , rtol=1E-3 ) @require_tf class A_ (lowercase__ ,lowercase__ ,unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () SCREAMING_SNAKE_CASE__ : Tuple = (TFOPTForCausalLM,) if is_tf_available() else () SCREAMING_SNAKE_CASE__ : Union[str, Any] = ( {"""feature-extraction""": TFOPTModel, """text-generation""": TFOPTForCausalLM} if is_tf_available() else {} ) SCREAMING_SNAKE_CASE__ : Dict = False SCREAMING_SNAKE_CASE__ : Union[str, Any] = False SCREAMING_SNAKE_CASE__ : Optional[int] = False SCREAMING_SNAKE_CASE__ : Optional[int] = 10 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = TFOPTModelTester(self ) UpperCAmelCase_ : Optional[Any] = ConfigTester(self , config_class=_snake_case ) def UpperCamelCase__ ( self ): """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_snake_case ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : int = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(lowercase_ , lowercase_ ): if hasattr(_snake_case , "weight" ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(_snake_case , "weight" ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 10, config.vocab_size + 10]: # build the embeddings UpperCAmelCase_ : Union[str, Any] = model_class(config=_snake_case ) UpperCAmelCase_ : str = _get_word_embedding_weight(_snake_case , model.get_input_embeddings() ) UpperCAmelCase_ : int = _get_word_embedding_weight(_snake_case , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(_snake_case ) UpperCAmelCase_ : Dict = _get_word_embedding_weight(_snake_case , model.get_input_embeddings() ) UpperCAmelCase_ : List[Any] = _get_word_embedding_weight(_snake_case , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. UpperCAmelCase_ : Union[str, Any] = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , _snake_case ) # check that weights remain the same after resizing UpperCAmelCase_ : List[Any] = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: UpperCAmelCase_ : List[Any] = False self.assertTrue(_snake_case ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , _snake_case ) UpperCAmelCase_ : Optional[Any] = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: UpperCAmelCase_ : str = False self.assertTrue(_snake_case ) def __a ( __lowerCamelCase ): return tf.constant(a__, dtype=tf.intaa ) @require_tf class A_ (unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Any = 99 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = tf.ones((4, 1) , dtype=tf.intaa ) * 2 UpperCAmelCase_ : Any = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) UpperCAmelCase_ : List[str] = input_ids.shape[0] UpperCAmelCase_ : str = OPTConfig( vocab_size=self.vocab_size , hidden_size=24 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class A_ (unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[str] = TFOPTModel.from_pretrained("facebook/opt-350m" ) UpperCAmelCase_ : Any = _long_tensor([[0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2]] ) UpperCAmelCase_ : Union[str, Any] = tf.not_equal(_snake_case , model.config.pad_token_id ) with tf.GradientTape(): UpperCAmelCase_ : Tuple = model(input_ids=_snake_case , attention_mask=_snake_case ).last_hidden_state UpperCAmelCase_ : Any = (1, 11, 512) self.assertEqual(output.shape , _snake_case ) UpperCAmelCase_ : Tuple = tf.constant( [[-0.28_73, -1.92_18, -0.30_33], [-1.27_10, -0.13_38, -0.19_02], [0.40_95, 0.12_14, -1.31_21]] ) self.assertTrue(np.allclose(output[:, :3, :3] , _snake_case , atol=4E-3 ) ) UpperCAmelCase_ : Union[str, Any] = tf.function(_snake_case , jit_compile=_snake_case ) UpperCAmelCase_ : List[Any] = xla_generate(_snake_case , _snake_case )[0] self.assertTrue(np.allclose(output[:, :3, :3] , _snake_case , atol=4E-2 ) ) @require_tf @slow class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" super().setUp() UpperCAmelCase_ : Optional[int] = "facebook/opt-350m" def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Any = TFOPTForCausalLM.from_pretrained(self.path_model ) UpperCAmelCase_ : str = GPTaTokenizer.from_pretrained(self.path_model ) UpperCAmelCase_ : List[Any] = [ "Today is a beautiful day and I want to", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False UpperCAmelCase_ : Tuple = tokenizer(_snake_case , return_tensors="tf" , padding=_snake_case , add_special_tokens=_snake_case ) UpperCAmelCase_ : Union[str, Any] = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) UpperCAmelCase_ : List[str] = tf.constant( [ [1.38_51, -13.89_23, -10.52_29, -10.75_33, -0.23_09, -10.23_84, -0.53_65, -9.09_47, -5.16_70], [-4.70_73, -10.62_76, -3.94_15, -21.52_42, -0.28_22, -0.28_22, -0.28_22, -0.28_22, -0.28_22], [0.62_47, -3.42_29, -8.91_79, -1.42_97, -14.16_50, 1.41_46, -9.02_18, -0.27_03, -0.27_03], [6.47_83, -1.99_13, -10.79_26, -2.33_36, 1.50_92, -0.99_74, -6.82_13, 1.34_77, 1.34_77], ] ) self.assertTrue(np.allclose(_snake_case , _snake_case , atol=1E-4 ) ) UpperCAmelCase_ : List[str] = tf.function(_snake_case , jit_compile=_snake_case ) UpperCAmelCase_ : List[Any] = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(_snake_case , _snake_case , atol=1E-4 ) ) @require_tf @slow class A_ (unittest.TestCase ): '''simple docstring''' @property def UpperCamelCase__ ( self ): """simple docstring""" return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = "facebook/opt-125m" UpperCAmelCase_ : Union[str, Any] = [ "Today is a beautiful day and I want to", "In the city of New York, the city", "Paris is the capital of France and the capital", "Computers and mobile phones have taken over the", ] UpperCAmelCase_ : Union[str, Any] = [] UpperCAmelCase_ : str = GPTaTokenizer.from_pretrained(_snake_case ) UpperCAmelCase_ : Dict = TFOPTForCausalLM.from_pretrained(_snake_case ) for prompt in self.prompts: UpperCAmelCase_ : str = tokenizer(_snake_case , return_tensors="tf" ).input_ids UpperCAmelCase_ : Any = model.generate(_snake_case , max_length=10 ) UpperCAmelCase_ : Optional[int] = tokenizer.batch_decode(_snake_case , skip_special_tokens=_snake_case ) predicted_outputs += generated_string self.assertListEqual(_snake_case , _snake_case ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = "facebook/opt-350m" UpperCAmelCase_ : Dict = GPTaTokenizer.from_pretrained(_snake_case ) UpperCAmelCase_ : Tuple = TFOPTForCausalLM.from_pretrained(_snake_case ) UpperCAmelCase_ : Optional[Any] = "left" # use different length sentences to test batching UpperCAmelCase_ : Dict = [ "Hello, my dog is a little", "Today, I", ] UpperCAmelCase_ : List[str] = tokenizer(_snake_case , return_tensors="tf" , padding=_snake_case ) UpperCAmelCase_ : List[str] = inputs["input_ids"] UpperCAmelCase_ : Optional[int] = model.generate(input_ids=_snake_case , attention_mask=inputs["attention_mask"] ) UpperCAmelCase_ : Dict = tokenizer(sentences[0] , return_tensors="tf" ).input_ids UpperCAmelCase_ : Any = model.generate(input_ids=_snake_case ) UpperCAmelCase_ : Optional[int] = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs["attention_mask"][-1] , tf.intaa ) ) UpperCAmelCase_ : List[str] = tokenizer(sentences[1] , return_tensors="tf" ).input_ids UpperCAmelCase_ : List[Any] = model.generate(input_ids=_snake_case , max_length=model.config.max_length - num_paddings ) UpperCAmelCase_ : str = tokenizer.batch_decode(_snake_case , skip_special_tokens=_snake_case ) UpperCAmelCase_ : Optional[Any] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=_snake_case ) UpperCAmelCase_ : str = tokenizer.decode(output_padded[0] , skip_special_tokens=_snake_case ) UpperCAmelCase_ : Dict = [ "Hello, my dog is a little bit of a dork.\nI\'m a little bit", "Today, I was in the middle of a conversation with a friend about the", ] self.assertListEqual(_snake_case , _snake_case ) self.assertListEqual(_snake_case , [non_padded_sentence, padded_sentence] ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = "facebook/opt-350m" UpperCAmelCase_ : Union[str, Any] = [ "Today is a beautiful day and I want to", "In the city of San Francisco, the city", "Paris is the capital of France and the capital", "Computers and mobile phones have taken over the", ] UpperCAmelCase_ : Dict = [] UpperCAmelCase_ : Tuple = GPTaTokenizer.from_pretrained(_snake_case ) UpperCAmelCase_ : Dict = TFOPTForCausalLM.from_pretrained(_snake_case ) for prompt in self.prompts: UpperCAmelCase_ : List[Any] = tokenizer(_snake_case , return_tensors="tf" ).input_ids UpperCAmelCase_ : List[Any] = model.generate(_snake_case , max_length=10 ) UpperCAmelCase_ : Dict = tokenizer.batch_decode(_snake_case , skip_special_tokens=_snake_case ) predicted_outputs += generated_string self.assertListEqual(_snake_case , _snake_case )
360
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = {'ctrl': 'https://huggingface.co/ctrl/resolve/main/config.json'} class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """ctrl""" SCREAMING_SNAKE_CASE__ : Optional[int] = ["""past_key_values"""] SCREAMING_SNAKE_CASE__ : List[str] = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase_=24_6534 , lowercase_=256 , lowercase_=1280 , lowercase_=8192 , lowercase_=48 , lowercase_=16 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=1E-6 , lowercase_=0.02 , lowercase_=True , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Tuple = vocab_size UpperCAmelCase_ : Union[str, Any] = n_positions UpperCAmelCase_ : List[str] = n_embd UpperCAmelCase_ : Dict = n_layer UpperCAmelCase_ : Optional[int] = n_head UpperCAmelCase_ : List[str] = dff UpperCAmelCase_ : Tuple = resid_pdrop UpperCAmelCase_ : Optional[Any] = embd_pdrop UpperCAmelCase_ : str = layer_norm_epsilon UpperCAmelCase_ : List[str] = initializer_range UpperCAmelCase_ : List[str] = use_cache super().__init__(**lowercase_ )
23
0
"""simple docstring""" # this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since the output of this script is fed into Makefile commands it doesn't print a newline after the results import re import subprocess import sys _a = subprocess.check_output('git merge-base main HEAD'.split()).decode('utf-8') _a = subprocess.check_output(f"""git diff --name-only {fork_point_sha}""".split()).decode('utf-8').split() _a = '|'.join(sys.argv[1:]) _a = re.compile(Rf"""^({joined_dirs}).*?\.py$""") _a = [x for x in modified_files if regex.match(x)] print(' '.join(relevant_modified_files), end='')
361
"""simple docstring""" def __a ( __lowerCamelCase ): assert isinstance(__lowerCamelCase, __lowerCamelCase ), f"""The input value of [n={number}] is not an integer""" if number == 1: return 2 elif number < 1: UpperCAmelCase_ : str = f"""The input value of [n={number}] has to be > 0""" raise ValueError(__lowerCamelCase ) else: UpperCAmelCase_ : List[str] = sylvester(number - 1 ) UpperCAmelCase_ : List[str] = num - 1 UpperCAmelCase_ : List[str] = num return lower * upper + 1 if __name__ == "__main__": print(f"""The 8th number in Sylvester's sequence: {sylvester(8)}""")
23
0
import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, is_tf_available, is_torch_available, ) _a = logging.getLogger(__name__) @dataclass(frozen=lowercase__ ) class A_ : '''simple docstring''' SCREAMING_SNAKE_CASE__ : str SCREAMING_SNAKE_CASE__ : str SCREAMING_SNAKE_CASE__ : Optional[str] = None SCREAMING_SNAKE_CASE__ : Optional[str] = None SCREAMING_SNAKE_CASE__ : Optional[str] = None @dataclass(frozen=lowercase__ ) class A_ : '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[int] SCREAMING_SNAKE_CASE__ : Optional[List[int]] = None SCREAMING_SNAKE_CASE__ : Optional[List[int]] = None SCREAMING_SNAKE_CASE__ : Optional[Union[int, float]] = None SCREAMING_SNAKE_CASE__ : Optional[int] = None if is_torch_available(): import torch from torch.utils.data import Dataset class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[InputFeatures] def __init__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ = None , lowercase_=False , lowercase_ = False , ): """simple docstring""" UpperCAmelCase_ : Any = hans_processors[task]() UpperCAmelCase_ : Tuple = os.path.join( _a , "cached_{}_{}_{}_{}".format( "dev" if evaluate else "train" , tokenizer.__class__.__name__ , str(_a ) , _a , ) , ) UpperCAmelCase_ : Union[str, Any] = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) UpperCAmelCase_ , UpperCAmelCase_ : Any = label_list[2], label_list[1] UpperCAmelCase_ : Any = label_list # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. UpperCAmelCase_ : Optional[int] = cached_features_file + ".lock" with FileLock(_a ): if os.path.exists(_a ) and not overwrite_cache: logger.info(F"""Loading features from cached file {cached_features_file}""" ) UpperCAmelCase_ : str = torch.load(_a ) else: logger.info(F"""Creating features from dataset file at {data_dir}""" ) UpperCAmelCase_ : Optional[Any] = ( processor.get_dev_examples(_a ) if evaluate else processor.get_train_examples(_a ) ) logger.info("Training examples: %s" , len(_a ) ) UpperCAmelCase_ : List[str] = hans_convert_examples_to_features(_a , _a , _a , _a ) logger.info("Saving features into cached file %s" , _a ) torch.save(self.features , _a ) def __len__( self ): """simple docstring""" return len(self.features ) def __getitem__( self , lowercase_ ): """simple docstring""" return self.features[i] def UpperCamelCase__ ( self ): """simple docstring""" return self.label_list if is_tf_available(): import tensorflow as tf class A_ : '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[InputFeatures] def __init__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ = 128 , lowercase_=False , lowercase_ = False , ): """simple docstring""" UpperCAmelCase_ : Tuple = hans_processors[task]() UpperCAmelCase_ : Union[str, Any] = processor.get_labels() if tokenizer.__class__ in ( RobertaTokenizer, RobertaTokenizerFast, XLMRobertaTokenizer, BartTokenizer, BartTokenizerFast, ): # HACK(label indices are swapped in RoBERTa pretrained model) UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = label_list[2], label_list[1] UpperCAmelCase_ : Optional[int] = label_list UpperCAmelCase_ : int = processor.get_dev_examples(_a ) if evaluate else processor.get_train_examples(_a ) UpperCAmelCase_ : Tuple = hans_convert_examples_to_features(_a , _a , _a , _a ) def gen(): for ex_index, ex in tqdm.tqdm(enumerate(self.features ) , desc="convert examples to features" ): if ex_index % 1_0000 == 0: logger.info("Writing example %d of %d" % (ex_index, len(_a )) ) yield ( { "example_id": 0, "input_ids": ex.input_ids, "attention_mask": ex.attention_mask, "token_type_ids": ex.token_type_ids, }, ex.label, ) UpperCAmelCase_ : Any = tf.data.Dataset.from_generator( _a , ( { "example_id": tf.intaa, "input_ids": tf.intaa, "attention_mask": tf.intaa, "token_type_ids": tf.intaa, }, tf.intaa, ) , ( { "example_id": tf.TensorShape([] ), "input_ids": tf.TensorShape([None, None] ), "attention_mask": tf.TensorShape([None, None] ), "token_type_ids": tf.TensorShape([None, None] ), }, tf.TensorShape([] ), ) , ) def UpperCamelCase__ ( self ): """simple docstring""" return self.dataset def __len__( self ): """simple docstring""" return len(self.features ) def __getitem__( self , lowercase_ ): """simple docstring""" return self.features[i] def UpperCamelCase__ ( self ): """simple docstring""" return self.label_list class A_ (lowercase__ ): '''simple docstring''' def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self._create_examples(self._read_tsv(os.path.join(_a , "heuristics_train_set.txt" ) ) , "train" ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self._create_examples(self._read_tsv(os.path.join(_a , "heuristics_evaluation_set.txt" ) ) , "dev" ) def UpperCamelCase__ ( self ): """simple docstring""" return ["contradiction", "entailment", "neutral"] def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = [] for i, line in enumerate(_a ): if i == 0: continue UpperCAmelCase_ : Dict = "%s-%s" % (set_type, line[0]) UpperCAmelCase_ : List[Any] = line[5] UpperCAmelCase_ : int = line[6] UpperCAmelCase_ : Optional[Any] = line[7][2:] if line[7].startswith("ex" ) else line[7] UpperCAmelCase_ : str = line[0] examples.append(InputExample(guid=_a , text_a=_a , text_b=_a , label=_a , pairID=_a ) ) return examples def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, ): UpperCAmelCase_ : str = {label: i for i, label in enumerate(UpperCamelCase__ )} UpperCAmelCase_ : Any = [] for ex_index, example in tqdm.tqdm(enumerate(UpperCamelCase__ ), desc="convert examples to features" ): if ex_index % 1_0000 == 0: logger.info("Writing example %d" % (ex_index) ) UpperCAmelCase_ : Dict = tokenizer( example.text_a, example.text_b, add_special_tokens=UpperCamelCase__, max_length=UpperCamelCase__, padding="max_length", truncation=UpperCamelCase__, return_overflowing_tokens=UpperCamelCase__, ) UpperCAmelCase_ : List[Any] = label_map[example.label] if example.label in label_map else 0 UpperCAmelCase_ : Dict = int(example.pairID ) features.append(InputFeatures(**UpperCamelCase__, label=UpperCamelCase__, pairID=UpperCamelCase__ ) ) for i, example in enumerate(examples[:5] ): logger.info("*** Example ***" ) logger.info(f"""guid: {example}""" ) logger.info(f"""features: {features[i]}""" ) return features _a = { 'hans': 3, } _a = { 'hans': HansProcessor, }
362
"""simple docstring""" import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class A_ (lowercase__ ,lowercase__ ,unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = IFImgaImgSuperResolutionPipeline SCREAMING_SNAKE_CASE__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""width""", """height"""} SCREAMING_SNAKE_CASE__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""original_image"""} ) SCREAMING_SNAKE_CASE__ : List[Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} def UpperCamelCase__ ( self ): """simple docstring""" return self._get_superresolution_dummy_components() def UpperCamelCase__ ( self , lowercase_ , lowercase_=0 ): """simple docstring""" if str(lowercase_ ).startswith("mps" ): UpperCAmelCase_ : Optional[Any] = torch.manual_seed(lowercase_ ) else: UpperCAmelCase_ : Union[str, Any] = torch.Generator(device=lowercase_ ).manual_seed(lowercase_ ) UpperCAmelCase_ : Any = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) UpperCAmelCase_ : Optional[int] = floats_tensor((1, 3, 16, 16) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) UpperCAmelCase_ : int = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def UpperCamelCase__ ( self ): """simple docstring""" # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_save_load_local() def UpperCamelCase__ ( self ): """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
23
0
"""simple docstring""" # Copyright (c) 2021-, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #################################################################################################### # # Note: If when running this conversion script you're getting an exception: # ModuleNotFoundError: No module named 'megatron.model.enums' # you need to tell python where to find the clone of Megatron-LM, e.g.: # # cd /tmp # git clone https://github.com/NVIDIA/Megatron-LM # PYTHONPATH=/tmp/Megatron-LM python src/transformers/models/megatron_gpt2/convert_megatron_gpt2_checkpoint.py ... # # if you already have it cloned elsewhere, simply adjust the path to the existing path # # If the training was done using a Megatron-LM fork, e.g., # https://github.com/microsoft/Megatron-DeepSpeed/ then chances are that you need to have that one # in your path, i.e., /path/to/Megatron-DeepSpeed/ # import argparse import os import re import zipfile import torch from transformers import AutoTokenizer, GPTaConfig def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=0 ): # Format the message. if name is None: UpperCAmelCase_ : List[Any] = None else: UpperCAmelCase_ : Dict = "." * max(0, spaces - 2 ) + "# {:" + str(50 - spaces ) + "s}" UpperCAmelCase_ : Dict = fmt.format(_a ) # Print and recurse (if needed). if isinstance(_a, _a ): if msg is not None: print(_a ) for k in val.keys(): recursive_print(_a, val[k], spaces + 2 ) elif isinstance(_a, torch.Tensor ): print(_a, ":", val.size() ) else: print(_a, ":", _a ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # Permutes layout of param tensor to [num_splits * num_heads * hidden_size, :] # for compatibility with later versions of NVIDIA Megatron-LM. # The inverse operation is performed inside Megatron-LM to read checkpoints: # https://github.com/NVIDIA/Megatron-LM/blob/v2.4/megatron/checkpointing.py#L209 # If param is the weight tensor of the self-attention block, the returned tensor # will have to be transposed one more time to be read by HuggingFace GPT2. UpperCAmelCase_ : int = param.size() if checkpoint_version == 1.0: # version 1.0 stores [num_heads * hidden_size * num_splits, :] UpperCAmelCase_ : List[str] = (num_heads, hidden_size, num_splits) + input_shape[1:] UpperCAmelCase_ : Any = param.view(*_a ) UpperCAmelCase_ : Dict = param.transpose(0, 2 ) UpperCAmelCase_ : Optional[int] = param.transpose(1, 2 ).contiguous() elif checkpoint_version >= 2.0: # other versions store [num_heads * num_splits * hidden_size, :] UpperCAmelCase_ : Any = (num_heads, num_splits, hidden_size) + input_shape[1:] UpperCAmelCase_ : Union[str, Any] = param.view(*_a ) UpperCAmelCase_ : List[str] = param.transpose(0, 1 ).contiguous() UpperCAmelCase_ : int = param.view(*_a ) return param def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # The converted output model. UpperCAmelCase_ : str = {} # old versions did not store training args UpperCAmelCase_ : Tuple = input_state_dict.get("args", _a ) if ds_args is not None: # do not make the user write a config file when the exact dimensions/sizes are already in the checkpoint # from pprint import pprint # pprint(vars(ds_args)) UpperCAmelCase_ : Any = ds_args.padded_vocab_size UpperCAmelCase_ : List[Any] = ds_args.max_position_embeddings UpperCAmelCase_ : Any = ds_args.hidden_size UpperCAmelCase_ : Any = ds_args.num_layers UpperCAmelCase_ : Any = ds_args.num_attention_heads UpperCAmelCase_ : Optional[Any] = ds_args.ffn_hidden_size # pprint(config) # The number of heads. UpperCAmelCase_ : Dict = config.n_head # The hidden_size per head. UpperCAmelCase_ : List[str] = config.n_embd // config.n_head # Megatron-LM checkpoint version if "checkpoint_version" in input_state_dict.keys(): UpperCAmelCase_ : Optional[Any] = input_state_dict["checkpoint_version"] else: UpperCAmelCase_ : Optional[Any] = 0.0 # The model. UpperCAmelCase_ : Any = input_state_dict["model"] # The language model. UpperCAmelCase_ : List[str] = model["language_model"] # The embeddings. UpperCAmelCase_ : List[Any] = lm["embedding"] # The word embeddings. UpperCAmelCase_ : List[Any] = embeddings["word_embeddings"]["weight"] # Truncate the embedding table to vocab_size rows. UpperCAmelCase_ : Any = word_embeddings[: config.vocab_size, :] UpperCAmelCase_ : int = word_embeddings # The position embeddings. UpperCAmelCase_ : str = embeddings["position_embeddings"]["weight"] # Read the causal mask dimension (seqlen). [max_sequence_length, hidden_size] UpperCAmelCase_ : List[str] = pos_embeddings.size(0 ) if n_positions != config.n_positions: raise ValueError( f"""pos_embeddings.max_sequence_length={n_positions} and config.n_positions={config.n_positions} don\'t match""" ) # Store the position embeddings. UpperCAmelCase_ : Optional[int] = pos_embeddings # The transformer. UpperCAmelCase_ : str = lm["transformer"] if "transformer" in lm.keys() else lm["encoder"] # The regex to extract layer names. UpperCAmelCase_ : List[str] = re.compile(r"layers\.(\d+)\.([a-z0-9_.]+)\.([a-z]+)" ) # The simple map of names for "automated" rules. UpperCAmelCase_ : int = { "attention.dense": ".attn.c_proj.", "self_attention.dense": ".attn.c_proj.", "mlp.dense_h_to_4h": ".mlp.c_fc.", "mlp.dense_4h_to_h": ".mlp.c_proj.", } # Extract the layers. for key, val in transformer.items(): # Match the name. UpperCAmelCase_ : int = layer_re.match(_a ) # Stop if that's not a layer if m is None: break # The index of the layer. UpperCAmelCase_ : List[str] = int(m.group(1 ) ) # The name of the operation. UpperCAmelCase_ : Tuple = m.group(2 ) # Is it a weight or a bias? UpperCAmelCase_ : str = m.group(3 ) # The name of the layer. UpperCAmelCase_ : Optional[Any] = f"""transformer.h.{layer_idx}""" # For layernorm(s), simply store the layer norm. if op_name.endswith("layernorm" ): UpperCAmelCase_ : Optional[int] = "ln_1" if op_name.startswith("input" ) else "ln_2" UpperCAmelCase_ : int = val # Transpose the QKV matrix. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "weight": # Insert a tensor of 1x1xDxD bias. UpperCAmelCase_ : Union[str, Any] = torch.tril(torch.ones((n_positions, n_positions), dtype=torch.floataa ) ).view( 1, 1, _a, _a ) UpperCAmelCase_ : Dict = causal_mask # Insert a "dummy" tensor for masked_bias. UpperCAmelCase_ : Optional[int] = torch.tensor(-1E4, dtype=torch.floataa ) UpperCAmelCase_ : List[str] = masked_bias UpperCAmelCase_ : Union[str, Any] = fix_query_key_value_ordering(_a, _a, 3, _a, _a ) # Megatron stores (3*D) x D but transformers-GPT2 expects D x 3*D. UpperCAmelCase_ : Optional[int] = out_val.transpose(0, 1 ).contiguous() # Store. UpperCAmelCase_ : str = out_val # Transpose the bias. elif ( op_name == "attention.query_key_value" or op_name == "self_attention.query_key_value" ) and weight_or_bias == "bias": UpperCAmelCase_ : str = fix_query_key_value_ordering(_a, _a, 3, _a, _a ) # Store. No change of shape. UpperCAmelCase_ : Any = out_val # Transpose the weights. elif weight_or_bias == "weight": UpperCAmelCase_ : Dict = megatron_to_transformers[op_name] UpperCAmelCase_ : Any = val.transpose(0, 1 ) # Copy the bias. elif weight_or_bias == "bias": UpperCAmelCase_ : Dict = megatron_to_transformers[op_name] UpperCAmelCase_ : Dict = val # DEBUG. assert config.n_layer == layer_idx + 1 # The final layernorm. UpperCAmelCase_ : Any = transformer["final_layernorm.weight"] UpperCAmelCase_ : Optional[Any] = transformer["final_layernorm.bias"] # For LM head, transformers' wants the matrix to weight embeddings. UpperCAmelCase_ : Any = word_embeddings # It should be done! return output_state_dict def __a ( ): # Create the argument parser. UpperCAmelCase_ : Tuple = argparse.ArgumentParser() parser.add_argument("--print-checkpoint-structure", action="store_true" ) parser.add_argument( "path_to_checkpoint", type=_a, help="Path to the checkpoint file (.zip archive or direct .pt file)", ) parser.add_argument( "--config_file", default="", type=_a, help="An optional config json file describing the pre-trained model.", ) UpperCAmelCase_ : List[Any] = parser.parse_args() # Extract the basename. UpperCAmelCase_ : Any = os.path.dirname(args.path_to_checkpoint ) # Load the model. # the .zip is very optional, let's keep it for backward compatibility print(f"""Extracting PyTorch state dictionary from {args.path_to_checkpoint}""" ) if args.path_to_checkpoint.endswith(".zip" ): with zipfile.ZipFile(args.path_to_checkpoint, "r" ) as checkpoint: with checkpoint.open("release/mp_rank_00/model_optim_rng.pt" ) as pytorch_dict: UpperCAmelCase_ : str = torch.load(_a, map_location="cpu" ) else: UpperCAmelCase_ : List[str] = torch.load(args.path_to_checkpoint, map_location="cpu" ) UpperCAmelCase_ : Any = input_state_dict.get("args", _a ) # Read the config, or default to the model released by NVIDIA. if args.config_file == "": if ds_args is not None: if ds_args.bias_gelu_fusion: UpperCAmelCase_ : Any = "gelu_fast" elif ds_args.openai_gelu: UpperCAmelCase_ : Any = "gelu_new" else: UpperCAmelCase_ : List[Any] = "gelu" else: # in the very early days this used to be "gelu_new" UpperCAmelCase_ : Any = "gelu_new" # Spell out all parameters in case the defaults change. UpperCAmelCase_ : Tuple = GPTaConfig( vocab_size=5_0257, n_positions=1024, n_embd=1024, n_layer=24, n_head=16, n_inner=4096, activation_function=_a, resid_pdrop=0.1, embd_pdrop=0.1, attn_pdrop=0.1, layer_norm_epsilon=1E-5, initializer_range=0.02, summary_type="cls_index", summary_use_proj=_a, summary_activation=_a, summary_proj_to_labels=_a, summary_first_dropout=0.1, scale_attn_weights=_a, use_cache=_a, bos_token_id=5_0256, eos_token_id=5_0256, ) else: UpperCAmelCase_ : Dict = GPTaConfig.from_json_file(args.config_file ) UpperCAmelCase_ : Optional[Any] = ["GPT2LMHeadModel"] # Convert. print("Converting" ) UpperCAmelCase_ : List[str] = convert_megatron_checkpoint(_a, _a, _a ) # Print the structure of converted state dict. if args.print_checkpoint_structure: recursive_print(_a, _a ) # Add tokenizer class info to config # see https://github.com/huggingface/transformers/issues/13906) if ds_args is not None: UpperCAmelCase_ : Optional[Any] = ds_args.tokenizer_type if tokenizer_type == "GPT2BPETokenizer": UpperCAmelCase_ : List[Any] = "gpt2" elif tokenizer_type == "PretrainedFromHF": UpperCAmelCase_ : List[Any] = ds_args.tokenizer_name_or_path else: raise ValueError(f"""Unrecognized tokenizer_type {tokenizer_type}""" ) else: UpperCAmelCase_ : Optional[Any] = "gpt2" UpperCAmelCase_ : Any = AutoTokenizer.from_pretrained(_a ) UpperCAmelCase_ : int = type(_a ).__name__ UpperCAmelCase_ : str = tokenizer_class # Store the config to file. print("Saving config" ) config.save_pretrained(_a ) # Save tokenizer based on args print(f"""Adding {tokenizer_class} tokenizer files""" ) tokenizer.save_pretrained(_a ) # Store the state_dict to file. UpperCAmelCase_ : Dict = os.path.join(_a, "pytorch_model.bin" ) print(f"""Saving checkpoint to \"{output_checkpoint_file}\"""" ) torch.save(_a, _a ) #################################################################################################### if __name__ == "__main__": main() ####################################################################################################
363
"""simple docstring""" import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = "ylacombe/bark-small" UpperCAmelCase_ : Union[str, Any] = tempfile.mkdtemp() UpperCAmelCase_ : List[str] = "en_speaker_1" UpperCAmelCase_ : Tuple = "This is a test string" UpperCAmelCase_ : List[Any] = "speaker_embeddings_path.json" UpperCAmelCase_ : Any = "speaker_embeddings" def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" return AutoTokenizer.from_pretrained(self.checkpoint , **lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.get_tokenizer() UpperCAmelCase_ : Union[str, Any] = BarkProcessor(tokenizer=lowercase_ ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase_ : Optional[int] = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) UpperCAmelCase_ : Dict = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) UpperCAmelCase_ : Union[str, Any] = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token="(BOS)" , eos_token="(EOS)" , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) UpperCAmelCase_ : int = 35 UpperCAmelCase_ : Optional[Any] = 2 UpperCAmelCase_ : List[Any] = 8 UpperCAmelCase_ : Optional[Any] = { "semantic_prompt": np.ones(lowercase_ ), "coarse_prompt": np.ones((nb_codebooks_coarse, seq_len) ), "fine_prompt": np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset UpperCAmelCase_ : Dict = processor(text=self.input_string , voice_preset=lowercase_ ) UpperCAmelCase_ : List[str] = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowercase_ , np.array([] ) ).tolist() ) # test loading voice preset from npz file UpperCAmelCase_ : Tuple = os.path.join(self.tmpdirname , "file.npz" ) np.savez(lowercase_ , **lowercase_ ) UpperCAmelCase_ : Optional[int] = processor(text=self.input_string , voice_preset=lowercase_ ) UpperCAmelCase_ : List[str] = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowercase_ , np.array([] ) ).tolist() ) # test loading voice preset from the hub UpperCAmelCase_ : Tuple = processor(text=self.input_string , voice_preset=self.voice_preset ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = self.get_tokenizer() UpperCAmelCase_ : Optional[Any] = BarkProcessor(tokenizer=lowercase_ ) UpperCAmelCase_ : Tuple = processor(text=self.input_string ) UpperCAmelCase_ : Union[str, Any] = tokenizer( self.input_string , padding="max_length" , max_length=256 , add_special_tokens=lowercase_ , return_attention_mask=lowercase_ , return_token_type_ids=lowercase_ , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
23
0
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { "facebook/wav2vec2-base-960h": "https://huggingface.co/facebook/wav2vec2-base-960h/resolve/main/config.json", # See all Wav2Vec2 models at https://huggingface.co/models?filter=wav2vec2 } class A_ (snake_case__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = """wav2vec2""" def __init__( self , lowercase_=32 , lowercase_=768 , lowercase_=12 , lowercase_=12 , lowercase_=3072 , lowercase_="gelu" , lowercase_=0.1 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=0.02 , lowercase_=1E-5 , lowercase_="group" , lowercase_="gelu" , lowercase_=(512, 512, 512, 512, 512, 512, 512) , lowercase_=(5, 2, 2, 2, 2, 2, 2) , lowercase_=(10, 3, 3, 3, 3, 2, 2) , lowercase_=False , lowercase_=128 , lowercase_=16 , lowercase_=False , lowercase_=True , lowercase_=0.05 , lowercase_=10 , lowercase_=2 , lowercase_=0.0 , lowercase_=10 , lowercase_=0 , lowercase_=320 , lowercase_=2 , lowercase_=0.1 , lowercase_=100 , lowercase_=256 , lowercase_=256 , lowercase_=0.1 , lowercase_="sum" , lowercase_=False , lowercase_=False , lowercase_=256 , lowercase_=(512, 512, 512, 512, 1500) , lowercase_=(5, 3, 3, 1, 1) , lowercase_=(1, 2, 3, 1, 1) , lowercase_=512 , lowercase_=0 , lowercase_=1 , lowercase_=2 , lowercase_=False , lowercase_=3 , lowercase_=2 , lowercase_=3 , lowercase_=None , lowercase_=None , **lowercase_ , ): """simple docstring""" super().__init__(**UpperCAmelCase_ , pad_token_id=UpperCAmelCase_ , bos_token_id=UpperCAmelCase_ , eos_token_id=UpperCAmelCase_ ) UpperCAmelCase_ : Tuple = hidden_size UpperCAmelCase_ : List[str] = feat_extract_norm UpperCAmelCase_ : Any = feat_extract_activation UpperCAmelCase_ : List[str] = list(UpperCAmelCase_ ) UpperCAmelCase_ : Any = list(UpperCAmelCase_ ) UpperCAmelCase_ : List[str] = list(UpperCAmelCase_ ) UpperCAmelCase_ : Union[str, Any] = conv_bias UpperCAmelCase_ : str = num_conv_pos_embeddings UpperCAmelCase_ : str = num_conv_pos_embedding_groups UpperCAmelCase_ : Optional[int] = len(self.conv_dim ) UpperCAmelCase_ : Any = num_hidden_layers UpperCAmelCase_ : List[str] = intermediate_size UpperCAmelCase_ : str = hidden_act UpperCAmelCase_ : int = num_attention_heads UpperCAmelCase_ : Optional[int] = hidden_dropout UpperCAmelCase_ : int = attention_dropout UpperCAmelCase_ : List[Any] = activation_dropout UpperCAmelCase_ : Tuple = feat_proj_dropout UpperCAmelCase_ : Union[str, Any] = final_dropout UpperCAmelCase_ : str = layerdrop UpperCAmelCase_ : Optional[Any] = layer_norm_eps UpperCAmelCase_ : Tuple = initializer_range UpperCAmelCase_ : Any = vocab_size UpperCAmelCase_ : Tuple = do_stable_layer_norm UpperCAmelCase_ : List[Any] = use_weighted_layer_sum if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( "Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==" " `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =" F""" {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,""" F""" `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCAmelCase_ : str = apply_spec_augment UpperCAmelCase_ : List[str] = mask_time_prob UpperCAmelCase_ : List[Any] = mask_time_length UpperCAmelCase_ : Dict = mask_time_min_masks UpperCAmelCase_ : Tuple = mask_feature_prob UpperCAmelCase_ : Dict = mask_feature_length UpperCAmelCase_ : int = mask_feature_min_masks # parameters for pretraining with codevector quantized representations UpperCAmelCase_ : Union[str, Any] = num_codevectors_per_group UpperCAmelCase_ : Any = num_codevector_groups UpperCAmelCase_ : List[Any] = contrastive_logits_temperature UpperCAmelCase_ : int = feat_quantizer_dropout UpperCAmelCase_ : Union[str, Any] = num_negatives UpperCAmelCase_ : Optional[Any] = codevector_dim UpperCAmelCase_ : Any = proj_codevector_dim UpperCAmelCase_ : str = diversity_loss_weight # ctc loss UpperCAmelCase_ : Dict = ctc_loss_reduction UpperCAmelCase_ : Any = ctc_zero_infinity # adapter UpperCAmelCase_ : Union[str, Any] = add_adapter UpperCAmelCase_ : int = adapter_kernel_size UpperCAmelCase_ : Tuple = adapter_stride UpperCAmelCase_ : Dict = num_adapter_layers UpperCAmelCase_ : Any = output_hidden_size or hidden_size UpperCAmelCase_ : Dict = adapter_attn_dim # SequenceClassification-specific parameter. Feel free to ignore for other classes. UpperCAmelCase_ : Tuple = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. UpperCAmelCase_ : Tuple = list(UpperCAmelCase_ ) UpperCAmelCase_ : Union[str, Any] = list(UpperCAmelCase_ ) UpperCAmelCase_ : Tuple = list(UpperCAmelCase_ ) UpperCAmelCase_ : List[str] = xvector_output_dim @property def UpperCamelCase__ ( self ): """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
364
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def __a ( __lowerCamelCase, __lowerCamelCase=False ): UpperCAmelCase_ : Optional[int] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""deit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""deit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""deit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""deit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""deit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""deit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""deit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""deit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""deit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""deit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "deit.embeddings.cls_token"), ("dist_token", "deit.embeddings.distillation_token"), ("patch_embed.proj.weight", "deit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "deit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "deit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" UpperCAmelCase_ : Dict = [(pair[0], pair[1][4:]) if pair[1].startswith("deit" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("norm.weight", "deit.layernorm.weight"), ("norm.bias", "deit.layernorm.bias"), ("head.weight", "cls_classifier.weight"), ("head.bias", "cls_classifier.bias"), ("head_dist.weight", "distillation_classifier.weight"), ("head_dist.bias", "distillation_classifier.bias"), ] ) return rename_keys def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=False ): for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase_ : int = "" else: UpperCAmelCase_ : Union[str, Any] = "deit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase_ : Tuple = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) UpperCAmelCase_ : Dict = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ : Union[str, Any] = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase_ : Any = in_proj_bias[: config.hidden_size] UpperCAmelCase_ : Optional[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase_ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase_ : List[Any] = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase_ : str = in_proj_bias[-config.hidden_size :] def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Tuple = dct.pop(__lowerCamelCase ) UpperCAmelCase_ : Tuple = val def __a ( ): UpperCAmelCase_ : Union[str, Any] = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ : str = Image.open(requests.get(__lowerCamelCase, stream=__lowerCamelCase ).raw ) return im @torch.no_grad() def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : List[str] = DeiTConfig() # all deit models have fine-tuned heads UpperCAmelCase_ : Optional[int] = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase_ : Tuple = 1000 UpperCAmelCase_ : str = "huggingface/label-files" UpperCAmelCase_ : str = "imagenet-1k-id2label.json" UpperCAmelCase_ : List[Any] = json.load(open(hf_hub_download(__lowerCamelCase, __lowerCamelCase, repo_type="dataset" ), "r" ) ) UpperCAmelCase_ : List[str] = {int(__lowerCamelCase ): v for k, v in idalabel.items()} UpperCAmelCase_ : Any = idalabel UpperCAmelCase_ : int = {v: k for k, v in idalabel.items()} UpperCAmelCase_ : Any = int(deit_name[-6:-4] ) UpperCAmelCase_ : Dict = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("tiny" ): UpperCAmelCase_ : Any = 192 UpperCAmelCase_ : Union[str, Any] = 768 UpperCAmelCase_ : Union[str, Any] = 12 UpperCAmelCase_ : int = 3 elif deit_name[9:].startswith("small" ): UpperCAmelCase_ : List[str] = 384 UpperCAmelCase_ : List[str] = 1536 UpperCAmelCase_ : Dict = 12 UpperCAmelCase_ : Any = 6 if deit_name[9:].startswith("base" ): pass elif deit_name[4:].startswith("large" ): UpperCAmelCase_ : int = 1024 UpperCAmelCase_ : List[Any] = 4096 UpperCAmelCase_ : Optional[int] = 24 UpperCAmelCase_ : int = 16 # load original model from timm UpperCAmelCase_ : Union[str, Any] = timm.create_model(__lowerCamelCase, pretrained=__lowerCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCAmelCase_ : Optional[Any] = timm_model.state_dict() UpperCAmelCase_ : Tuple = create_rename_keys(__lowerCamelCase, __lowerCamelCase ) for src, dest in rename_keys: rename_key(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) read_in_q_k_v(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) # load HuggingFace model UpperCAmelCase_ : str = DeiTForImageClassificationWithTeacher(__lowerCamelCase ).eval() model.load_state_dict(__lowerCamelCase ) # Check outputs on an image, prepared by DeiTImageProcessor UpperCAmelCase_ : Union[str, Any] = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 UpperCAmelCase_ : Optional[Any] = DeiTImageProcessor(size=__lowerCamelCase, crop_size=config.image_size ) UpperCAmelCase_ : Any = image_processor(images=prepare_img(), return_tensors="pt" ) UpperCAmelCase_ : int = encoding["pixel_values"] UpperCAmelCase_ : Optional[Any] = model(__lowerCamelCase ) UpperCAmelCase_ : Any = timm_model(__lowerCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__lowerCamelCase, outputs.logits, atol=1E-3 ) Path(__lowerCamelCase ).mkdir(exist_ok=__lowerCamelCase ) print(f"""Saving model {deit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCamelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--deit_name', default='vit_deit_base_distilled_patch16_224', type=str, help='Name of the DeiT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _a = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
23
0
"""simple docstring""" def __a ( __lowerCamelCase, __lowerCamelCase ): def get_matched_characters(__lowerCamelCase, __lowerCamelCase ) -> str: UpperCAmelCase_ : Tuple = [] UpperCAmelCase_ : Optional[Any] = min(len(_stra ), len(_stra ) ) // 2 for i, l in enumerate(_stra ): UpperCAmelCase_ : Optional[Any] = int(max(0, i - limit ) ) UpperCAmelCase_ : int = int(min(i + limit + 1, len(_stra ) ) ) if l in _stra[left:right]: matched.append(SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : int = f"""{_stra[0:_stra.index(SCREAMING_SNAKE_CASE_ )]} {_stra[_stra.index(SCREAMING_SNAKE_CASE_ ) + 1:]}""" return "".join(SCREAMING_SNAKE_CASE_ ) # matching characters UpperCAmelCase_ : List[str] = get_matched_characters(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : Optional[Any] = get_matched_characters(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) UpperCAmelCase_ : Optional[int] = len(SCREAMING_SNAKE_CASE_ ) # transposition UpperCAmelCase_ : int = ( len([(ca, ca) for ca, ca in zip(SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_ ) if ca != ca] ) // 2 ) if not match_count: UpperCAmelCase_ : str = 0.0 else: UpperCAmelCase_ : Any = ( 1 / 3 * ( match_count / len(SCREAMING_SNAKE_CASE_ ) + match_count / len(SCREAMING_SNAKE_CASE_ ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters UpperCAmelCase_ : str = 0 for ca, ca in zip(stra[:4], stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler('hello', 'world'))
365
"""simple docstring""" import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights UpperCAmelCase_ : Optional[Any] = FlaxDiffusionPipeline.from_pretrained( "hf-internal-testing/tiny-stable-diffusion-pipe" , safety_checker=lowercase_ , cache_dir=lowercase_ ) UpperCAmelCase_ : List[Any] = [t[-1] for t in os.walk(os.path.join(lowercase_ , os.listdir(lowercase_ )[0] , "snapshots" ) )] UpperCAmelCase_ : Dict = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith(".bin" ) for f in files ) @slow @require_flax class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( "hf-internal-testing/tiny-stable-diffusion-pipe" , safety_checker=lowercase_ ) UpperCAmelCase_ : Tuple = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : List[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase_ : List[str] = 4 UpperCAmelCase_ : Tuple = jax.device_count() UpperCAmelCase_ : Optional[int] = num_samples * [prompt] UpperCAmelCase_ : List[Any] = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : int = replicate(lowercase_ ) UpperCAmelCase_ : str = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : Dict = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_51_47_45 ) < 1E-3 assert np.abs(np.abs(lowercase_ , dtype=np.floataa ).sum() - 4_99_47.8_75 ) < 5E-1 UpperCAmelCase_ : List[Any] = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(lowercase_ ) == num_samples def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="flax" , safety_checker=lowercase_ ) UpperCAmelCase_ : Optional[int] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : str = jax.random.PRNGKey(0 ) UpperCAmelCase_ : Union[str, Any] = 50 UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[str] = num_samples * [prompt] UpperCAmelCase_ : Union[str, Any] = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Any = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : int = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_65_24_01) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_38_38_08.2) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ ) UpperCAmelCase_ : Any = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : str = jax.random.PRNGKey(0 ) UpperCAmelCase_ : str = 50 UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : Any = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Dict = replicate(lowercase_ ) UpperCAmelCase_ : str = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Union[str, Any] = shard(lowercase_ ) UpperCAmelCase_ : List[Any] = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : str = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa ) UpperCAmelCase_ : List[Any] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : Dict = jax.random.PRNGKey(0 ) UpperCAmelCase_ : Optional[int] = 50 UpperCAmelCase_ : Optional[int] = jax.device_count() UpperCAmelCase_ : str = num_samples * [prompt] UpperCAmelCase_ : int = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Union[str, Any] = replicate(lowercase_ ) UpperCAmelCase_ : Union[str, Any] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[int] = shard(lowercase_ ) UpperCAmelCase_ : Any = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = FlaxDDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="scaled_linear" , set_alpha_to_one=lowercase_ , steps_offset=1 , ) UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , scheduler=lowercase_ , safety_checker=lowercase_ , ) UpperCAmelCase_ : List[Any] = scheduler.create_state() UpperCAmelCase_ : int = scheduler_state UpperCAmelCase_ : Union[str, Any] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : Optional[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase_ : int = 50 UpperCAmelCase_ : str = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : int = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : int = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = shard(lowercase_ ) UpperCAmelCase_ : Any = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_45_04_39_45) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_34_76_93.5) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : Union[str, Any] = jax.random.split(jax.random.PRNGKey(0 ) , lowercase_ ) UpperCAmelCase_ , UpperCAmelCase_ : str = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ , ) UpperCAmelCase_ : Any = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = pipeline.prepare_inputs(lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : List[Any] = pipeline(lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase_ : int = images[2, 0, 256, 10:17, 1] # With memory efficient attention UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ , use_memory_efficient_attention=lowercase_ , ) UpperCAmelCase_ : str = replicate(lowercase_ ) UpperCAmelCase_ : str = pipeline.prepare_inputs(lowercase_ ) UpperCAmelCase_ : Optional[int] = shard(lowercase_ ) UpperCAmelCase_ : str = pipeline(lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase_ : Optional[int] = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1E-2
23
0
"""simple docstring""" from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
366
"""simple docstring""" from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean _a = 0 _a = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] _a = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right _a = tuple[int, int] class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ): """simple docstring""" UpperCAmelCase_ : int = pos_x UpperCAmelCase_ : List[Any] = pos_y UpperCAmelCase_ : Union[str, Any] = (pos_y, pos_x) UpperCAmelCase_ : Any = goal_x UpperCAmelCase_ : Dict = goal_y UpperCAmelCase_ : Any = g_cost UpperCAmelCase_ : List[str] = parent UpperCAmelCase_ : int = self.calculate_heuristic() UpperCAmelCase_ : Any = self.g_cost + self.h_cost def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.pos_x - self.goal_x UpperCAmelCase_ : Union[str, Any] = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowercase_ ) + abs(lowercase_ ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self , lowercase_ ): """simple docstring""" return self.f_cost < other.f_cost class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Tuple = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , lowercase_ ) UpperCAmelCase_ : List[Any] = Node(goal[1] , goal[0] , goal[1] , goal[0] , 9_9999 , lowercase_ ) UpperCAmelCase_ : str = [self.start] UpperCAmelCase_ : list[Node] = [] UpperCAmelCase_ : int = False def UpperCamelCase__ ( self ): """simple docstring""" while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() UpperCAmelCase_ : List[str] = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(lowercase_ ) self.closed_nodes.append(lowercase_ ) UpperCAmelCase_ : str = self.get_successors(lowercase_ ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowercase_ ) else: # retrieve the best current path UpperCAmelCase_ : Union[str, Any] = self.open_nodes.pop(self.open_nodes.index(lowercase_ ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowercase_ ) else: self.open_nodes.append(lowercase_ ) return [self.start.pos] def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Any = [] for action in delta: UpperCAmelCase_ : str = parent.pos_x + action[1] UpperCAmelCase_ : int = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowercase_ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowercase_ , lowercase_ , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , lowercase_ , ) ) return successors def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = node UpperCAmelCase_ : int = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) UpperCAmelCase_ : Optional[int] = current_node.parent path.reverse() return path class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Dict = AStar(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = AStar(lowercase_ , lowercase_ ) UpperCAmelCase_ : Tuple = False def UpperCamelCase__ ( self ): """simple docstring""" while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() UpperCAmelCase_ : List[str] = self.fwd_astar.open_nodes.pop(0 ) UpperCAmelCase_ : List[Any] = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowercase_ , lowercase_ ) self.fwd_astar.closed_nodes.append(lowercase_ ) self.bwd_astar.closed_nodes.append(lowercase_ ) UpperCAmelCase_ : Tuple = current_bwd_node UpperCAmelCase_ : str = current_fwd_node UpperCAmelCase_ : Dict = { self.fwd_astar: self.fwd_astar.get_successors(lowercase_ ), self.bwd_astar: self.bwd_astar.get_successors(lowercase_ ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowercase_ ) else: # retrieve the best current path UpperCAmelCase_ : List[Any] = astar.open_nodes.pop( astar.open_nodes.index(lowercase_ ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowercase_ ) else: astar.open_nodes.append(lowercase_ ) return [self.fwd_astar.start.pos] def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.fwd_astar.retrace_path(lowercase_ ) UpperCAmelCase_ : int = self.bwd_astar.retrace_path(lowercase_ ) bwd_path.pop() bwd_path.reverse() UpperCAmelCase_ : Any = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] _a = (0, 0) _a = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) _a = time.time() _a = AStar(init, goal) _a = a_star.search() _a = time.time() - start_time print(f"""AStar execution time = {end_time:f} seconds""") _a = time.time() _a = BidirectionalAStar(init, goal) _a = time.time() - bd_start_time print(f"""BidirectionalAStar execution time = {bd_end_time:f} seconds""")
23
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = { 'facebook/timesformer': 'https://huggingface.co/facebook/timesformer/resolve/main/config.json', } class A_ (A_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """timesformer""" def __init__( self , lowercase_=224 , lowercase_=16 , lowercase_=3 , lowercase_=8 , lowercase_=768 , lowercase_=12 , lowercase_=12 , lowercase_=3072 , lowercase_="gelu" , lowercase_=0.0 , lowercase_=0.0 , lowercase_=0.02 , lowercase_=1E-6 , lowercase_=True , lowercase_="divided_space_time" , lowercase_=0 , **lowercase_ , ): """simple docstring""" super().__init__(**_lowerCamelCase ) UpperCAmelCase_ : int = image_size UpperCAmelCase_ : Dict = patch_size UpperCAmelCase_ : List[str] = num_channels UpperCAmelCase_ : Optional[Any] = num_frames UpperCAmelCase_ : str = hidden_size UpperCAmelCase_ : Any = num_hidden_layers UpperCAmelCase_ : Optional[int] = num_attention_heads UpperCAmelCase_ : List[str] = intermediate_size UpperCAmelCase_ : int = hidden_act UpperCAmelCase_ : int = hidden_dropout_prob UpperCAmelCase_ : List[Any] = attention_probs_dropout_prob UpperCAmelCase_ : Union[str, Any] = initializer_range UpperCAmelCase_ : Any = layer_norm_eps UpperCAmelCase_ : Optional[int] = qkv_bias UpperCAmelCase_ : Dict = attention_type UpperCAmelCase_ : Dict = drop_path_rate
367
"""simple docstring""" import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : str = (PNDMScheduler,) SCREAMING_SNAKE_CASE__ : str = (("""num_inference_steps""", 50),) def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : int = { "num_train_timesteps": 1000, "beta_start": 0.00_01, "beta_end": 0.02, "beta_schedule": "linear", } config.update(**lowercase_ ) return config def UpperCamelCase__ ( self , lowercase_=0 , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = dict(self.forward_default_kwargs ) UpperCAmelCase_ : List[str] = kwargs.pop("num_inference_steps" , lowercase_ ) UpperCAmelCase_ : Union[str, Any] = self.dummy_sample UpperCAmelCase_ : Dict = 0.1 * sample UpperCAmelCase_ : Dict = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : List[Any] = self.get_scheduler_config(**lowercase_ ) UpperCAmelCase_ : Dict = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals UpperCAmelCase_ : List[Any] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) UpperCAmelCase_ : Optional[int] = scheduler_class.from_pretrained(lowercase_ ) new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals UpperCAmelCase_ : int = dummy_past_residuals[:] UpperCAmelCase_ : List[str] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : str = new_scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" UpperCAmelCase_ : Optional[int] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Dict = new_scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCamelCase__ ( self ): """simple docstring""" pass def UpperCamelCase__ ( self , lowercase_=0 , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = dict(self.forward_default_kwargs ) UpperCAmelCase_ : str = kwargs.pop("num_inference_steps" , lowercase_ ) UpperCAmelCase_ : Optional[int] = self.dummy_sample UpperCAmelCase_ : List[str] = 0.1 * sample UpperCAmelCase_ : Tuple = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : str = self.get_scheduler_config() UpperCAmelCase_ : Dict = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # copy over dummy past residuals (must be after setting timesteps) UpperCAmelCase_ : List[Any] = dummy_past_residuals[:] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(lowercase_ ) UpperCAmelCase_ : Dict = scheduler_class.from_pretrained(lowercase_ ) # copy over dummy past residuals new_scheduler.set_timesteps(lowercase_ ) # copy over dummy past residual (must be after setting timesteps) UpperCAmelCase_ : Optional[Any] = dummy_past_residuals[:] UpperCAmelCase_ : Union[str, Any] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Dict = new_scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" UpperCAmelCase_ : List[str] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : int = new_scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ , **lowercase_ ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1E-5, "Scheduler outputs are not identical" def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = self.scheduler_classes[0] UpperCAmelCase_ : Union[str, Any] = self.get_scheduler_config(**lowercase_ ) UpperCAmelCase_ : List[Any] = scheduler_class(**lowercase_ ) UpperCAmelCase_ : Tuple = 10 UpperCAmelCase_ : List[str] = self.dummy_model() UpperCAmelCase_ : str = self.dummy_sample_deter scheduler.set_timesteps(lowercase_ ) for i, t in enumerate(scheduler.prk_timesteps ): UpperCAmelCase_ : Tuple = model(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[int] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ ).prev_sample for i, t in enumerate(scheduler.plms_timesteps ): UpperCAmelCase_ : Any = model(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , lowercase_ , lowercase_ ).prev_sample return sample def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = dict(self.forward_default_kwargs ) UpperCAmelCase_ : Optional[Any] = kwargs.pop("num_inference_steps" , lowercase_ ) for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : Any = self.get_scheduler_config() UpperCAmelCase_ : Tuple = scheduler_class(**lowercase_ ) UpperCAmelCase_ : str = self.dummy_sample UpperCAmelCase_ : List[Any] = 0.1 * sample if num_inference_steps is not None and hasattr(lowercase_ , "set_timesteps" ): scheduler.set_timesteps(lowercase_ ) elif num_inference_steps is not None and not hasattr(lowercase_ , "set_timesteps" ): UpperCAmelCase_ : List[Any] = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) UpperCAmelCase_ : List[str] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] UpperCAmelCase_ : List[str] = dummy_past_residuals[:] UpperCAmelCase_ : str = scheduler.step_prk(lowercase_ , 0 , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Any = scheduler.step_prk(lowercase_ , 1 , lowercase_ , **lowercase_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , 0 , lowercase_ , **lowercase_ ).prev_sample UpperCAmelCase_ : Optional[Any] = scheduler.step_plms(lowercase_ , 1 , lowercase_ , **lowercase_ ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def UpperCamelCase__ ( self ): """simple docstring""" for timesteps in [100, 1000]: self.check_over_configs(num_train_timesteps=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for steps_offset in [0, 1]: self.check_over_configs(steps_offset=lowercase_ ) UpperCAmelCase_ : Optional[int] = self.scheduler_classes[0] UpperCAmelCase_ : int = self.get_scheduler_config(steps_offset=1 ) UpperCAmelCase_ : Optional[Any] = scheduler_class(**lowercase_ ) scheduler.set_timesteps(10 ) assert torch.equal( scheduler.timesteps , torch.LongTensor( [901, 851, 851, 801, 801, 751, 751, 701, 701, 651, 651, 601, 601, 501, 401, 301, 201, 101, 1] ) , ) def UpperCamelCase__ ( self ): """simple docstring""" for beta_start, beta_end in zip([0.00_01, 0.0_01] , [0.0_02, 0.02] ): self.check_over_configs(beta_start=lowercase_ , beta_end=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for t in [1, 5, 10]: self.check_over_forward(time_step=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" for t, num_inference_steps in zip([1, 5, 10] , [10, 50, 100] ): self.check_over_forward(num_inference_steps=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" # earlier version of set_timesteps() caused an error indexing alpha's with inference steps as power of 3 UpperCAmelCase_ : List[Any] = 27 for scheduler_class in self.scheduler_classes: UpperCAmelCase_ : List[Any] = self.dummy_sample UpperCAmelCase_ : Optional[int] = 0.1 * sample UpperCAmelCase_ : List[str] = self.get_scheduler_config() UpperCAmelCase_ : List[str] = scheduler_class(**lowercase_ ) scheduler.set_timesteps(lowercase_ ) # before power of 3 fix, would error on first step, so we only need to do two for i, t in enumerate(scheduler.prk_timesteps[:2] ): UpperCAmelCase_ : List[str] = scheduler.step_prk(lowercase_ , lowercase_ , lowercase_ ).prev_sample def UpperCamelCase__ ( self ): """simple docstring""" with self.assertRaises(lowercase_ ): UpperCAmelCase_ : List[str] = self.scheduler_classes[0] UpperCAmelCase_ : str = self.get_scheduler_config() UpperCAmelCase_ : Tuple = scheduler_class(**lowercase_ ) scheduler.step_plms(self.dummy_sample , 1 , self.dummy_sample ).prev_sample def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.full_loop() UpperCAmelCase_ : Any = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Dict = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 1_98.13_18 ) < 1E-2 assert abs(result_mean.item() - 0.25_80 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = self.full_loop(prediction_type="v_prediction" ) UpperCAmelCase_ : str = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Tuple = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 67.39_86 ) < 1E-2 assert abs(result_mean.item() - 0.08_78 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase_ : Union[str, Any] = self.full_loop(set_alpha_to_one=lowercase_ , beta_start=0.01 ) UpperCAmelCase_ : List[Any] = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : int = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 2_30.03_99 ) < 1E-2 assert abs(result_mean.item() - 0.29_95 ) < 1E-3 def UpperCamelCase__ ( self ): """simple docstring""" # We specify different beta, so that the first alpha is 0.99 UpperCAmelCase_ : Tuple = self.full_loop(set_alpha_to_one=lowercase_ , beta_start=0.01 ) UpperCAmelCase_ : int = torch.sum(torch.abs(lowercase_ ) ) UpperCAmelCase_ : Tuple = torch.mean(torch.abs(lowercase_ ) ) assert abs(result_sum.item() - 1_86.94_82 ) < 1E-2 assert abs(result_mean.item() - 0.24_34 ) < 1E-3
23
0
import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class A_ (lowercase__ ,lowercase__ ): '''simple docstring''' @register_to_config def __init__( self , *, lowercase_ = 4 , lowercase_ = 768 , lowercase_ , lowercase_ , ): """simple docstring""" super().__init__() UpperCAmelCase_ : List[Any] = nn.Parameter(torch.zeros(_a ) ) # parameters for additional clip time embeddings UpperCAmelCase_ : Tuple = nn.Linear(_a , _a ) UpperCAmelCase_ : int = nn.Linear(_a , _a ) # parameters for encoder hidden states UpperCAmelCase_ : List[Any] = clip_extra_context_tokens UpperCAmelCase_ : Any = nn.Linear( _a , self.clip_extra_context_tokens * cross_attention_dim ) UpperCAmelCase_ : Any = nn.Linear(_a , _a ) UpperCAmelCase_ : Tuple = nn.LayerNorm(_a ) def UpperCamelCase__ ( self , *, lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" if do_classifier_free_guidance: # Add the classifier free guidance embeddings to the image embeddings UpperCAmelCase_ : List[str] = image_embeddings.shape[0] UpperCAmelCase_ : str = self.learned_classifier_free_guidance_embeddings.unsqueeze(0 ) UpperCAmelCase_ : Any = classifier_free_guidance_embeddings.expand( _a , -1 ) UpperCAmelCase_ : Optional[Any] = torch.cat([classifier_free_guidance_embeddings, image_embeddings] , dim=0 ) # The image embeddings batch size and the text embeddings batch size are equal assert image_embeddings.shape[0] == prompt_embeds.shape[0] UpperCAmelCase_ : Tuple = prompt_embeds.shape[0] # "Specifically, we modify the architecture described in Nichol et al. (2021) by projecting and # adding CLIP embeddings to the existing timestep embedding, ... UpperCAmelCase_ : List[str] = self.embedding_proj(_a ) UpperCAmelCase_ : Dict = self.clip_image_embeddings_project_to_time_embeddings(_a ) UpperCAmelCase_ : str = time_projected_image_embeddings + time_projected_prompt_embeds # ... and by projecting CLIP embeddings into four # extra tokens of context that are concatenated to the sequence of outputs from the GLIDE text encoder" UpperCAmelCase_ : str = self.clip_extra_context_tokens_proj(_a ) UpperCAmelCase_ : Any = clip_extra_context_tokens.reshape(_a , -1 , self.clip_extra_context_tokens ) UpperCAmelCase_ : Optional[int] = clip_extra_context_tokens.permute(0 , 2 , 1 ) UpperCAmelCase_ : int = self.encoder_hidden_states_proj(_a ) UpperCAmelCase_ : Any = self.text_encoder_hidden_states_norm(_a ) UpperCAmelCase_ : Dict = torch.cat([clip_extra_context_tokens, text_encoder_hidden_states] , dim=1 ) return text_encoder_hidden_states, additive_clip_time_embeddings
368
"""simple docstring""" import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels _a = object() # For specifying empty leaf dict `{}` _a = object() def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Any = tuple((re.compile(x + "$" ) for x in qs) ) for i in range(len(__lowerCamelCase ) - len(__lowerCamelCase ) + 1 ): UpperCAmelCase_ : List[str] = [x.match(__lowerCamelCase ) for x, y in zip(__lowerCamelCase, ks[i:] )] if matches and all(__lowerCamelCase ): return True return False def __a ( __lowerCamelCase ): def replace(__lowerCamelCase, __lowerCamelCase ): for rule, replacement in rules: if _match(__lowerCamelCase, __lowerCamelCase ): return replacement return val return replace def __a ( ): return [ # embeddings (("transformer", "wpe", "embedding"), P("mp", __lowerCamelCase )), (("transformer", "wte", "embedding"), P("mp", __lowerCamelCase )), # atention (("attention", "(q_proj|k_proj|v_proj)", "kernel"), P(__lowerCamelCase, "mp" )), (("attention", "out_proj", "kernel"), P("mp", __lowerCamelCase )), (("attention", "out_proj", "bias"), None), # mlp (("mlp", "c_fc", "kernel"), P(__lowerCamelCase, "mp" )), (("mlp", "c_fc", "bias"), P("mp" )), (("mlp", "c_proj", "kernel"), P("mp", __lowerCamelCase )), (("mlp", "c_proj", "bias"), None), # layer norms ((r"ln_\d+", "bias"), None), ((r"\d+", r"ln_\d+", "scale"), None), (("ln_f", "bias"), None), (("ln_f", "scale"), None), ] def __a ( __lowerCamelCase ): UpperCAmelCase_ : List[str] = _get_partition_rules() UpperCAmelCase_ : Any = _replacement_rules(__lowerCamelCase ) UpperCAmelCase_ : Any = {k: _unmatched for k in flatten_dict(__lowerCamelCase )} UpperCAmelCase_ : Dict = {k: replace(__lowerCamelCase, __lowerCamelCase ) for k, v in initd.items()} assert _unmatched not in result.values(), "Incomplete partition spec." return freeze(unflatten_dict(__lowerCamelCase ) )
23
0
"""simple docstring""" import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _a = datasets.logging.get_logger(__name__) _a = '\\n@InProceedings{moosavi2019minimum,\n author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube},\n title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection},\n year = {2019},\n booktitle = {Proceedings of the 57th Annual Meeting of\n the Association for Computational Linguistics (Volume 1: Long Papers)},\n publisher = {Association for Computational Linguistics},\n address = {Florence, Italy},\n}\n\n@inproceedings{10.3115/1072399.1072405,\nauthor = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette},\ntitle = {A Model-Theoretic Coreference Scoring Scheme},\nyear = {1995},\nisbn = {1558604022},\npublisher = {Association for Computational Linguistics},\naddress = {USA},\nurl = {https://doi.org/10.3115/1072399.1072405},\ndoi = {10.3115/1072399.1072405},\nbooktitle = {Proceedings of the 6th Conference on Message Understanding},\npages = {45–52},\nnumpages = {8},\nlocation = {Columbia, Maryland},\nseries = {MUC6 ’95}\n}\n\n@INPROCEEDINGS{Bagga98algorithmsfor,\n author = {Amit Bagga and Breck Baldwin},\n title = {Algorithms for Scoring Coreference Chains},\n booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference},\n year = {1998},\n pages = {563--566}\n}\n\n@INPROCEEDINGS{Luo05oncoreference,\n author = {Xiaoqiang Luo},\n title = {On coreference resolution performance metrics},\n booktitle = {In Proc. of HLT/EMNLP},\n year = {2005},\n pages = {25--32},\n publisher = {URL}\n}\n\n@inproceedings{moosavi-strube-2016-coreference,\n title = "Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric",\n author = "Moosavi, Nafise Sadat and\n Strube, Michael",\n booktitle = "Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",\n month = aug,\n year = "2016",\n address = "Berlin, Germany",\n publisher = "Association for Computational Linguistics",\n url = "https://www.aclweb.org/anthology/P16-1060",\n doi = "10.18653/v1/P16-1060",\n pages = "632--642",\n}\n\n' _a = '\\nCoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which\nimplements of the common evaluation metrics including MUC [Vilain et al, 1995],\nB-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005],\nLEA [Moosavi and Strube, 2016] and the averaged CoNLL score\n(the average of the F1 values of MUC, B-cubed and CEAFe)\n[Denis and Baldridge, 2009a; Pradhan et al., 2011].\n\nThis wrapper of CoVal currently only work with CoNLL line format:\nThe CoNLL format has one word per line with all the annotation for this word in column separated by spaces:\nColumn Type Description\n1 Document ID This is a variation on the document filename\n2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc.\n3 Word number\n4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release.\n5 Part-of-Speech\n6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the "([pos] [word])" string (or leaf) and concatenating the items in the rows of that column.\n7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a "-"\n8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7.\n9 Word sense This is the word sense of the word in Column 3.\n10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data.\n11 Named Entities These columns identifies the spans representing various named entities.\n12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7.\nN Coreference Coreference chain information encoded in a parenthesis structure.\nMore informations on the format can be found here (section "*_conll File Format"): http://www.conll.cemantix.org/2012/data.html\n\nDetails on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md\n\nCoVal code was written by @ns-moosavi.\nSome parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py\nThe test suite is taken from https://github.com/conll/reference-coreference-scorers/\nMention evaluation and the test suite are added by @andreasvc.\nParsing CoNLL files is developed by Leo Born.\n' _a = '\nCalculates coreference evaluation metrics.\nArgs:\n predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format.\n Each prediction is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format.\n Each reference is a word with its annotations as a string made of columns joined with spaces.\n Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation)\n See the details on the format in the description of the metric.\n keep_singletons: After extracting all mentions of key or system files,\n mentions whose corresponding coreference chain is of size one,\n are considered as singletons. The default evaluation mode will include\n singletons in evaluations if they are included in the key or the system files.\n By setting \'keep_singletons=False\', all singletons in the key and system files\n will be excluded from the evaluation.\n NP_only: Most of the recent coreference resolvers only resolve NP mentions and\n leave out the resolution of VPs. By setting the \'NP_only\' option, the scorer will only evaluate the resolution of NPs.\n min_span: By setting \'min_span\', the scorer reports the results based on automatically detected minimum spans.\n Minimum spans are determined using the MINA algorithm.\n\nReturns:\n \'mentions\': mentions\n \'muc\': MUC metric [Vilain et al, 1995]\n \'bcub\': B-cubed [Bagga and Baldwin, 1998]\n \'ceafe\': CEAFe [Luo et al., 2005]\n \'lea\': LEA [Moosavi and Strube, 2016]\n \'conll_score\': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe)\n\nExamples:\n\n >>> coval = datasets.load_metric(\'coval\')\n >>> words = [\'bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -\',\n ... \'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)\',\n ... \'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)\',\n ... \'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -\',\n ... \'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -\',\n ... \'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -\']\n >>> references = [words]\n >>> predictions = [words]\n >>> results = coval.compute(predictions=predictions, references=references)\n >>> print(results) # doctest:+ELLIPSIS\n {\'mentions/recall\': 1.0,[...] \'conll_score\': 100.0}\n' def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=False, __lowerCamelCase=False, __lowerCamelCase=True, __lowerCamelCase=False, __lowerCamelCase="dummy_doc" ) -> List[Any]: UpperCAmelCase_ : str = {doc: key_lines} UpperCAmelCase_ : str = {doc: sys_lines} UpperCAmelCase_ : List[Any] = {} UpperCAmelCase_ : Dict = 0 UpperCAmelCase_ : str = 0 UpperCAmelCase_ : Union[str, Any] = 0 UpperCAmelCase_ : Tuple = 0 UpperCAmelCase_ : List[Any] = 0 UpperCAmelCase_ : Union[str, Any] = 0 UpperCAmelCase_ : Dict = reader.get_doc_mentions(__lowerCamelCase, key_doc_lines[doc], __lowerCamelCase ) key_singletons_num += singletons_num if NP_only or min_span: UpperCAmelCase_ : List[str] = reader.set_annotated_parse_trees(__lowerCamelCase, key_doc_lines[doc], __lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : Tuple = reader.get_doc_mentions(__lowerCamelCase, sys_doc_lines[doc], __lowerCamelCase ) sys_singletons_num += singletons_num if NP_only or min_span: UpperCAmelCase_ : Optional[Any] = reader.set_annotated_parse_trees(__lowerCamelCase, key_doc_lines[doc], __lowerCamelCase, __lowerCamelCase ) if remove_nested: UpperCAmelCase_ : List[Any] = reader.remove_nested_coref_mentions(__lowerCamelCase, __lowerCamelCase ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters UpperCAmelCase_ : Dict = reader.remove_nested_coref_mentions(__lowerCamelCase, __lowerCamelCase ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters UpperCAmelCase_ : Optional[int] = reader.get_mention_assignments(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : Any = reader.get_mention_assignments(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : int = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( "Number of removed nested coreferring mentions in the key " f"""annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}""" ) logger.info( "Number of resulting singleton clusters in the key " f"""annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}""" ) if not keep_singletons: logger.info( f"""{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system """ "files, respectively" ) return doc_coref_infos def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) -> str: UpperCAmelCase_ : Union[str, Any] = get_coref_infos(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : str = {} UpperCAmelCase_ : List[str] = 0 UpperCAmelCase_ : Dict = 0 for name, metric in metrics: UpperCAmelCase_ : Dict = evaluator.evaluate_documents(__lowerCamelCase, __lowerCamelCase, beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f"""{name}/recall""": recall, f"""{name}/precision""": precision, f"""{name}/f1""": fa} ) logger.info( name.ljust(10 ), f"""Recall: {recall * 100:.2f}""", f""" Precision: {precision * 100:.2f}""", f""" F1: {fa * 100:.2f}""", ) if conll_subparts_num == 3: UpperCAmelCase_ : List[Any] = (conll / 3) * 100 logger.info(f"""CoNLL score: {conll:.2f}""" ) output_scores.update({"conll_score": conll} ) return output_scores def __a ( __lowerCamelCase ) -> Tuple: UpperCAmelCase_ : Dict = False for line in key_lines: if not line.startswith("#" ): if len(line.split() ) > 6: UpperCAmelCase_ : Any = line.split()[5] if not parse_col == "-": UpperCAmelCase_ : List[str] = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Sequence(datasets.Value("string" ) ), "references": datasets.Sequence(datasets.Value("string" ) ), } ) , codebase_urls=["https://github.com/ns-moosavi/coval"] , reference_urls=[ "https://github.com/ns-moosavi/coval", "https://www.aclweb.org/anthology/P16-1060", "http://www.conll.cemantix.org/2012/data.html", ] , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_=True , lowercase_=False , lowercase_=False , lowercase_=False ): """simple docstring""" UpperCAmelCase_ : Dict = [ ("mentions", evaluator.mentions), ("muc", evaluator.muc), ("bcub", evaluator.b_cubed), ("ceafe", evaluator.ceafe), ("lea", evaluator.lea), ] if min_span: UpperCAmelCase_ : List[str] = util.check_gold_parse_annotation(_a ) if not has_gold_parse: raise NotImplementedError("References should have gold parse annotation to use 'min_span'." ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" UpperCAmelCase_ : int = evaluate( key_lines=_a , sys_lines=_a , metrics=_a , NP_only=_a , remove_nested=_a , keep_singletons=_a , min_span=_a , ) return score
369
"""simple docstring""" import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow _a = logging.getLogger() @unittest.skip("""Temporarily disable the doc tests.""" ) @require_torch @require_tf @slow class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , lowercase_ = True , ): """simple docstring""" UpperCAmelCase_ : List[str] = [file for file in os.listdir(lowercase_ ) if os.path.isfile(os.path.join(lowercase_ , lowercase_ ) )] if identifier is not None: UpperCAmelCase_ : Dict = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(lowercase_ , lowercase_ ): for n_ in n_identifier: UpperCAmelCase_ : str = [file for file in files if n_ not in file] else: UpperCAmelCase_ : Any = [file for file in files if n_identifier not in file] UpperCAmelCase_ : Union[str, Any] = ignore_files or [] ignore_files.append("__init__.py" ) UpperCAmelCase_ : Optional[int] = [file for file in files if file not in ignore_files] for file in files: # Open all files print("Testing" , lowercase_ ) if only_modules: UpperCAmelCase_ : str = file.split("." )[0] try: UpperCAmelCase_ : str = getattr(lowercase_ , lowercase_ ) UpperCAmelCase_ : Tuple = doctest.DocTestSuite(lowercase_ ) UpperCAmelCase_ : int = unittest.TextTestRunner().run(lowercase_ ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: UpperCAmelCase_ : Optional[Any] = doctest.testfile(str(".." / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = Path("src/transformers" ) UpperCAmelCase_ : str = "modeling" UpperCAmelCase_ : Optional[Any] = [ "modeling_ctrl.py", "modeling_tf_ctrl.py", ] self.analyze_directory(lowercase_ , identifier=lowercase_ , ignore_files=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = Path("src/transformers" ) UpperCAmelCase_ : Any = "tokenization" self.analyze_directory(lowercase_ , identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = Path("src/transformers" ) UpperCAmelCase_ : List[Any] = "configuration" self.analyze_directory(lowercase_ , identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = Path("src/transformers" ) UpperCAmelCase_ : List[Any] = ["configuration", "modeling", "tokenization"] self.analyze_directory(lowercase_ , n_identifier=lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = Path("docs/source" ) UpperCAmelCase_ : Union[str, Any] = ["favicon.ico"] self.analyze_directory(lowercase_ , ignore_files=lowercase_ , only_modules=lowercase_ )
23
0
"""simple docstring""" from __future__ import annotations import math _a = '2020.9.26' _a = 'xcodz-dot, cclaus, dhruvmanila' def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): if not all(isinstance(_lowerCAmelCase, (float, int) ) for val in locals().values() ): UpperCAmelCase_ : Tuple = f"""Input values must either be float or int: {list(locals().values() )}""" raise TypeError(_lowerCAmelCase ) UpperCAmelCase_ : List[str] = ((x * distance) / (z + distance)) * scale UpperCAmelCase_ : List[Any] = ((y * distance) / (z + distance)) * scale return projected_x, projected_y def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): if not isinstance(_lowerCAmelCase, _lowerCAmelCase ): raise TypeError("Axis must be a str" ) UpperCAmelCase_ : List[str] = locals() del input_variables["axis"] if not all(isinstance(_lowerCAmelCase, (float, int) ) for val in input_variables.values() ): UpperCAmelCase_ : Union[str, Any] = ( """Input values except axis must either be float or int: """ f"""{list(input_variables.values() )}""" ) raise TypeError(_lowerCAmelCase ) UpperCAmelCase_ : Optional[int] = (angle % 360) / 450 * 180 / math.pi if axis == "z": UpperCAmelCase_ : Dict = x * math.cos(_lowerCAmelCase ) - y * math.sin(_lowerCAmelCase ) UpperCAmelCase_ : Any = y * math.cos(_lowerCAmelCase ) + x * math.sin(_lowerCAmelCase ) UpperCAmelCase_ : Union[str, Any] = z elif axis == "x": UpperCAmelCase_ : Tuple = y * math.cos(_lowerCAmelCase ) - z * math.sin(_lowerCAmelCase ) UpperCAmelCase_ : Any = z * math.cos(_lowerCAmelCase ) + y * math.sin(_lowerCAmelCase ) UpperCAmelCase_ : Optional[int] = x elif axis == "y": UpperCAmelCase_ : Tuple = x * math.cos(_lowerCAmelCase ) - z * math.sin(_lowerCAmelCase ) UpperCAmelCase_ : Optional[Any] = z * math.cos(_lowerCAmelCase ) + x * math.sin(_lowerCAmelCase ) UpperCAmelCase_ : List[str] = y else: raise ValueError("not a valid axis, choose one of 'x', 'y', 'z'" ) return new_x, new_y, new_z if __name__ == "__main__": import doctest doctest.testmod() print(f"""{convert_to_ad(1.0, 2.0, 3.0, 10.0, 10.0) = }""") print(f"""{rotate(1.0, 2.0, 3.0, 'y', 90.0) = }""")
370
"""simple docstring""" import warnings from ...utils import is_sklearn_available, requires_backends if is_sklearn_available(): from scipy.stats import pearsonr, spearmanr from sklearn.metrics import fa_score, matthews_corrcoef _a = ( 'This metric will be removed from the library soon, metrics should be handled with the 🤗 Evaluate ' 'library. You can have a look at this example script for pointers: ' 'https://github.com/huggingface/transformers/blob/main/examples/pytorch/text-classification/run_glue.py' ) def __a ( __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) return (preds == labels).mean() def __a ( __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) UpperCAmelCase_ : Optional[Any] = simple_accuracy(__lowerCamelCase, __lowerCamelCase ) UpperCAmelCase_ : List[Any] = fa_score(y_true=__lowerCamelCase, y_pred=__lowerCamelCase ) return { "acc": acc, "f1": fa, "acc_and_f1": (acc + fa) / 2, } def __a ( __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) UpperCAmelCase_ : Any = pearsonr(__lowerCamelCase, __lowerCamelCase )[0] UpperCAmelCase_ : Optional[Any] = spearmanr(__lowerCamelCase, __lowerCamelCase )[0] return { "pearson": pearson_corr, "spearmanr": spearman_corr, "corr": (pearson_corr + spearman_corr) / 2, } def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) assert len(__lowerCamelCase ) == len(__lowerCamelCase ), f"""Predictions and labels have mismatched lengths {len(__lowerCamelCase )} and {len(__lowerCamelCase )}""" if task_name == "cola": return {"mcc": matthews_corrcoef(__lowerCamelCase, __lowerCamelCase )} elif task_name == "sst-2": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "mrpc": return acc_and_fa(__lowerCamelCase, __lowerCamelCase ) elif task_name == "sts-b": return pearson_and_spearman(__lowerCamelCase, __lowerCamelCase ) elif task_name == "qqp": return acc_and_fa(__lowerCamelCase, __lowerCamelCase ) elif task_name == "mnli": return {"mnli/acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "mnli-mm": return {"mnli-mm/acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "qnli": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "rte": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "wnli": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} elif task_name == "hans": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} else: raise KeyError(__lowerCamelCase ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): warnings.warn(__lowerCamelCase, __lowerCamelCase ) requires_backends(__lowerCamelCase, "sklearn" ) if len(__lowerCamelCase ) != len(__lowerCamelCase ): raise ValueError(f"""Predictions and labels have mismatched lengths {len(__lowerCamelCase )} and {len(__lowerCamelCase )}""" ) if task_name == "xnli": return {"acc": simple_accuracy(__lowerCamelCase, __lowerCamelCase )} else: raise KeyError(__lowerCamelCase )
23
0
"""simple docstring""" from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=__lowerCamelCase ) class A_ (__lowerCamelCase ): '''simple docstring''' # `task` is not a ClassVar since we want it to be part of the `asdict` output for JSON serialization SCREAMING_SNAKE_CASE__ : str = field(default="""summarization""" ,metadata={"""include_in_asdict_even_if_is_default""": True} ) SCREAMING_SNAKE_CASE__ : ClassVar[Features] = Features({"""text""": Value("""string""" )} ) SCREAMING_SNAKE_CASE__ : ClassVar[Features] = Features({"""summary""": Value("""string""" )} ) SCREAMING_SNAKE_CASE__ : str = "text" SCREAMING_SNAKE_CASE__ : str = "summary" @property def UpperCamelCase__ ( self ): """simple docstring""" return {self.text_column: "text", self.summary_column: "summary"}
371
"""simple docstring""" import json import os from typing import Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a = logging.get_logger(__name__) _a = {'vocab_file': 'vocab.json'} _a = { 'vocab_file': { 'mgp-str': 'https://huggingface.co/alibaba-damo/mgp-str-base/blob/main/vocab.json', } } _a = {'mgp-str': 27} class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : List[str] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : str = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , lowercase_ , lowercase_="[GO]" , lowercase_="[GO]" , lowercase_="[s]" , lowercase_="[GO]" , **lowercase_ ): """simple docstring""" super().__init__( unk_token=lowercase_ , bos_token=lowercase_ , eos_token=lowercase_ , pad_token=lowercase_ , **lowercase_ , ) with open(lowercase_ , encoding="utf-8" ) as vocab_handle: UpperCAmelCase_ : Dict = json.load(lowercase_ ) UpperCAmelCase_ : Dict = {v: k for k, v in self.vocab.items()} @property def UpperCamelCase__ ( self ): """simple docstring""" return len(self.vocab ) def UpperCamelCase__ ( self ): """simple docstring""" return dict(self.vocab , **self.added_tokens_encoder ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = [] for s in text: char_tokens.extend(lowercase_ ) return char_tokens def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.vocab.get(lowercase_ , self.vocab.get(self.unk_token ) ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.decoder.get(lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if not os.path.isdir(lowercase_ ): logger.error("Vocabulary path ({}) should be a directory".format(lowercase_ ) ) return UpperCAmelCase_ : Optional[int] = os.path.join( lowercase_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) with open(lowercase_ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.vocab , indent=2 , sort_keys=lowercase_ , ensure_ascii=lowercase_ ) + "\n" ) return (vocab_file,)
23
0
"""simple docstring""" from __future__ import annotations _a = 1.6_0_2_1e-1_9 # units = C def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, ): if (conductivity, electron_conc, mobility).count(0 ) != 1: raise ValueError("You cannot supply more or less than 2 values" ) elif conductivity < 0: raise ValueError("Conductivity cannot be negative" ) elif electron_conc < 0: raise ValueError("Electron concentration cannot be negative" ) elif mobility < 0: raise ValueError("mobility cannot be negative" ) elif conductivity == 0: return ( "conductivity", mobility * electron_conc * ELECTRON_CHARGE, ) elif electron_conc == 0: return ( "electron_conc", conductivity / (mobility * ELECTRON_CHARGE), ) else: return ( "mobility", conductivity / (electron_conc * ELECTRON_CHARGE), ) if __name__ == "__main__": import doctest doctest.testmod()
350
"""simple docstring""" import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency _a = { 'E': 12.70, 'T': 9.06, 'A': 8.17, 'O': 7.51, 'I': 6.97, 'N': 6.75, 'S': 6.33, 'H': 6.09, 'R': 5.99, 'D': 4.25, 'L': 4.03, 'C': 2.78, 'U': 2.76, 'M': 2.41, 'W': 2.36, 'F': 2.23, 'G': 2.02, 'Y': 1.97, 'P': 1.93, 'B': 1.29, 'V': 0.98, 'K': 0.77, 'J': 0.15, 'X': 0.15, 'Q': 0.10, 'Z': 0.07, } _a = 'ETAOINSHRDLCUMWFGYPBVKJXQZ' _a = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def __a ( __lowerCamelCase ): UpperCAmelCase_ : Any = {letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def __a ( __lowerCamelCase ): return x[0] def __a ( __lowerCamelCase ): UpperCAmelCase_ : Any = get_letter_count(__lowerCamelCase ) UpperCAmelCase_ : dict[int, list[str]] = { freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(__lowerCamelCase ) UpperCAmelCase_ : dict[int, str] = {} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find, reverse=__lowerCamelCase ) UpperCAmelCase_ : Any = "".join(freq_to_letter[freq] ) UpperCAmelCase_ : str = list(freq_to_letter_str.items() ) freq_pairs.sort(key=__lowerCamelCase, reverse=__lowerCamelCase ) UpperCAmelCase_ : list[str] = [freq_pair[1] for freq_pair in freq_pairs] return "".join(__lowerCamelCase ) def __a ( __lowerCamelCase ): UpperCAmelCase_ : Any = get_frequency_order(__lowerCamelCase ) UpperCAmelCase_ : int = 0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
23
0
"""simple docstring""" import argparse from pathlib import Path import fairseq import torch from fairseq.models.xmod import XMODModel as FairseqXmodModel from packaging import version from transformers import XmodConfig, XmodForMaskedLM, XmodForSequenceClassification from transformers.utils import logging if version.parse(fairseq.__version__) < version.parse('0.12.2'): raise Exception('requires fairseq >= 0.12.2') if version.parse(fairseq.__version__) > version.parse('2'): raise Exception('requires fairseq < v2') logging.set_verbosity_info() _a = logging.get_logger(__name__) _a = """Hello, World!""" _a = """en_XX""" def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Optional[Any] = Path("data_bin" ) UpperCAmelCase_ : Dict = FairseqXmodModel.from_pretrained( model_name_or_path=str(Path(__lowerCamelCase ).parent ), checkpoint_file=Path(__lowerCamelCase ).name, _name="xmod_base", arch="xmod_base", task="multilingual_masked_lm", data_name_or_path=str(__lowerCamelCase ), bpe="sentencepiece", sentencepiece_model=str(Path(__lowerCamelCase ).parent / "sentencepiece.bpe.model" ), src_dict=str(data_dir / "dict.txt" ), ) xmod.eval() # disable dropout print(__lowerCamelCase ) UpperCAmelCase_ : Optional[Any] = xmod.model.encoder.sentence_encoder UpperCAmelCase_ : Optional[int] = XmodConfig( vocab_size=xmod_sent_encoder.embed_tokens.num_embeddings, hidden_size=xmod.cfg.model.encoder_embed_dim, num_hidden_layers=xmod.cfg.model.encoder_layers, num_attention_heads=xmod.cfg.model.encoder_attention_heads, intermediate_size=xmod.cfg.model.encoder_ffn_embed_dim, max_position_embeddings=514, type_vocab_size=1, layer_norm_eps=1E-5, pre_norm=xmod.cfg.model.encoder_normalize_before, adapter_reduction_factor=getattr(xmod.cfg.model, "bottleneck", 2 ), adapter_layer_norm=xmod.cfg.model.adapter_layer_norm, adapter_reuse_layer_norm=xmod.cfg.model.adapter_reuse_layer_norm, ln_before_adapter=xmod.cfg.model.ln_before_adapter, languages=xmod.cfg.model.languages, ) if classification_head: UpperCAmelCase_ : int = xmod.model.classification_heads["mnli"].out_proj.weight.shape[0] print("Our X-MOD config:", __lowerCamelCase ) UpperCAmelCase_ : List[str] = XmodForSequenceClassification(__lowerCamelCase ) if classification_head else XmodForMaskedLM(__lowerCamelCase ) model.eval() # Now let's copy all the weights. # Embeddings UpperCAmelCase_ : List[str] = xmod_sent_encoder.embed_tokens.weight UpperCAmelCase_ : str = xmod_sent_encoder.embed_positions.weight UpperCAmelCase_ : Optional[Any] = torch.zeros_like( model.roberta.embeddings.token_type_embeddings.weight ) # just zero them out b/c xmod doesn't use them. UpperCAmelCase_ : List[Any] = xmod_sent_encoder.layernorm_embedding.weight UpperCAmelCase_ : Union[str, Any] = xmod_sent_encoder.layernorm_embedding.bias for i in range(config.num_hidden_layers ): # Encoder: start of layer UpperCAmelCase_ : List[str] = model.roberta.encoder.layer[i] UpperCAmelCase_ : Any = xmod_sent_encoder.layers[i] # self attention UpperCAmelCase_ : str = layer.attention.self if not ( xmod_layer.self_attn.k_proj.weight.data.shape == xmod_layer.self_attn.q_proj.weight.data.shape == xmod_layer.self_attn.v_proj.weight.data.shape == torch.Size((config.hidden_size, config.hidden_size) ) ): raise AssertionError("Dimensions of self-attention weights do not match." ) UpperCAmelCase_ : Union[str, Any] = xmod_layer.self_attn.q_proj.weight UpperCAmelCase_ : str = xmod_layer.self_attn.q_proj.bias UpperCAmelCase_ : Any = xmod_layer.self_attn.k_proj.weight UpperCAmelCase_ : Dict = xmod_layer.self_attn.k_proj.bias UpperCAmelCase_ : Tuple = xmod_layer.self_attn.v_proj.weight UpperCAmelCase_ : Dict = xmod_layer.self_attn.v_proj.bias # self-attention output UpperCAmelCase_ : str = layer.attention.output if self_output.dense.weight.shape != xmod_layer.self_attn.out_proj.weight.shape: raise AssertionError("Dimensions of self-attention output weights do not match." ) UpperCAmelCase_ : Union[str, Any] = xmod_layer.self_attn.out_proj.weight UpperCAmelCase_ : Optional[Any] = xmod_layer.self_attn.out_proj.bias UpperCAmelCase_ : Dict = xmod_layer.self_attn_layer_norm.weight UpperCAmelCase_ : Dict = xmod_layer.self_attn_layer_norm.bias # intermediate UpperCAmelCase_ : List[str] = layer.intermediate if intermediate.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError("Dimensions of intermediate weights do not match." ) UpperCAmelCase_ : str = xmod_layer.fca.weight UpperCAmelCase_ : Tuple = xmod_layer.fca.bias # output UpperCAmelCase_ : List[Any] = layer.output if bert_output.dense.weight.shape != xmod_layer.fca.weight.shape: raise AssertionError("Dimensions of feed-forward weights do not match." ) UpperCAmelCase_ : Optional[Any] = xmod_layer.fca.weight UpperCAmelCase_ : List[Any] = xmod_layer.fca.bias UpperCAmelCase_ : Any = xmod_layer.final_layer_norm.weight UpperCAmelCase_ : List[Any] = xmod_layer.final_layer_norm.bias if bert_output.adapter_layer_norm is not None: UpperCAmelCase_ : Union[str, Any] = xmod_layer.adapter_layer_norm.weight UpperCAmelCase_ : List[str] = xmod_layer.adapter_layer_norm.bias if sorted(bert_output.adapter_modules.keys() ) != sorted(xmod_layer.adapter_modules.keys() ): raise AssertionError("Lists of language adapters do not match." ) for lang_code, adapter in xmod_layer.adapter_modules.items(): UpperCAmelCase_ : Union[str, Any] = bert_output.adapter_modules[lang_code] UpperCAmelCase_ : Tuple = xmod_layer.adapter_modules[lang_code] UpperCAmelCase_ : Optional[Any] = from_adapter.fca.weight UpperCAmelCase_ : Any = from_adapter.fca.bias UpperCAmelCase_ : List[Any] = from_adapter.fca.weight UpperCAmelCase_ : int = from_adapter.fca.bias # end of layer if xmod_sent_encoder.layer_norm is not None: UpperCAmelCase_ : int = xmod_sent_encoder.layer_norm.weight UpperCAmelCase_ : Any = xmod_sent_encoder.layer_norm.bias if classification_head: UpperCAmelCase_ : Optional[int] = xmod.model.classification_heads["mnli"].dense.weight UpperCAmelCase_ : Any = xmod.model.classification_heads["mnli"].dense.bias UpperCAmelCase_ : Any = xmod.model.classification_heads["mnli"].out_proj.weight UpperCAmelCase_ : Optional[int] = xmod.model.classification_heads["mnli"].out_proj.bias else: # LM Head UpperCAmelCase_ : int = xmod.model.encoder.lm_head.dense.weight UpperCAmelCase_ : str = xmod.model.encoder.lm_head.dense.bias UpperCAmelCase_ : Optional[Any] = xmod.model.encoder.lm_head.layer_norm.weight UpperCAmelCase_ : Optional[Any] = xmod.model.encoder.lm_head.layer_norm.bias UpperCAmelCase_ : List[Any] = xmod.model.encoder.lm_head.weight UpperCAmelCase_ : Optional[Any] = xmod.model.encoder.lm_head.bias # Let's check that we get the same results. UpperCAmelCase_ : List[str] = xmod.encode(__lowerCamelCase ).unsqueeze(0 ) # batch of size 1 model.roberta.set_default_language(__lowerCamelCase ) UpperCAmelCase_ : List[str] = model(__lowerCamelCase )[0] if classification_head: UpperCAmelCase_ : Optional[Any] = xmod.model.classification_heads["mnli"](xmod.extract_features(__lowerCamelCase ) ) else: UpperCAmelCase_ : Dict = xmod.model(__lowerCamelCase, lang_id=[SAMPLE_LANGUAGE] )[0] print(our_output.shape, their_output.shape ) UpperCAmelCase_ : str = torch.max(torch.abs(our_output - their_output ) ).item() print(f"""max_absolute_diff = {max_absolute_diff}""" ) # ~ 1e-7 UpperCAmelCase_ : List[Any] = torch.allclose(__lowerCamelCase, __lowerCamelCase, atol=1E-3 ) print("Do both models output the same tensors?", "🔥" if success else "💩" ) if not success: raise Exception("Something went wRoNg" ) Path(__lowerCamelCase ).mkdir(parents=__lowerCamelCase, exist_ok=__lowerCamelCase ) print(f"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--xmod_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) parser.add_argument( '--classification_head', action='store_true', help='Whether to convert a final classification head.' ) _a = parser.parse_args() convert_xmod_checkpoint_to_pytorch( args.xmod_checkpoint_path, args.pytorch_dump_folder_path, args.classification_head )
351
"""simple docstring""" import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) _a = logging.getLogger() def __a ( ): UpperCAmelCase_ : Tuple = argparse.ArgumentParser() parser.add_argument("-f" ) UpperCAmelCase_ : Dict = parser.parse_args() return args.f class A_ (lowercase__ ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = logging.StreamHandler(sys.stdout ) logger.addHandler(lowercase_ ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , "run_glue_deebert.py" ) with patch.object(lowercase_ , "argv" , lowercase_ ): UpperCAmelCase_ : List[str] = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(lowercase_ , 0.6_66 ) @slow @require_torch_non_multi_gpu def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = "\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n ".split() self.run_and_check(lowercase_ ) UpperCAmelCase_ : Optional[Any] = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowercase_ ) UpperCAmelCase_ : Dict = "\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n ".split() self.run_and_check(lowercase_ )
23
0
"""simple docstring""" def __a ( __lowerCamelCase = 10 ): if not isinstance(lowerCamelCase__, lowerCamelCase__ ) or n < 0: raise ValueError("Invalid input" ) UpperCAmelCase_ : Union[str, Any] = 10**n UpperCAmelCase_ : Optional[Any] = 2_8433 * (pow(2, 783_0457, lowerCamelCase__ )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(f"""{solution(10) = }""")
352
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _a = {'configuration_unispeech': ['UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP', 'UniSpeechConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST', 'UniSpeechForCTC', 'UniSpeechForPreTraining', 'UniSpeechForSequenceClassification', 'UniSpeechModel', 'UniSpeechPreTrainedModel', ] if TYPE_CHECKING: from .configuration_unispeech import UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP, UniSpeechConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_unispeech import ( UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST, UniSpeechForCTC, UniSpeechForPreTraining, UniSpeechForSequenceClassification, UniSpeechModel, UniSpeechPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
23
0
"""simple docstring""" from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _a = {'configuration_mmbt': ['MMBTConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ['MMBTForClassification', 'MMBTModel', 'ModalEmbeddings'] if TYPE_CHECKING: from .configuration_mmbt import MMBTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mmbt import MMBTForClassification, MMBTModel, ModalEmbeddings else: import sys _a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
353
"""simple docstring""" from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) _a = logging.get_logger(__name__) # pylint: disable=invalid-name _a = '\n Examples:\n ```py\n >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained("kandinsky-community/kandinsky-2-2-prior")\n >>> pipe_prior.to("cuda")\n >>> prompt = "red cat, 4k photo"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> zero_image_emb = out.negative_image_embeds\n >>> pipe = KandinskyV22Pipeline.from_pretrained("kandinsky-community/kandinsky-2-2-decoder")\n >>> pipe.to("cuda")\n >>> image = pipe(\n ... image_embeds=image_emb,\n ... negative_image_embeds=zero_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=50,\n ... ).images\n >>> image[0].save("cat.png")\n ```\n' def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=8 ): UpperCAmelCase_ : List[str] = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 UpperCAmelCase_ : Tuple = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class A_ (lowercase__ ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , ): """simple docstring""" super().__init__() self.register_modules( unet=lowercase_ , scheduler=lowercase_ , movq=lowercase_ , ) UpperCAmelCase_ : int = 2 ** (len(self.movq.config.block_out_channels ) - 1) def UpperCamelCase__ ( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ ): """simple docstring""" if latents is None: UpperCAmelCase_ : Dict = randn_tensor(lowercase_ , generator=lowercase_ , device=lowercase_ , dtype=lowercase_ ) else: if latents.shape != shape: raise ValueError(F"""Unexpected latents shape, got {latents.shape}, expected {shape}""" ) UpperCAmelCase_ : str = latents.to(lowercase_ ) UpperCAmelCase_ : Dict = latents * scheduler.init_noise_sigma return latents def UpperCamelCase__ ( self , lowercase_=0 ): """simple docstring""" if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError("Please install accelerate via `pip install accelerate`" ) UpperCAmelCase_ : Any = torch.device(F"""cuda:{gpu_id}""" ) UpperCAmelCase_ : int = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(lowercase_ , lowercase_ ) def UpperCamelCase__ ( self , lowercase_=0 ): """simple docstring""" if is_accelerate_available() and is_accelerate_version(">=" , "0.17.0.dev0" ): from accelerate import cpu_offload_with_hook else: raise ImportError("`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher." ) UpperCAmelCase_ : Any = torch.device(F"""cuda:{gpu_id}""" ) if self.device.type != "cpu": self.to("cpu" , silence_dtype_warnings=lowercase_ ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) UpperCAmelCase_ : List[Any] = None for cpu_offloaded_model in [self.unet, self.movq]: UpperCAmelCase_ , UpperCAmelCase_ : str = cpu_offload_with_hook(lowercase_ , lowercase_ , prev_module_hook=lowercase_ ) # We'll offload the last model manually. UpperCAmelCase_ : Tuple = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def UpperCamelCase__ ( self ): """simple docstring""" if not hasattr(self.unet , "_hf_hook" ): return self.device for module in self.unet.modules(): if ( hasattr(lowercase_ , "_hf_hook" ) and hasattr(module._hf_hook , "execution_device" ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(lowercase_ ) def __call__( self , lowercase_ , lowercase_ , lowercase_ = 512 , lowercase_ = 512 , lowercase_ = 100 , lowercase_ = 4.0 , lowercase_ = 1 , lowercase_ = None , lowercase_ = None , lowercase_ = "pil" , lowercase_ = True , ): """simple docstring""" UpperCAmelCase_ : str = self._execution_device UpperCAmelCase_ : List[Any] = guidance_scale > 1.0 if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = torch.cat(lowercase_ , dim=0 ) UpperCAmelCase_ : Any = image_embeds.shape[0] * num_images_per_prompt if isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : List[Any] = torch.cat(lowercase_ , dim=0 ) if do_classifier_free_guidance: UpperCAmelCase_ : Tuple = image_embeds.repeat_interleave(lowercase_ , dim=0 ) UpperCAmelCase_ : List[str] = negative_image_embeds.repeat_interleave(lowercase_ , dim=0 ) UpperCAmelCase_ : Optional[Any] = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=lowercase_ ) self.scheduler.set_timesteps(lowercase_ , device=lowercase_ ) UpperCAmelCase_ : List[Any] = self.scheduler.timesteps UpperCAmelCase_ : List[str] = self.unet.config.in_channels UpperCAmelCase_ , UpperCAmelCase_ : Dict = downscale_height_and_width(lowercase_ , lowercase_ , self.movq_scale_factor ) # create initial latent UpperCAmelCase_ : int = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , lowercase_ , lowercase_ , lowercase_ , self.scheduler , ) for i, t in enumerate(self.progress_bar(lowercase_ ) ): # expand the latents if we are doing classifier free guidance UpperCAmelCase_ : Tuple = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents UpperCAmelCase_ : Union[str, Any] = {"image_embeds": image_embeds} UpperCAmelCase_ : Optional[Any] = self.unet( sample=lowercase_ , timestep=lowercase_ , encoder_hidden_states=lowercase_ , added_cond_kwargs=lowercase_ , return_dict=lowercase_ , )[0] if do_classifier_free_guidance: UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = noise_pred.split(latents.shape[1] , dim=1 ) UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = noise_pred.chunk(2 ) UpperCAmelCase_ , UpperCAmelCase_ : Tuple = variance_pred.chunk(2 ) UpperCAmelCase_ : int = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) UpperCAmelCase_ : str = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , "variance_type" ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): UpperCAmelCase_ , UpperCAmelCase_ : Dict = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase_ : List[str] = self.scheduler.step( lowercase_ , lowercase_ , lowercase_ , generator=lowercase_ , )[0] # post-processing UpperCAmelCase_ : Tuple = self.movq.decode(lowercase_ , force_not_quantize=lowercase_ )["sample"] if output_type not in ["pt", "np", "pil"]: raise ValueError(F"""Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}""" ) if output_type in ["np", "pil"]: UpperCAmelCase_ : List[Any] = image * 0.5 + 0.5 UpperCAmelCase_ : int = image.clamp(0 , 1 ) UpperCAmelCase_ : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": UpperCAmelCase_ : Dict = self.numpy_to_pil(lowercase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowercase_ )
23
0
"""simple docstring""" import argparse import logging import os import time import timeit import datasets import numpy as np import pycuda.autoinit # noqa: F401 import pycuda.driver as cuda import tensorrt as trt import torch from absl import logging as absl_logging from accelerate import Accelerator from datasets import load_dataset, load_metric from torch.utils.data import DataLoader from utils_qa import postprocess_qa_predictions import transformers from transformers import AutoTokenizer, EvalPrediction, default_data_collator, set_seed from transformers.trainer_pt_utils import nested_concat, nested_truncate _a = trt.Logger(trt.Logger.WARNING) _a = absl_logging.get_absl_logger() absl_logger.setLevel(logging.WARNING) _a = logging.getLogger(__name__) _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--onnx_model_path', default=None, type=str, required=True, help='Path to ONNX model: ', ) parser.add_argument( '--output_dir', default=None, type=str, required=True, help='The output directory where the model checkpoints and predictions will be written.', ) # Other parameters parser.add_argument( '--tokenizer_name', default='', type=str, required=True, help='Pretrained tokenizer name or path if not the same as model_name', ) parser.add_argument( '--version_2_with_negative', action='store_true', help='If true, the SQuAD examples contain some that do not have an answer.', ) parser.add_argument( '--null_score_diff_threshold', type=float, default=0.0, help='If null_score - best_non_null is greater than the threshold predict null.', ) parser.add_argument( '--max_seq_length', default=384, type=int, help=( 'The maximum total input sequence length after WordPiece tokenization. Sequences ' 'longer than this will be truncated, and sequences shorter than this will be padded.' ), ) parser.add_argument( '--doc_stride', default=128, type=int, help='When splitting up a long document into chunks, how much stride to take between chunks.', ) parser.add_argument('--per_device_eval_batch_size', default=8, type=int, help='Batch size per GPU/CPU for evaluation.') parser.add_argument( '--n_best_size', default=20, type=int, help='The total number of n-best predictions to generate in the nbest_predictions.json output file.', ) parser.add_argument( '--max_answer_length', default=30, type=int, help=( 'The maximum length of an answer that can be generated. This is needed because the start ' 'and end predictions are not conditioned on one another.' ), ) parser.add_argument('--seed', type=int, default=42, help='random seed for initialization') parser.add_argument( '--dataset_name', type=str, default=None, required=True, help='The name of the dataset to use (via the datasets library).', ) parser.add_argument( '--dataset_config_name', type=str, default=None, help='The configuration name of the dataset to use (via the datasets library).', ) parser.add_argument( '--preprocessing_num_workers', type=int, default=4, help='A csv or a json file containing the training data.' ) parser.add_argument('--overwrite_cache', action='store_true', help='Overwrite the cached training and evaluation sets') parser.add_argument( '--fp16', action='store_true', help='Whether to use 16-bit (mixed) precision instead of 32-bit', ) parser.add_argument( '--int8', action='store_true', help='Whether to use INT8', ) _a = parser.parse_args() if args.tokenizer_name: _a = AutoTokenizer.from_pretrained(args.tokenizer_name, use_fast=True) else: raise ValueError( 'You are instantiating a new tokenizer from scratch. This is not supported by this script.' 'You can do it from another script, save it, and load it from here, using --tokenizer_name.' ) logger.info('Training/evaluation parameters %s', args) _a = args.per_device_eval_batch_size _a = (args.eval_batch_size, args.max_seq_length) # TRT Engine properties _a = True _a = 'temp_engine/bert-fp32.engine' if args.fpaa: _a = 'temp_engine/bert-fp16.engine' if args.inta: _a = 'temp_engine/bert-int8.engine' # import ONNX file if not os.path.exists('temp_engine'): os.makedirs('temp_engine') _a = 1 << (int)(trt.NetworkDefinitionCreationFlag.EXPLICIT_BATCH) with trt.Builder(TRT_LOGGER) as builder, builder.create_network(EXPLICIT_BATCH) as network, trt.OnnxParser( network, TRT_LOGGER ) as parser: with open(args.onnx_model_path, 'rb') as model: if not parser.parse(model.read()): for error in range(parser.num_errors): print(parser.get_error(error)) # Query input names and shapes from parsed TensorRT network _a = [network.get_input(i) for i in range(network.num_inputs)] _a = [_input.name for _input in network_inputs] # ex: ["actual_input1"] with builder.create_builder_config() as config: _a = 1 << 50 if STRICT_TYPES: config.set_flag(trt.BuilderFlag.STRICT_TYPES) if args.fpaa: config.set_flag(trt.BuilderFlag.FPaa) if args.inta: config.set_flag(trt.BuilderFlag.INTa) _a = builder.create_optimization_profile() config.add_optimization_profile(profile) for i in range(len(input_names)): profile.set_shape(input_names[i], INPUT_SHAPE, INPUT_SHAPE, INPUT_SHAPE) _a = builder.build_engine(network, config) # serialize_engine and store in file (can be directly loaded and deserialized): with open(engine_name, 'wb') as f: f.write(engine.serialize()) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : List[str] = np.asarray(inputs["input_ids"], dtype=np.intaa ) UpperCAmelCase_ : Dict = np.asarray(inputs["attention_mask"], dtype=np.intaa ) UpperCAmelCase_ : Optional[Any] = np.asarray(inputs["token_type_ids"], dtype=np.intaa ) # Copy inputs cuda.memcpy_htod_async(d_inputs[0], input_ids.ravel(), _A ) cuda.memcpy_htod_async(d_inputs[1], attention_mask.ravel(), _A ) cuda.memcpy_htod_async(d_inputs[2], token_type_ids.ravel(), _A ) # start time UpperCAmelCase_ : Union[str, Any] = time.time() # Run inference context.execute_async( bindings=[int(_A ) for d_inp in d_inputs] + [int(_A ), int(_A )], stream_handle=stream.handle ) # Transfer predictions back from GPU cuda.memcpy_dtoh_async(_A, _A, _A ) cuda.memcpy_dtoh_async(_A, _A, _A ) # Synchronize the stream and take time stream.synchronize() # end time UpperCAmelCase_ : int = time.time() UpperCAmelCase_ : Tuple = end_time - start_time UpperCAmelCase_ : Dict = (h_outputa, h_outputa) # print(outputs) return outputs, infer_time # Initialize the accelerator. We will let the accelerator handle device placement for us in this example. _a = Accelerator() # Make one log on every process with the configuration for debugging. logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO, ) # Setup logging, we only want one process per machine to log things on the screen. # accelerator.is_local_main_process is only True for one process per machine. logger.setLevel(logging.INFO if accelerator.is_local_main_process else logging.ERROR) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_info() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # If passed along, set the training seed now. if args.seed is not None: set_seed(args.seed) # Get the datasets: you can either provide your own CSV/JSON/TXT training and evaluation files (see below) # or just provide the name of one of the public datasets available on the hub at https://huggingface.co/datasets/ # (the dataset will be downloaded automatically from the datasets Hub). # # For CSV/JSON files, this script will use the column called 'text' or the first column if no column called # 'text' is found. You can easily tweak this behavior (see below). if args.dataset_name is not None: # Downloading and loading a dataset from the hub. _a = load_dataset(args.dataset_name, args.dataset_config_name) else: raise ValueError('Evaluation requires a dataset name') # See more about loading any type of standard or custom dataset (from files, python dict, pandas DataFrame, etc) at # https://huggingface.co/docs/datasets/loading_datasets.html. # Preprocessing the datasets. # Preprocessing is slighlty different for training and evaluation. _a = raw_datasets['validation'].column_names _a = 'question' if 'question' in column_names else column_names[0] _a = 'context' if 'context' in column_names else column_names[1] _a = 'answers' if 'answers' in column_names else column_names[2] # Padding side determines if we do (question|context) or (context|question). _a = tokenizer.padding_side == 'right' if args.max_seq_length > tokenizer.model_max_length: logger.warning( f"""The max_seq_length passed ({args.max_seq_length}) is larger than the maximum length for the""" f"""model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.""" ) _a = min(args.max_seq_length, tokenizer.model_max_length) def __a ( __lowerCamelCase ): UpperCAmelCase_ : Optional[int] = [q.lstrip() for q in examples[question_column_name]] # Tokenize our examples with truncation and maybe padding, but keep the overflows using a stride. This results # in one example possible giving several features when a context is long, each of those features having a # context that overlaps a bit the context of the previous feature. UpperCAmelCase_ : Union[str, Any] = tokenizer( examples[question_column_name if pad_on_right else context_column_name], examples[context_column_name if pad_on_right else question_column_name], truncation="only_second" if pad_on_right else "only_first", max_length=_A, stride=args.doc_stride, return_overflowing_tokens=_A, return_offsets_mapping=_A, padding="max_length", ) # Since one example might give us several features if it has a long context, we need a map from a feature to # its corresponding example. This key gives us just that. UpperCAmelCase_ : Optional[Any] = tokenized_examples.pop("overflow_to_sample_mapping" ) # For evaluation, we will need to convert our predictions to substrings of the context, so we keep the # corresponding example_id and we will store the offset mappings. UpperCAmelCase_ : Optional[Any] = [] for i in range(len(tokenized_examples["input_ids"] ) ): # Grab the sequence corresponding to that example (to know what is the context and what is the question). UpperCAmelCase_ : List[Any] = tokenized_examples.sequence_ids(_A ) UpperCAmelCase_ : Tuple = 1 if pad_on_right else 0 # One example can give several spans, this is the index of the example containing this span of text. UpperCAmelCase_ : Optional[int] = sample_mapping[i] tokenized_examples["example_id"].append(examples["id"][sample_index] ) # Set to None the offset_mapping that are not part of the context so it's easy to determine if a token # position is part of the context or not. UpperCAmelCase_ : List[str] = [ (o if sequence_ids[k] == context_index else None) for k, o in enumerate(tokenized_examples["offset_mapping"][i] ) ] return tokenized_examples _a = raw_datasets['validation'] # Validation Feature Creation _a = eval_examples.map( prepare_validation_features, batched=True, num_proc=args.preprocessing_num_workers, remove_columns=column_names, load_from_cache_file=not args.overwrite_cache, desc='Running tokenizer on validation dataset', ) _a = default_data_collator _a = eval_dataset.remove_columns(['example_id', 'offset_mapping']) _a = DataLoader( eval_dataset_for_model, collate_fn=data_collator, batch_size=args.per_device_eval_batch_size ) def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase="eval" ): UpperCAmelCase_ : Union[str, Any] = postprocess_qa_predictions( examples=_A, features=_A, predictions=_A, version_2_with_negative=args.version_2_with_negative, n_best_size=args.n_best_size, max_answer_length=args.max_answer_length, null_score_diff_threshold=args.null_score_diff_threshold, output_dir=args.output_dir, prefix=_A, ) # Format the result to the format the metric expects. if args.version_2_with_negative: UpperCAmelCase_ : List[str] = [ {"id": k, "prediction_text": v, "no_answer_probability": 0.0} for k, v in predictions.items() ] else: UpperCAmelCase_ : List[Any] = [{"id": k, "prediction_text": v} for k, v in predictions.items()] UpperCAmelCase_ : Optional[Any] = [{"id": ex["id"], "answers": ex[answer_column_name]} for ex in examples] return EvalPrediction(predictions=_A, label_ids=_A ) _a = load_metric('squad_v2' if args.version_2_with_negative else 'squad') # Evaluation! logger.info('Loading ONNX model %s for evaluation', args.onnx_model_path) with open(engine_name, 'rb') as f, trt.Runtime(TRT_LOGGER) as runtime, runtime.deserialize_cuda_engine( f.read() ) as engine, engine.create_execution_context() as context: # setup for TRT inferrence for i in range(len(input_names)): context.set_binding_shape(i, INPUT_SHAPE) assert context.all_binding_shapes_specified def __a ( __lowerCamelCase ): return trt.volume(engine.get_binding_shape(_A ) ) * engine.get_binding_dtype(_A ).itemsize # Allocate device memory for inputs and outputs. _a = [cuda.mem_alloc(binding_nbytes(binding)) for binding in engine if engine.binding_is_input(binding)] # Allocate output buffer _a = cuda.pagelocked_empty(tuple(context.get_binding_shape(3)), dtype=np.floataa) _a = cuda.pagelocked_empty(tuple(context.get_binding_shape(4)), dtype=np.floataa) _a = cuda.mem_alloc(h_outputa.nbytes) _a = cuda.mem_alloc(h_outputa.nbytes) # Create a stream in which to copy inputs/outputs and run inference. _a = cuda.Stream() # Evaluation logger.info('***** Running Evaluation *****') logger.info(f""" Num examples = {len(eval_dataset)}""") logger.info(f""" Batch size = {args.per_device_eval_batch_size}""") _a = 0.0 _a = 0 _a = timeit.default_timer() _a = None for step, batch in enumerate(eval_dataloader): _a = model_infer(batch, context, d_inputs, h_outputa, h_outputa, d_outputa, d_outputa, stream) total_time += infer_time niter += 1 _a = outputs _a = torch.tensor(start_logits) _a = torch.tensor(end_logits) # necessary to pad predictions and labels for being gathered _a = accelerator.pad_across_processes(start_logits, dim=1, pad_index=-100) _a = accelerator.pad_across_processes(end_logits, dim=1, pad_index=-100) _a = (accelerator.gather(start_logits).cpu().numpy(), accelerator.gather(end_logits).cpu().numpy()) _a = logits if all_preds is None else nested_concat(all_preds, logits, padding_index=-100) if all_preds is not None: _a = nested_truncate(all_preds, len(eval_dataset)) _a = timeit.default_timer() - start_time logger.info(' Evaluation done in total %f secs (%f sec per example)', evalTime, evalTime / len(eval_dataset)) # Inference time from TRT logger.info('Average Inference Time = {:.3f} ms'.format(total_time * 1_000 / niter)) logger.info('Total Inference Time = {:.3f} ms'.format(total_time * 1_000)) logger.info('Total Number of Inference = %d', niter) _a = post_processing_function(eval_examples, eval_dataset, all_preds) _a = metric.compute(predictions=prediction.predictions, references=prediction.label_ids) logger.info(f"""Evaluation metrics: {eval_metric}""")
354
"""simple docstring""" import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _a = logging.get_logger(__name__) _a = { 'facebook/detr-resnet-50': 'https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json', # See all DETR models at https://huggingface.co/models?filter=detr } class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """detr""" SCREAMING_SNAKE_CASE__ : str = ["""past_key_values"""] SCREAMING_SNAKE_CASE__ : Union[str, Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , lowercase_=True , lowercase_=None , lowercase_=3 , lowercase_=100 , lowercase_=6 , lowercase_=2048 , lowercase_=8 , lowercase_=6 , lowercase_=2048 , lowercase_=8 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=True , lowercase_="relu" , lowercase_=256 , lowercase_=0.1 , lowercase_=0.0 , lowercase_=0.0 , lowercase_=0.02 , lowercase_=1.0 , lowercase_=False , lowercase_="sine" , lowercase_="resnet50" , lowercase_=True , lowercase_=False , lowercase_=1 , lowercase_=5 , lowercase_=2 , lowercase_=1 , lowercase_=1 , lowercase_=5 , lowercase_=2 , lowercase_=0.1 , **lowercase_ , ): """simple docstring""" if backbone_config is not None and use_timm_backbone: raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`." ) if not use_timm_backbone: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) UpperCAmelCase_ : Union[str, Any] = CONFIG_MAPPING["resnet"](out_features=["stage4"] ) elif isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = backbone_config.get("model_type" ) UpperCAmelCase_ : int = CONFIG_MAPPING[backbone_model_type] UpperCAmelCase_ : Any = config_class.from_dict(lowercase_ ) # set timm attributes to None UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = None, None, None UpperCAmelCase_ : int = use_timm_backbone UpperCAmelCase_ : int = backbone_config UpperCAmelCase_ : List[Any] = num_channels UpperCAmelCase_ : int = num_queries UpperCAmelCase_ : Union[str, Any] = d_model UpperCAmelCase_ : str = encoder_ffn_dim UpperCAmelCase_ : Tuple = encoder_layers UpperCAmelCase_ : List[Any] = encoder_attention_heads UpperCAmelCase_ : Union[str, Any] = decoder_ffn_dim UpperCAmelCase_ : Optional[Any] = decoder_layers UpperCAmelCase_ : Union[str, Any] = decoder_attention_heads UpperCAmelCase_ : Optional[int] = dropout UpperCAmelCase_ : List[str] = attention_dropout UpperCAmelCase_ : Any = activation_dropout UpperCAmelCase_ : str = activation_function UpperCAmelCase_ : Tuple = init_std UpperCAmelCase_ : Optional[Any] = init_xavier_std UpperCAmelCase_ : Optional[Any] = encoder_layerdrop UpperCAmelCase_ : Optional[int] = decoder_layerdrop UpperCAmelCase_ : Tuple = encoder_layers UpperCAmelCase_ : int = auxiliary_loss UpperCAmelCase_ : Optional[Any] = position_embedding_type UpperCAmelCase_ : Tuple = backbone UpperCAmelCase_ : Optional[int] = use_pretrained_backbone UpperCAmelCase_ : Dict = dilation # Hungarian matcher UpperCAmelCase_ : Union[str, Any] = class_cost UpperCAmelCase_ : Any = bbox_cost UpperCAmelCase_ : int = giou_cost # Loss coefficients UpperCAmelCase_ : str = mask_loss_coefficient UpperCAmelCase_ : Any = dice_loss_coefficient UpperCAmelCase_ : Optional[Any] = bbox_loss_coefficient UpperCAmelCase_ : List[str] = giou_loss_coefficient UpperCAmelCase_ : List[Any] = eos_coefficient super().__init__(is_encoder_decoder=lowercase_ , **lowercase_ ) @property def UpperCamelCase__ ( self ): """simple docstring""" return self.encoder_attention_heads @property def UpperCamelCase__ ( self ): """simple docstring""" return self.d_model @classmethod def UpperCamelCase__ ( cls , lowercase_ , **lowercase_ ): """simple docstring""" return cls(backbone_config=lowercase_ , **lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: UpperCAmelCase_ : Union[str, Any] = self.backbone_config.to_dict() UpperCAmelCase_ : str = self.__class__.model_type return output class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : int = version.parse("""1.11""" ) @property def UpperCamelCase__ ( self ): """simple docstring""" return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("pixel_mask", {0: "batch"}), ] ) @property def UpperCamelCase__ ( self ): """simple docstring""" return 1E-5 @property def UpperCamelCase__ ( self ): """simple docstring""" return 12
23
0
"""simple docstring""" def __a ( __lowerCamelCase ): UpperCAmelCase_ : int = abs(__lowerCamelCase ) UpperCAmelCase_ : Any = 0 while n > 0: res += n % 10 n //= 10 return res def __a ( __lowerCamelCase ): UpperCAmelCase_ : Any = abs(__lowerCamelCase ) return n if n < 10 else n % 10 + sum_of_digits(n // 10 ) def __a ( __lowerCamelCase ): return sum(int(__lowerCamelCase ) for c in str(abs(__lowerCamelCase ) ) ) def __a ( ): from collections.abc import Callable from timeit import timeit def benchmark_a_function(__lowerCamelCase, __lowerCamelCase ) -> None: UpperCAmelCase_ : Any = f"""{func.__name__}({value})""" UpperCAmelCase_ : int = timeit(f"""__main__.{call}""", setup="import __main__" ) print(f"""{call:56} = {func(__lowerCamelCase )} -- {timing:.4f} seconds""" ) for value in (26_2144, 1125_8999_0684_2624, 126_7650_6002_2822_9401_4967_0320_5376): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(__lowerCamelCase, __lowerCamelCase ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
355
"""simple docstring""" _a = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(100_000)] def __a ( __lowerCamelCase ): UpperCAmelCase_ : Optional[int] = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 10_0000] number //= 10_0000 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution _a = [None] * 10_000_000 _a = True _a = False def __a ( __lowerCamelCase ): if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore UpperCAmelCase_ : Dict = chain(next_number(__lowerCamelCase ) ) UpperCAmelCase_ : List[str] = number_chain while number < 1000_0000: UpperCAmelCase_ : List[Any] = number_chain number *= 10 return number_chain def __a ( __lowerCamelCase = 1000_0000 ): for i in range(1, __lowerCamelCase ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(__lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() print(f"""{solution() = }""")
23
0
"""simple docstring""" from __future__ import annotations def __a ( __lowerCamelCase ): UpperCAmelCase_ : Tuple = len(__lowerCamelCase ) # We need to create solution object to save path. UpperCAmelCase_ : Tuple = [[0 for _ in range(__lowerCamelCase )] for _ in range(__lowerCamelCase )] UpperCAmelCase_ : Any = run_maze(__lowerCamelCase, 0, 0, __lowerCamelCase ) if solved: print("\n".join(str(__lowerCamelCase ) for row in solutions ) ) else: print("No solution exists!" ) return solved def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Tuple = len(__lowerCamelCase ) # Final check point. if i == j == (size - 1): UpperCAmelCase_ : Optional[int] = 1 return True UpperCAmelCase_ : List[Any] = (not i < 0) and (not j < 0) # Check lower bounds UpperCAmelCase_ : Dict = (i < size) and (j < size) # Check upper bounds if lower_flag and upper_flag: # check for already visited and block points. UpperCAmelCase_ : Tuple = (not solutions[i][j]) and (not maze[i][j]) if block_flag: # check visited UpperCAmelCase_ : List[Any] = 1 # check for directions if ( run_maze(__lowerCamelCase, i + 1, __lowerCamelCase, __lowerCamelCase ) or run_maze(__lowerCamelCase, __lowerCamelCase, j + 1, __lowerCamelCase ) or run_maze(__lowerCamelCase, i - 1, __lowerCamelCase, __lowerCamelCase ) or run_maze(__lowerCamelCase, __lowerCamelCase, j - 1, __lowerCamelCase ) ): return True UpperCAmelCase_ : List[str] = 0 return False return False if __name__ == "__main__": import doctest doctest.testmod()
356
"""simple docstring""" def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # Return True if there is node that has not iterated. UpperCAmelCase_ : List[Any] = [False] * len(__lowerCamelCase ) UpperCAmelCase_ : Any = [] queue.append(__lowerCamelCase ) UpperCAmelCase_ : Tuple = True while queue: UpperCAmelCase_ : str = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(__lowerCamelCase ) UpperCAmelCase_ : Any = True UpperCAmelCase_ : Union[str, Any] = u return visited[t] def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): # This array is filled by BFS and to store path UpperCAmelCase_ : List[str] = [-1] * (len(__lowerCamelCase )) UpperCAmelCase_ : Any = 0 while bfs(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : int = float("Inf" ) UpperCAmelCase_ : Tuple = sink while s != source: # Find the minimum value in select path UpperCAmelCase_ : Tuple = min(__lowerCamelCase, graph[parent[s]][s] ) UpperCAmelCase_ : Dict = parent[s] max_flow += path_flow UpperCAmelCase_ : Optional[Any] = sink while v != source: UpperCAmelCase_ : List[str] = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow UpperCAmelCase_ : Optional[int] = parent[v] return max_flow _a = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] _a , _a = 0, 5 print(ford_fulkerson(graph, source, sink))
23
0
"""simple docstring""" import os def __a ( ): UpperCAmelCase_ : int = os.path.dirname(os.path.realpath(A__ ) ) UpperCAmelCase_ : str = os.path.join(A__, "triangle.txt" ) with open(A__ ) as f: UpperCAmelCase_ : Any = f.readlines() UpperCAmelCase_ : Union[str, Any] = [] for line in triangle: UpperCAmelCase_ : str = [] for number in line.strip().split(" " ): numbers_from_line.append(int(A__ ) ) a.append(A__ ) for i in range(1, len(A__ ) ): for j in range(len(a[i] ) ): UpperCAmelCase_ : List[Any] = a[i - 1][j] if j != len(a[i - 1] ) else 0 UpperCAmelCase_ : Any = a[i - 1][j - 1] if j > 0 else 0 a[i][j] += max(A__, A__ ) return max(a[-1] ) if __name__ == "__main__": print(solution())
357
"""simple docstring""" import datasets _a = '\\n@InProceedings{conneau2018xnli,\n author = "Conneau, Alexis\n and Rinott, Ruty\n and Lample, Guillaume\n and Williams, Adina\n and Bowman, Samuel R.\n and Schwenk, Holger\n and Stoyanov, Veselin",\n title = "XNLI: Evaluating Cross-lingual Sentence Representations",\n booktitle = "Proceedings of the 2018 Conference on Empirical Methods\n in Natural Language Processing",\n year = "2018",\n publisher = "Association for Computational Linguistics",\n location = "Brussels, Belgium",\n}\n' _a = '\\nXNLI is a subset of a few thousand examples from MNLI which has been translated\ninto a 14 different languages (some low-ish resource). As with MNLI, the goal is\nto predict textual entailment (does sentence A imply/contradict/neither sentence\nB) and is a classification task (given two sentences, predict one of three\nlabels).\n' _a = '\nComputes XNLI score which is just simple accuracy.\nArgs:\n predictions: Predicted labels.\n references: Ground truth labels.\nReturns:\n \'accuracy\': accuracy\nExamples:\n\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> xnli_metric = datasets.load_metric("xnli")\n >>> results = xnli_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n' def __a ( __lowerCamelCase, __lowerCamelCase ): return (preds == labels).mean() @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64" if self.config_name != "sts-b" else "float32" ), "references": datasets.Value("int64" if self.config_name != "sts-b" else "float32" ), } ) , codebase_urls=[] , reference_urls=[] , format="numpy" , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" return {"accuracy": simple_accuracy(lowercase_ , lowercase_ )}
23
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _a = { 'configuration_bigbird_pegasus': [ 'BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BigBirdPegasusConfig', 'BigBirdPegasusOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST', 'BigBirdPegasusForCausalLM', 'BigBirdPegasusForConditionalGeneration', 'BigBirdPegasusForQuestionAnswering', 'BigBirdPegasusForSequenceClassification', 'BigBirdPegasusModel', 'BigBirdPegasusPreTrainedModel', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys _a = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
358
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from .feature_extraction_utils import BatchFeature, FeatureExtractionMixin from .utils import PaddingStrategy, TensorType, is_tf_tensor, is_torch_tensor, logging, to_numpy _a = logging.get_logger(__name__) class A_ (lowercase__ ): '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , **lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = feature_size UpperCAmelCase_ : Any = sampling_rate UpperCAmelCase_ : Any = padding_value UpperCAmelCase_ : str = kwargs.pop("padding_side" , "right" ) UpperCAmelCase_ : List[str] = kwargs.pop("return_attention_mask" , lowercase_ ) super().__init__(**lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = True , lowercase_ = None , lowercase_ = False , lowercase_ = None , lowercase_ = None , lowercase_ = None , ): """simple docstring""" # If we have a list of dicts, let's convert it in a dict of lists # We do this to allow using this method as a collate_fn function in PyTorch Dataloader if isinstance(lowercase_ , (list, tuple) ) and isinstance(processed_features[0] , (dict, BatchFeature) ): UpperCAmelCase_ : Dict = { key: [example[key] for example in processed_features] for key in processed_features[0].keys() } # The model's main input name, usually `input_values`, has be passed for padding if self.model_input_names[0] not in processed_features: raise ValueError( "You should supply an instance of `transformers.BatchFeature` or list of `transformers.BatchFeature`" F""" to this method that includes {self.model_input_names[0]}, but you provided""" F""" {list(processed_features.keys() )}""" ) UpperCAmelCase_ : Tuple = processed_features[self.model_input_names[0]] UpperCAmelCase_ : List[str] = ( return_attention_mask if return_attention_mask is not None else self.return_attention_mask ) if len(lowercase_ ) == 0: if return_attention_mask: UpperCAmelCase_ : Union[str, Any] = [] return processed_features # If we have PyTorch/TF tensors or lists as inputs, we cast them as Numpy arrays # and rebuild them afterwards if no return_tensors is specified # Note that we lose the specific device the tensor may be on for PyTorch UpperCAmelCase_ : List[str] = required_input[0] if isinstance(lowercase_ , (list, tuple) ): # first_element might be an empty list/tuple in some edge cases so we grab the first non empty element. UpperCAmelCase_ : Any = 0 while len(required_input[index] ) == 0: index += 1 if index < len(lowercase_ ): UpperCAmelCase_ : Optional[Any] = required_input[index][0] if return_tensors is None: if is_tf_tensor(lowercase_ ): UpperCAmelCase_ : Dict = "tf" elif is_torch_tensor(lowercase_ ): UpperCAmelCase_ : Any = "pt" elif isinstance(lowercase_ , (int, float, list, tuple, np.ndarray) ): UpperCAmelCase_ : str = "np" else: raise ValueError( F"""type of {first_element} unknown: {type(lowercase_ )}. """ "Should be one of a python, numpy, pytorch or tensorflow object." ) for key, value in processed_features.items(): if isinstance(value[0] , (int, float) ): UpperCAmelCase_ : Optional[int] = to_numpy(lowercase_ ) else: UpperCAmelCase_ : List[str] = [to_numpy(lowercase_ ) for v in value] # Convert padding_strategy in PaddingStrategy UpperCAmelCase_ : Dict = self._get_padding_strategies(padding=lowercase_ , max_length=lowercase_ ) UpperCAmelCase_ : str = processed_features[self.model_input_names[0]] UpperCAmelCase_ : int = len(lowercase_ ) if not all(len(lowercase_ ) == batch_size for v in processed_features.values() ): raise ValueError("Some items in the output dictionary have a different batch size than others." ) UpperCAmelCase_ : int = [] for i in range(lowercase_ ): UpperCAmelCase_ : str = {k: v[i] for k, v in processed_features.items()} # truncation UpperCAmelCase_ : List[str] = self._truncate( lowercase_ , max_length=lowercase_ , pad_to_multiple_of=lowercase_ , truncation=lowercase_ , ) truncated_inputs.append(lowercase_ ) if padding_strategy == PaddingStrategy.LONGEST: # make sure that `max_length` cannot be longer than the longest truncated length UpperCAmelCase_ : str = max(len(input_slice[self.model_input_names[0]] ) for input_slice in truncated_inputs ) UpperCAmelCase_ : Dict = PaddingStrategy.MAX_LENGTH UpperCAmelCase_ : List[str] = {} for i in range(lowercase_ ): # padding UpperCAmelCase_ : int = self._pad( truncated_inputs[i] , max_length=lowercase_ , padding_strategy=lowercase_ , pad_to_multiple_of=lowercase_ , return_attention_mask=lowercase_ , ) for key, value in outputs.items(): if key not in batch_outputs: UpperCAmelCase_ : Any = [] if value.dtype is np.dtype(np.floataa ): UpperCAmelCase_ : List[Any] = value.astype(np.floataa ) batch_outputs[key].append(lowercase_ ) return BatchFeature(lowercase_ , tensor_type=lowercase_ ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = PaddingStrategy.DO_NOT_PAD , lowercase_ = None , lowercase_ = None , ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = processed_features[self.model_input_names[0]] if padding_strategy == PaddingStrategy.LONGEST: UpperCAmelCase_ : Tuple = len(lowercase_ ) if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCAmelCase_ : Tuple = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCAmelCase_ : Dict = padding_strategy != PaddingStrategy.DO_NOT_PAD and len(lowercase_ ) < max_length if return_attention_mask and "attention_mask" not in processed_features: UpperCAmelCase_ : Optional[int] = np.ones(len(lowercase_ ) , dtype=np.intaa ) if needs_to_be_padded: UpperCAmelCase_ : Dict = max_length - len(lowercase_ ) if self.padding_side == "right": if return_attention_mask: UpperCAmelCase_ : List[Any] = np.pad( processed_features["attention_mask"] , (0, difference) ) UpperCAmelCase_ : Dict = ((0, difference), (0, 0)) if self.feature_size > 1 else (0, difference) UpperCAmelCase_ : Optional[Any] = np.pad( lowercase_ , lowercase_ , "constant" , constant_values=self.padding_value ) elif self.padding_side == "left": if return_attention_mask: UpperCAmelCase_ : Optional[Any] = np.pad( processed_features["attention_mask"] , (difference, 0) ) UpperCAmelCase_ : Dict = ((difference, 0), (0, 0)) if self.feature_size > 1 else (difference, 0) UpperCAmelCase_ : str = np.pad( lowercase_ , lowercase_ , "constant" , constant_values=self.padding_value ) else: raise ValueError("Invalid padding strategy:" + str(self.padding_side ) ) return processed_features def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = None , lowercase_ = None , ): """simple docstring""" if not truncation: return processed_features elif truncation and max_length is None: raise ValueError("When setting ``truncation=True``, make sure that ``max_length`` is defined." ) UpperCAmelCase_ : Optional[int] = processed_features[self.model_input_names[0]] # find `max_length` that fits `pad_to_multiple_of` if max_length is not None and pad_to_multiple_of is not None and (max_length % pad_to_multiple_of != 0): UpperCAmelCase_ : Union[str, Any] = ((max_length // pad_to_multiple_of) + 1) * pad_to_multiple_of UpperCAmelCase_ : Optional[Any] = len(lowercase_ ) > max_length if needs_to_be_truncated: UpperCAmelCase_ : int = processed_features[self.model_input_names[0]][:max_length] if "attention_mask" in processed_features: UpperCAmelCase_ : Dict = processed_features["attention_mask"][:max_length] return processed_features def UpperCamelCase__ ( self , lowercase_=False , lowercase_=None ): """simple docstring""" # Get padding strategy if padding is not False: if padding is True: UpperCAmelCase_ : Dict = PaddingStrategy.LONGEST # Default to pad to the longest sequence in the batch elif not isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : Optional[Any] = PaddingStrategy(lowercase_ ) elif isinstance(lowercase_ , lowercase_ ): UpperCAmelCase_ : int = padding else: UpperCAmelCase_ : str = PaddingStrategy.DO_NOT_PAD # Set max length if needed if max_length is None: if padding_strategy == PaddingStrategy.MAX_LENGTH: raise ValueError( F"""When setting ``padding={PaddingStrategy.MAX_LENGTH}``, make sure that max_length is defined""" ) # Test if we have a padding value if padding_strategy != PaddingStrategy.DO_NOT_PAD and (self.padding_value is None): raise ValueError( "Asking to pad but the feature_extractor does not have a padding value. Please select a value to use" " as `padding_value`. For example: `feature_extractor.padding_value = 0.0`." ) return padding_strategy
23
0
import dataclasses import json import sys import types from argparse import ArgumentDefaultsHelpFormatter, ArgumentParser, ArgumentTypeError from copy import copy from enum import Enum from inspect import isclass from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Literal, NewType, Optional, Tuple, Union, get_type_hints import yaml _a = NewType('DataClass', Any) _a = NewType('DataClassType', Any) def __a ( __lowerCamelCase ): if isinstance(a__, a__ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise ArgumentTypeError( f"""Truthy value expected: got {v} but expected one of yes/no, true/false, t/f, y/n, 1/0 (case insensitive).""" ) def __a ( __lowerCamelCase ): UpperCAmelCase_ : Tuple = {str(a__ ): choice for choice in choices} return lambda __lowerCamelCase : str_to_choice.get(a__, a__ ) def __a ( *, __lowerCamelCase = None, __lowerCamelCase = None, __lowerCamelCase = dataclasses.MISSING, __lowerCamelCase = dataclasses.MISSING, __lowerCamelCase = None, **__lowerCamelCase, ): if metadata is None: # Important, don't use as default param in function signature because dict is mutable and shared across function calls UpperCAmelCase_ : Any = {} if aliases is not None: UpperCAmelCase_ : List[Any] = aliases if help is not None: UpperCAmelCase_ : Optional[int] = help return dataclasses.field(metadata=a__, default=a__, default_factory=a__, **a__ ) class A_ (_lowercase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = 42 def __init__( self , lowercase_ , **lowercase_ ): """simple docstring""" # To make the default appear when using --help if "formatter_class" not in kwargs: UpperCAmelCase_ : List[str] = ArgumentDefaultsHelpFormatter super().__init__(**__UpperCamelCase ) if dataclasses.is_dataclass(__UpperCamelCase ): UpperCAmelCase_ : Union[str, Any] = [dataclass_types] UpperCAmelCase_ : Optional[Any] = list(__UpperCamelCase ) for dtype in self.dataclass_types: self._add_dataclass_arguments(__UpperCamelCase ) @staticmethod def UpperCamelCase__ ( lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = F"""--{field.name}""" UpperCAmelCase_ : int = field.metadata.copy() # field.metadata is not used at all by Data Classes, # it is provided as a third-party extension mechanism. if isinstance(field.type , __UpperCamelCase ): raise RuntimeError( "Unresolved type detected, which should have been done with the help of " "`typing.get_type_hints` method by default" ) UpperCAmelCase_ : Optional[Any] = kwargs.pop("aliases" , [] ) if isinstance(__UpperCamelCase , __UpperCamelCase ): UpperCAmelCase_ : str = [aliases] UpperCAmelCase_ : List[Any] = getattr(field.type , "__origin__" , field.type ) if origin_type is Union or (hasattr(__UpperCamelCase , "UnionType" ) and isinstance(__UpperCamelCase , types.UnionType )): if str not in field.type.__args__ and ( len(field.type.__args__ ) != 2 or type(__UpperCamelCase ) not in field.type.__args__ ): raise ValueError( "Only `Union[X, NoneType]` (i.e., `Optional[X]`) is allowed for `Union` because" " the argument parser only supports one type per argument." F""" Problem encountered in field \'{field.name}\'.""" ) if type(__UpperCamelCase ) not in field.type.__args__: # filter `str` in Union UpperCAmelCase_ : int = field.type.__args__[0] if field.type.__args__[1] == str else field.type.__args__[1] UpperCAmelCase_ : List[str] = getattr(field.type , "__origin__" , field.type ) elif bool not in field.type.__args__: # filter `NoneType` in Union (except for `Union[bool, NoneType]`) UpperCAmelCase_ : Optional[int] = ( field.type.__args__[0] if isinstance(__UpperCamelCase , field.type.__args__[1] ) else field.type.__args__[1] ) UpperCAmelCase_ : Dict = getattr(field.type , "__origin__" , field.type ) # A variable to store kwargs for a boolean field, if needed # so that we can init a `no_*` complement argument (see below) UpperCAmelCase_ : List[Any] = {} if origin_type is Literal or (isinstance(field.type , __UpperCamelCase ) and issubclass(field.type , __UpperCamelCase )): if origin_type is Literal: UpperCAmelCase_ : List[Any] = field.type.__args__ else: UpperCAmelCase_ : List[Any] = [x.value for x in field.type] UpperCAmelCase_ : int = make_choice_type_function(kwargs["choices"] ) if field.default is not dataclasses.MISSING: UpperCAmelCase_ : List[Any] = field.default else: UpperCAmelCase_ : Optional[int] = True elif field.type is bool or field.type == Optional[bool]: # Copy the currect kwargs to use to instantiate a `no_*` complement argument below. # We do not initialize it here because the `no_*` alternative must be instantiated after the real argument UpperCAmelCase_ : Optional[int] = copy(__UpperCamelCase ) # Hack because type=bool in argparse does not behave as we want. UpperCAmelCase_ : Union[str, Any] = string_to_bool if field.type is bool or (field.default is not None and field.default is not dataclasses.MISSING): # Default value is False if we have no default when of type bool. UpperCAmelCase_ : Dict = False if field.default is dataclasses.MISSING else field.default # This is the value that will get picked if we don't include --field_name in any way UpperCAmelCase_ : Union[str, Any] = default # This tells argparse we accept 0 or 1 value after --field_name UpperCAmelCase_ : Optional[int] = "?" # This is the value that will get picked if we do --field_name (without value) UpperCAmelCase_ : Any = True elif isclass(__UpperCamelCase ) and issubclass(__UpperCamelCase , __UpperCamelCase ): UpperCAmelCase_ : Optional[Any] = field.type.__args__[0] UpperCAmelCase_ : List[Any] = "+" if field.default_factory is not dataclasses.MISSING: UpperCAmelCase_ : Optional[int] = field.default_factory() elif field.default is dataclasses.MISSING: UpperCAmelCase_ : List[str] = True else: UpperCAmelCase_ : Any = field.type if field.default is not dataclasses.MISSING: UpperCAmelCase_ : Optional[Any] = field.default elif field.default_factory is not dataclasses.MISSING: UpperCAmelCase_ : str = field.default_factory() else: UpperCAmelCase_ : Optional[Any] = True parser.add_argument(__UpperCamelCase , *__UpperCamelCase , **__UpperCamelCase ) # Add a complement `no_*` argument for a boolean field AFTER the initial field has already been added. # Order is important for arguments with the same destination! # We use a copy of earlier kwargs because the original kwargs have changed a lot before reaching down # here and we do not need those changes/additional keys. if field.default is True and (field.type is bool or field.type == Optional[bool]): UpperCAmelCase_ : List[str] = False parser.add_argument(F"""--no_{field.name}""" , action="store_false" , dest=field.name , **__UpperCamelCase ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" if hasattr(__UpperCamelCase , "_argument_group_name" ): UpperCAmelCase_ : List[Any] = self.add_argument_group(dtype._argument_group_name ) else: UpperCAmelCase_ : Any = self try: UpperCAmelCase_ : Tuple = get_type_hints(__UpperCamelCase ) except NameError: raise RuntimeError( F"""Type resolution failed for {dtype}. Try declaring the class in global scope or """ "removing line of `from __future__ import annotations` which opts in Postponed " "Evaluation of Annotations (PEP 563)" ) except TypeError as ex: # Remove this block when we drop Python 3.9 support if sys.version_info[:2] < (3, 10) and "unsupported operand type(s) for |" in str(__UpperCamelCase ): UpperCAmelCase_ : str = ".".join(map(__UpperCamelCase , sys.version_info[:3] ) ) raise RuntimeError( F"""Type resolution failed for {dtype} on Python {python_version}. Try removing """ "line of `from __future__ import annotations` which opts in union types as " "`X | Y` (PEP 604) via Postponed Evaluation of Annotations (PEP 563). To " "support Python versions that lower than 3.10, you need to use " "`typing.Union[X, Y]` instead of `X | Y` and `typing.Optional[X]` instead of " "`X | None`." ) from ex raise for field in dataclasses.fields(__UpperCamelCase ): if not field.init: continue UpperCAmelCase_ : List[str] = type_hints[field.name] self._parse_dataclass_field(__UpperCamelCase , __UpperCamelCase ) def UpperCamelCase__ ( self , lowercase_=None , lowercase_=False , lowercase_=True , lowercase_=None , lowercase_=None , ): """simple docstring""" if args_file_flag or args_filename or (look_for_args_file and len(sys.argv )): UpperCAmelCase_ : List[Any] = [] if args_filename: args_files.append(Path(__UpperCamelCase ) ) elif look_for_args_file and len(sys.argv ): args_files.append(Path(sys.argv[0] ).with_suffix(".args" ) ) # args files specified via command line flag should overwrite default args files so we add them last if args_file_flag: # Create special parser just to extract the args_file_flag values UpperCAmelCase_ : Dict = ArgumentParser() args_file_parser.add_argument(__UpperCamelCase , type=__UpperCamelCase , action="append" ) # Use only remaining args for further parsing (remove the args_file_flag) UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = args_file_parser.parse_known_args(args=__UpperCamelCase ) UpperCAmelCase_ : Tuple = vars(__UpperCamelCase ).get(args_file_flag.lstrip("-" ) , __UpperCamelCase ) if cmd_args_file_paths: args_files.extend([Path(__UpperCamelCase ) for p in cmd_args_file_paths] ) UpperCAmelCase_ : str = [] for args_file in args_files: if args_file.exists(): file_args += args_file.read_text().split() # in case of duplicate arguments the last one has precedence # args specified via the command line should overwrite args from files, so we add them last UpperCAmelCase_ : Any = file_args + args if args is not None else file_args + sys.argv[1:] UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = self.parse_known_args(args=__UpperCamelCase ) UpperCAmelCase_ : List[Any] = [] for dtype in self.dataclass_types: UpperCAmelCase_ : Any = {f.name for f in dataclasses.fields(__UpperCamelCase ) if f.init} UpperCAmelCase_ : Union[str, Any] = {k: v for k, v in vars(__UpperCamelCase ).items() if k in keys} for k in keys: delattr(__UpperCamelCase , __UpperCamelCase ) UpperCAmelCase_ : List[Any] = dtype(**__UpperCamelCase ) outputs.append(__UpperCamelCase ) if len(namespace.__dict__ ) > 0: # additional namespace. outputs.append(__UpperCamelCase ) if return_remaining_strings: return (*outputs, remaining_args) else: if remaining_args: raise ValueError(F"""Some specified arguments are not used by the HfArgumentParser: {remaining_args}""" ) return (*outputs,) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = False ): """simple docstring""" UpperCAmelCase_ : str = set(args.keys() ) UpperCAmelCase_ : List[str] = [] for dtype in self.dataclass_types: UpperCAmelCase_ : List[str] = {f.name for f in dataclasses.fields(__UpperCamelCase ) if f.init} UpperCAmelCase_ : Any = {k: v for k, v in args.items() if k in keys} unused_keys.difference_update(inputs.keys() ) UpperCAmelCase_ : Optional[int] = dtype(**__UpperCamelCase ) outputs.append(__UpperCamelCase ) if not allow_extra_keys and unused_keys: raise ValueError(F"""Some keys are not used by the HfArgumentParser: {sorted(__UpperCamelCase )}""" ) return tuple(__UpperCamelCase ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = False ): """simple docstring""" with open(Path(__UpperCamelCase ) , encoding="utf-8" ) as open_json_file: UpperCAmelCase_ : Optional[Any] = json.loads(open_json_file.read() ) UpperCAmelCase_ : Tuple = self.parse_dict(__UpperCamelCase , allow_extra_keys=__UpperCamelCase ) return tuple(__UpperCamelCase ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = False ): """simple docstring""" UpperCAmelCase_ : int = self.parse_dict(yaml.safe_load(Path(__UpperCamelCase ).read_text() ) , allow_extra_keys=__UpperCamelCase ) return tuple(__UpperCamelCase )
359
"""simple docstring""" import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : str = torch.nn.Linear(10 , 10 ) UpperCAmelCase_ : List[str] = torch.optim.SGD(model.parameters() , 0.1 ) UpperCAmelCase_ : Optional[Any] = Accelerator() UpperCAmelCase_ : Tuple = accelerator.prepare(lowercase_ ) try: pickle.loads(pickle.dumps(lowercase_ ) ) except Exception as e: self.fail(F"""Accelerated optimizer pickling failed with {e}""" ) AcceleratorState._reset_state()
23
0
"""simple docstring""" from pathlib import Path import numpy as np from PIL import Image def __a ( __lowerCamelCase ): UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ : str = rgb[:, :, 0], rgb[:, :, 1], rgb[:, :, 2] return 0.2989 * r + 0.5870 * g + 0.1140 * b def __a ( __lowerCamelCase ): return (gray > 127) & (gray <= 255) def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Optional[int] = np.zeros_like(__lowerCamelCase ) UpperCAmelCase_ : Union[str, Any] = np.zeros( (image.shape[0] + kernel.shape[0] - 1, image.shape[1] + kernel.shape[1] - 1) ) # Copy image to padded image UpperCAmelCase_ : List[Any] = image # Iterate over image & apply kernel for x in range(image.shape[1] ): for y in range(image.shape[0] ): UpperCAmelCase_ : List[Any] = ( kernel * image_padded[y : y + kernel.shape[0], x : x + kernel.shape[1]] ).sum() UpperCAmelCase_ : List[str] = int(summation > 0 ) return output if __name__ == "__main__": # read original image _a = Path(__file__).resolve().parent / 'image_data' / 'lena.jpg' _a = np.array(Image.open(lena_path)) # kernel to be applied _a = np.array([[0, 1, 0], [1, 1, 1], [0, 1, 0]]) _a = dilation(gray_to_binary(rgb_to_gray(lena)), structuring_element) # Save the output image _a = Image.fromarray(output).convert('RGB') pil_img.save('result_dilation.png')
360
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a = logging.get_logger(__name__) _a = {'ctrl': 'https://huggingface.co/ctrl/resolve/main/config.json'} class A_ (lowercase__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Dict = """ctrl""" SCREAMING_SNAKE_CASE__ : Optional[int] = ["""past_key_values"""] SCREAMING_SNAKE_CASE__ : List[str] = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , lowercase_=24_6534 , lowercase_=256 , lowercase_=1280 , lowercase_=8192 , lowercase_=48 , lowercase_=16 , lowercase_=0.1 , lowercase_=0.1 , lowercase_=1E-6 , lowercase_=0.02 , lowercase_=True , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : Tuple = vocab_size UpperCAmelCase_ : Union[str, Any] = n_positions UpperCAmelCase_ : List[str] = n_embd UpperCAmelCase_ : Dict = n_layer UpperCAmelCase_ : Optional[int] = n_head UpperCAmelCase_ : List[str] = dff UpperCAmelCase_ : Tuple = resid_pdrop UpperCAmelCase_ : Optional[Any] = embd_pdrop UpperCAmelCase_ : str = layer_norm_epsilon UpperCAmelCase_ : List[str] = initializer_range UpperCAmelCase_ : List[str] = use_cache super().__init__(**lowercase_ )
23
0
"""simple docstring""" import math def __a ( __lowerCamelCase ): UpperCAmelCase_ : Optional[int] = [] UpperCAmelCase_ : Optional[Any] = 2 UpperCAmelCase_ : Dict = int(math.sqrt(__lowerCAmelCase ) ) # Size of every segment UpperCAmelCase_ : Tuple = [True] * (end + 1) UpperCAmelCase_ : Any = [] while start <= end: if temp[start] is True: in_prime.append(__lowerCAmelCase ) for i in range(start * start, end + 1, __lowerCAmelCase ): UpperCAmelCase_ : Any = False start += 1 prime += in_prime UpperCAmelCase_ : Union[str, Any] = end + 1 UpperCAmelCase_ : Optional[int] = min(2 * end, __lowerCAmelCase ) while low <= n: UpperCAmelCase_ : str = [True] * (high - low + 1) for each in in_prime: UpperCAmelCase_ : Optional[int] = math.floor(low / each ) * each if t < low: t += each for j in range(__lowerCAmelCase, high + 1, __lowerCAmelCase ): UpperCAmelCase_ : Optional[Any] = False for j in range(len(__lowerCAmelCase ) ): if temp[j] is True: prime.append(j + low ) UpperCAmelCase_ : Dict = high + 1 UpperCAmelCase_ : List[str] = min(high + end, __lowerCAmelCase ) return prime print(sieve(10**6))
361
"""simple docstring""" def __a ( __lowerCamelCase ): assert isinstance(__lowerCamelCase, __lowerCamelCase ), f"""The input value of [n={number}] is not an integer""" if number == 1: return 2 elif number < 1: UpperCAmelCase_ : str = f"""The input value of [n={number}] has to be > 0""" raise ValueError(__lowerCamelCase ) else: UpperCAmelCase_ : List[str] = sylvester(number - 1 ) UpperCAmelCase_ : List[str] = num - 1 UpperCAmelCase_ : List[str] = num return lower * upper + 1 if __name__ == "__main__": print(f"""The 8th number in Sylvester's sequence: {sylvester(8)}""")
23
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _a = { 'configuration_convnext': ['CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ConvNextConfig', 'ConvNextOnnxConfig'] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = ['ConvNextFeatureExtractor'] _a = ['ConvNextImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'ConvNextForImageClassification', 'ConvNextModel', 'ConvNextPreTrainedModel', 'ConvNextBackbone', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ 'TFConvNextForImageClassification', 'TFConvNextModel', 'TFConvNextPreTrainedModel', ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys _a = _LazyModule(__name__, globals()['__file__'], _import_structure)
362
"""simple docstring""" import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class A_ (lowercase__ ,lowercase__ ,unittest.TestCase ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[str] = IFImgaImgSuperResolutionPipeline SCREAMING_SNAKE_CASE__ : Optional[int] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {"""width""", """height"""} SCREAMING_SNAKE_CASE__ : List[str] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""original_image"""} ) SCREAMING_SNAKE_CASE__ : List[Any] = PipelineTesterMixin.required_optional_params - {"""latents"""} def UpperCamelCase__ ( self ): """simple docstring""" return self._get_superresolution_dummy_components() def UpperCamelCase__ ( self , lowercase_ , lowercase_=0 ): """simple docstring""" if str(lowercase_ ).startswith("mps" ): UpperCAmelCase_ : Optional[Any] = torch.manual_seed(lowercase_ ) else: UpperCAmelCase_ : Union[str, Any] = torch.Generator(device=lowercase_ ).manual_seed(lowercase_ ) UpperCAmelCase_ : Any = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) UpperCAmelCase_ : Optional[int] = floats_tensor((1, 3, 16, 16) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) UpperCAmelCase_ : int = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def UpperCamelCase__ ( self ): """simple docstring""" # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def UpperCamelCase__ ( self ): """simple docstring""" self._test_save_load_local() def UpperCamelCase__ ( self ): """simple docstring""" self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
23
0
"""simple docstring""" def __a ( __lowerCamelCase = 400_0000 ): UpperCAmelCase_ : Any = [0, 1] UpperCAmelCase_ : Dict = 0 while fib[i] <= n: fib.append(fib[i] + fib[i + 1] ) if fib[i + 2] > n: break i += 1 UpperCAmelCase_ : Dict = 0 for j in range(len(__snake_case ) - 1 ): if fib[j] % 2 == 0: total += fib[j] return total if __name__ == "__main__": print(f"""{solution() = }""")
363
"""simple docstring""" import os import shutil import tempfile import unittest import numpy as np from transformers import AutoTokenizer, BarkProcessor from transformers.testing_utils import require_torch, slow @require_torch class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Union[str, Any] = "ylacombe/bark-small" UpperCAmelCase_ : Union[str, Any] = tempfile.mkdtemp() UpperCAmelCase_ : List[str] = "en_speaker_1" UpperCAmelCase_ : Tuple = "This is a test string" UpperCAmelCase_ : List[Any] = "speaker_embeddings_path.json" UpperCAmelCase_ : Any = "speaker_embeddings" def UpperCamelCase__ ( self , **lowercase_ ): """simple docstring""" return AutoTokenizer.from_pretrained(self.checkpoint , **lowercase_ ) def UpperCamelCase__ ( self ): """simple docstring""" shutil.rmtree(self.tmpdirname ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.get_tokenizer() UpperCAmelCase_ : Union[str, Any] = BarkProcessor(tokenizer=lowercase_ ) processor.save_pretrained(self.tmpdirname ) UpperCAmelCase_ : Optional[int] = BarkProcessor.from_pretrained(self.tmpdirname ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[Any] = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) processor.save_pretrained( self.tmpdirname , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , speaker_embeddings_directory=self.speaker_embeddings_directory , ) UpperCAmelCase_ : Dict = self.get_tokenizer(bos_token="(BOS)" , eos_token="(EOS)" ) UpperCAmelCase_ : Union[str, Any] = BarkProcessor.from_pretrained( self.tmpdirname , self.speaker_embeddings_dict_path , bos_token="(BOS)" , eos_token="(EOS)" , ) self.assertEqual(processor.tokenizer.get_vocab() , tokenizer_add_kwargs.get_vocab() ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = BarkProcessor.from_pretrained( pretrained_processor_name_or_path=self.checkpoint , speaker_embeddings_dict_path=self.speaker_embeddings_dict_path , ) UpperCAmelCase_ : int = 35 UpperCAmelCase_ : Optional[Any] = 2 UpperCAmelCase_ : List[Any] = 8 UpperCAmelCase_ : Optional[Any] = { "semantic_prompt": np.ones(lowercase_ ), "coarse_prompt": np.ones((nb_codebooks_coarse, seq_len) ), "fine_prompt": np.ones((nb_codebooks_total, seq_len) ), } # test providing already loaded voice_preset UpperCAmelCase_ : Dict = processor(text=self.input_string , voice_preset=lowercase_ ) UpperCAmelCase_ : List[str] = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowercase_ , np.array([] ) ).tolist() ) # test loading voice preset from npz file UpperCAmelCase_ : Tuple = os.path.join(self.tmpdirname , "file.npz" ) np.savez(lowercase_ , **lowercase_ ) UpperCAmelCase_ : Optional[int] = processor(text=self.input_string , voice_preset=lowercase_ ) UpperCAmelCase_ : List[str] = inputs["history_prompt"] for key in voice_preset: self.assertListEqual(voice_preset[key].tolist() , processed_voice_preset.get(lowercase_ , np.array([] ) ).tolist() ) # test loading voice preset from the hub UpperCAmelCase_ : Tuple = processor(text=self.input_string , voice_preset=self.voice_preset ) def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Optional[int] = self.get_tokenizer() UpperCAmelCase_ : Optional[Any] = BarkProcessor(tokenizer=lowercase_ ) UpperCAmelCase_ : Tuple = processor(text=self.input_string ) UpperCAmelCase_ : Union[str, Any] = tokenizer( self.input_string , padding="max_length" , max_length=256 , add_special_tokens=lowercase_ , return_attention_mask=lowercase_ , return_token_type_ids=lowercase_ , ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key].squeeze().tolist() )
23
0
"""simple docstring""" from collections.abc import Callable def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : float = a UpperCAmelCase_ : float = b if function(_a ) == 0: # one of the a or b is a root for the function return a elif function(_a ) == 0: return b elif ( function(_a ) * function(_a ) > 0 ): # if none of these are root and they are both positive or negative, # then this algorithm can't find the root raise ValueError("could not find root in given interval." ) else: UpperCAmelCase_ : float = start + (end - start) / 2.0 while abs(start - mid ) > 10**-7: # until precisely equals to 10^-7 if function(_a ) == 0: return mid elif function(_a ) * function(_a ) < 0: UpperCAmelCase_ : List[str] = mid else: UpperCAmelCase_ : int = mid UpperCAmelCase_ : Dict = start + (end - start) / 2.0 return mid def __a ( __lowerCamelCase ): return x**3 - 2 * x - 5 if __name__ == "__main__": print(bisection(f, 1, 1_000)) import doctest doctest.testmod()
364
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() _a = logging.get_logger(__name__) def __a ( __lowerCamelCase, __lowerCamelCase=False ): UpperCAmelCase_ : Optional[int] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f"""blocks.{i}.norm1.weight""", f"""deit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((f"""blocks.{i}.norm1.bias""", f"""deit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((f"""blocks.{i}.attn.proj.weight""", f"""deit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.attn.proj.bias""", f"""deit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((f"""blocks.{i}.norm2.weight""", f"""deit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((f"""blocks.{i}.norm2.bias""", f"""deit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.weight""", f"""deit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc1.bias""", f"""deit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.weight""", f"""deit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((f"""blocks.{i}.mlp.fc2.bias""", f"""deit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ("cls_token", "deit.embeddings.cls_token"), ("dist_token", "deit.embeddings.distillation_token"), ("patch_embed.proj.weight", "deit.embeddings.patch_embeddings.projection.weight"), ("patch_embed.proj.bias", "deit.embeddings.patch_embeddings.projection.bias"), ("pos_embed", "deit.embeddings.position_embeddings"), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" UpperCAmelCase_ : Dict = [(pair[0], pair[1][4:]) if pair[1].startswith("deit" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("norm.weight", "deit.layernorm.weight"), ("norm.bias", "deit.layernorm.bias"), ("head.weight", "cls_classifier.weight"), ("head.bias", "cls_classifier.bias"), ("head_dist.weight", "distillation_classifier.weight"), ("head_dist.bias", "distillation_classifier.bias"), ] ) return rename_keys def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase=False ): for i in range(config.num_hidden_layers ): if base_model: UpperCAmelCase_ : int = "" else: UpperCAmelCase_ : Union[str, Any] = "deit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) UpperCAmelCase_ : Tuple = state_dict.pop(f"""blocks.{i}.attn.qkv.weight""" ) UpperCAmelCase_ : Dict = state_dict.pop(f"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict UpperCAmelCase_ : Union[str, Any] = in_proj_weight[ : config.hidden_size, : ] UpperCAmelCase_ : Any = in_proj_bias[: config.hidden_size] UpperCAmelCase_ : Optional[Any] = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] UpperCAmelCase_ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] UpperCAmelCase_ : List[Any] = in_proj_weight[ -config.hidden_size :, : ] UpperCAmelCase_ : str = in_proj_bias[-config.hidden_size :] def __a ( __lowerCamelCase, __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : Tuple = dct.pop(__lowerCamelCase ) UpperCAmelCase_ : Tuple = val def __a ( ): UpperCAmelCase_ : Union[str, Any] = "http://images.cocodataset.org/val2017/000000039769.jpg" UpperCAmelCase_ : str = Image.open(requests.get(__lowerCamelCase, stream=__lowerCamelCase ).raw ) return im @torch.no_grad() def __a ( __lowerCamelCase, __lowerCamelCase ): UpperCAmelCase_ : List[str] = DeiTConfig() # all deit models have fine-tuned heads UpperCAmelCase_ : Optional[int] = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size UpperCAmelCase_ : Tuple = 1000 UpperCAmelCase_ : str = "huggingface/label-files" UpperCAmelCase_ : str = "imagenet-1k-id2label.json" UpperCAmelCase_ : List[Any] = json.load(open(hf_hub_download(__lowerCamelCase, __lowerCamelCase, repo_type="dataset" ), "r" ) ) UpperCAmelCase_ : List[str] = {int(__lowerCamelCase ): v for k, v in idalabel.items()} UpperCAmelCase_ : Any = idalabel UpperCAmelCase_ : int = {v: k for k, v in idalabel.items()} UpperCAmelCase_ : Any = int(deit_name[-6:-4] ) UpperCAmelCase_ : Dict = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("tiny" ): UpperCAmelCase_ : Any = 192 UpperCAmelCase_ : Union[str, Any] = 768 UpperCAmelCase_ : Union[str, Any] = 12 UpperCAmelCase_ : int = 3 elif deit_name[9:].startswith("small" ): UpperCAmelCase_ : List[str] = 384 UpperCAmelCase_ : List[str] = 1536 UpperCAmelCase_ : Dict = 12 UpperCAmelCase_ : Any = 6 if deit_name[9:].startswith("base" ): pass elif deit_name[4:].startswith("large" ): UpperCAmelCase_ : int = 1024 UpperCAmelCase_ : List[Any] = 4096 UpperCAmelCase_ : Optional[int] = 24 UpperCAmelCase_ : int = 16 # load original model from timm UpperCAmelCase_ : Union[str, Any] = timm.create_model(__lowerCamelCase, pretrained=__lowerCamelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys UpperCAmelCase_ : Optional[Any] = timm_model.state_dict() UpperCAmelCase_ : Tuple = create_rename_keys(__lowerCamelCase, __lowerCamelCase ) for src, dest in rename_keys: rename_key(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) read_in_q_k_v(__lowerCamelCase, __lowerCamelCase, __lowerCamelCase ) # load HuggingFace model UpperCAmelCase_ : str = DeiTForImageClassificationWithTeacher(__lowerCamelCase ).eval() model.load_state_dict(__lowerCamelCase ) # Check outputs on an image, prepared by DeiTImageProcessor UpperCAmelCase_ : Union[str, Any] = int( (256 / 224) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 UpperCAmelCase_ : Optional[Any] = DeiTImageProcessor(size=__lowerCamelCase, crop_size=config.image_size ) UpperCAmelCase_ : Any = image_processor(images=prepare_img(), return_tensors="pt" ) UpperCAmelCase_ : int = encoding["pixel_values"] UpperCAmelCase_ : Optional[Any] = model(__lowerCamelCase ) UpperCAmelCase_ : Any = timm_model(__lowerCamelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__lowerCamelCase, outputs.logits, atol=1E-3 ) Path(__lowerCamelCase ).mkdir(exist_ok=__lowerCamelCase ) print(f"""Saving model {deit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCamelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCamelCase ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( '--deit_name', default='vit_deit_base_distilled_patch16_224', type=str, help='Name of the DeiT timm model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) _a = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
23
0
"""simple docstring""" from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Sequence, Value from .base import TaskTemplate @dataclass(frozen=lowercase__ ) class A_ (lowercase__ ): '''simple docstring''' # `task` is not a ClassVar since we want it to be part of the `asdict` output for JSON serialization SCREAMING_SNAKE_CASE__ : str = field(default="""question-answering-extractive""" ,metadata={"""include_in_asdict_even_if_is_default""": True} ) SCREAMING_SNAKE_CASE__ : ClassVar[Features] = Features({"""question""": Value("""string""" ), """context""": Value("""string""" )} ) SCREAMING_SNAKE_CASE__ : ClassVar[Features] = Features( { """answers""": Sequence( { """text""": Value("""string""" ), """answer_start""": Value("""int32""" ), } ) } ) SCREAMING_SNAKE_CASE__ : str = "question" SCREAMING_SNAKE_CASE__ : str = "context" SCREAMING_SNAKE_CASE__ : str = "answers" @property def UpperCamelCase__ ( self ): """simple docstring""" return {self.question_column: "question", self.context_column: "context", self.answers_column: "answers"}
365
"""simple docstring""" import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights UpperCAmelCase_ : Optional[Any] = FlaxDiffusionPipeline.from_pretrained( "hf-internal-testing/tiny-stable-diffusion-pipe" , safety_checker=lowercase_ , cache_dir=lowercase_ ) UpperCAmelCase_ : List[Any] = [t[-1] for t in os.walk(os.path.join(lowercase_ , os.listdir(lowercase_ )[0] , "snapshots" ) )] UpperCAmelCase_ : Dict = [item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith(".bin" ) for f in files ) @slow @require_flax class A_ (unittest.TestCase ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = FlaxStableDiffusionPipeline.from_pretrained( "hf-internal-testing/tiny-stable-diffusion-pipe" , safety_checker=lowercase_ ) UpperCAmelCase_ : Tuple = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : List[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase_ : List[str] = 4 UpperCAmelCase_ : Tuple = jax.device_count() UpperCAmelCase_ : Optional[int] = num_samples * [prompt] UpperCAmelCase_ : List[Any] = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : int = replicate(lowercase_ ) UpperCAmelCase_ : str = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : Dict = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1_51_47_45 ) < 1E-3 assert np.abs(np.abs(lowercase_ , dtype=np.floataa ).sum() - 4_99_47.8_75 ) < 5E-1 UpperCAmelCase_ : List[Any] = pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(lowercase_ ) == num_samples def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="flax" , safety_checker=lowercase_ ) UpperCAmelCase_ : Optional[int] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : str = jax.random.PRNGKey(0 ) UpperCAmelCase_ : Union[str, Any] = 50 UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[str] = num_samples * [prompt] UpperCAmelCase_ : Union[str, Any] = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Any = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : int = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05_65_24_01) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_38_38_08.2) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ ) UpperCAmelCase_ : Any = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : str = jax.random.PRNGKey(0 ) UpperCAmelCase_ : str = 50 UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : Any = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Dict = replicate(lowercase_ ) UpperCAmelCase_ : str = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Union[str, Any] = shard(lowercase_ ) UpperCAmelCase_ : List[Any] = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ , UpperCAmelCase_ : str = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa ) UpperCAmelCase_ : List[Any] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : Dict = jax.random.PRNGKey(0 ) UpperCAmelCase_ : Optional[int] = 50 UpperCAmelCase_ : Optional[int] = jax.device_count() UpperCAmelCase_ : str = num_samples * [prompt] UpperCAmelCase_ : int = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : Union[str, Any] = replicate(lowercase_ ) UpperCAmelCase_ : Union[str, Any] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[int] = shard(lowercase_ ) UpperCAmelCase_ : Any = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04_00_39_06) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_37_35_16.75) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[Any] = FlaxDDIMScheduler( beta_start=0.0_00_85 , beta_end=0.0_12 , beta_schedule="scaled_linear" , set_alpha_to_one=lowercase_ , steps_offset=1 , ) UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , scheduler=lowercase_ , safety_checker=lowercase_ , ) UpperCAmelCase_ : List[Any] = scheduler.create_state() UpperCAmelCase_ : int = scheduler_state UpperCAmelCase_ : Union[str, Any] = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : Optional[Any] = jax.random.PRNGKey(0 ) UpperCAmelCase_ : int = 50 UpperCAmelCase_ : str = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : int = pipeline.prepare_inputs(lowercase_ ) # shard inputs and rng UpperCAmelCase_ : int = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = jax.random.split(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = shard(lowercase_ ) UpperCAmelCase_ : Any = pipeline(lowercase_ , lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.0_45_04_39_45) ) < 1E-3 assert np.abs((np.abs(lowercase_ , dtype=np.floataa ).sum() - 2_34_76_93.5) ) < 5E-1 def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : int = ( "A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of" " field, close up, split lighting, cinematic" ) UpperCAmelCase_ : List[str] = jax.device_count() UpperCAmelCase_ : List[Any] = num_samples * [prompt] UpperCAmelCase_ : Union[str, Any] = jax.random.split(jax.random.PRNGKey(0 ) , lowercase_ ) UpperCAmelCase_ , UpperCAmelCase_ : str = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ , ) UpperCAmelCase_ : Any = replicate(lowercase_ ) UpperCAmelCase_ : List[str] = pipeline.prepare_inputs(lowercase_ ) UpperCAmelCase_ : List[str] = shard(lowercase_ ) UpperCAmelCase_ : List[Any] = pipeline(lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase_ : int = images[2, 0, 256, 10:17, 1] # With memory efficient attention UpperCAmelCase_ , UpperCAmelCase_ : int = FlaxStableDiffusionPipeline.from_pretrained( "CompVis/stable-diffusion-v1-4" , revision="bf16" , dtype=jnp.bfloataa , safety_checker=lowercase_ , use_memory_efficient_attention=lowercase_ , ) UpperCAmelCase_ : str = replicate(lowercase_ ) UpperCAmelCase_ : str = pipeline.prepare_inputs(lowercase_ ) UpperCAmelCase_ : Optional[int] = shard(lowercase_ ) UpperCAmelCase_ : str = pipeline(lowercase_ , lowercase_ , lowercase_ , jit=lowercase_ ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) UpperCAmelCase_ : Optional[int] = images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1E-2
23
0
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a = logging.get_logger(__name__) _a = '▁' _a = {'vocab_file': 'sentencepiece.bpe.model'} _a = { 'vocab_file': { 'facebook/xglm-564M': 'https://huggingface.co/facebook/xglm-564M/resolve/main/sentencepiece.bpe.model', } } _a = { 'facebook/xglm-564M': 2_048, } class A_ (lowerCamelCase_ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE__ : Tuple = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE__ : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE__ : List[str] = ["""input_ids""", """attention_mask"""] def __init__( self , lowercase_ , lowercase_="<s>" , lowercase_="</s>" , lowercase_="</s>" , lowercase_="<s>" , lowercase_="<unk>" , lowercase_="<pad>" , lowercase_ = None , **lowercase_ , ): """simple docstring""" UpperCAmelCase_ : str = {} if sp_model_kwargs is None else sp_model_kwargs # Compatibility with the original tokenizer UpperCAmelCase_ : Any = 7 UpperCAmelCase_ : List[str] = [F"""<madeupword{i}>""" for i in range(self.num_madeup_words )] UpperCAmelCase_ : Any = kwargs.get("additional_special_tokens" , [] ) kwargs["additional_special_tokens"] += [ word for word in madeup_words if word not in kwargs["additional_special_tokens"] ] super().__init__( bos_token=_UpperCAmelCase , eos_token=_UpperCAmelCase , unk_token=_UpperCAmelCase , sep_token=_UpperCAmelCase , cls_token=_UpperCAmelCase , pad_token=_UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **_UpperCAmelCase , ) UpperCAmelCase_ : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(_UpperCAmelCase ) ) UpperCAmelCase_ : Tuple = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab UpperCAmelCase_ : int = 1 # Mimic fairseq token-to-id alignment for the first 4 token UpperCAmelCase_ : Any = {"<s>": 0, "<pad>": 1, "</s>": 2, "<unk>": 3} UpperCAmelCase_ : Tuple = len(self.sp_model ) UpperCAmelCase_ : Dict = {F"""<madeupword{i}>""": sp_size + i + self.fairseq_offset for i in range(self.num_madeup_words )} self.fairseq_tokens_to_ids.update(_UpperCAmelCase ) UpperCAmelCase_ : int = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.__dict__.copy() UpperCAmelCase_ : int = None UpperCAmelCase_ : Optional[int] = self.sp_model.serialized_model_proto() return state def __setstate__( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Any = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): UpperCAmelCase_ : Tuple = {} UpperCAmelCase_ : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if token_ids_a is None: return [self.sep_token_id] + token_ids_a UpperCAmelCase_ : Union[str, Any] = [self.sep_token_id] return sep + token_ids_a + sep + sep + token_ids_a def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None , lowercase_ = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_UpperCAmelCase , token_ids_a=_UpperCAmelCase , already_has_special_tokens=_UpperCAmelCase ) if token_ids_a is None: return [1] + ([0] * len(_UpperCAmelCase )) return [1] + ([0] * len(_UpperCAmelCase )) + [1, 1] + ([0] * len(_UpperCAmelCase )) def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" UpperCAmelCase_ : int = [self.sep_token_id] if token_ids_a is None: return len(sep + token_ids_a ) * [0] return len(sep + token_ids_a + sep + sep + token_ids_a ) * [0] @property def UpperCamelCase__ ( self ): """simple docstring""" return len(self.sp_model ) + self.fairseq_offset + self.num_madeup_words def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Dict = {self.convert_ids_to_tokens(_UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" return self.sp_model.encode(_UpperCAmelCase , out_type=_UpperCAmelCase ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCAmelCase_ : Dict = self.sp_model.PieceToId(_UpperCAmelCase ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : str = "".join(_UpperCAmelCase ).replace(_UpperCAmelCase , " " ).strip() return out_string def UpperCamelCase__ ( self , lowercase_ , lowercase_ = None ): """simple docstring""" if not os.path.isdir(_UpperCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase_ : Tuple = os.path.join( _UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , _UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(_UpperCAmelCase , "wb" ) as fi: UpperCAmelCase_ : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(_UpperCAmelCase ) return (out_vocab_file,)
366
"""simple docstring""" from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean _a = 0 _a = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] _a = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right _a = tuple[int, int] class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , lowercase_ , ): """simple docstring""" UpperCAmelCase_ : int = pos_x UpperCAmelCase_ : List[Any] = pos_y UpperCAmelCase_ : Union[str, Any] = (pos_y, pos_x) UpperCAmelCase_ : Any = goal_x UpperCAmelCase_ : Dict = goal_y UpperCAmelCase_ : Any = g_cost UpperCAmelCase_ : List[str] = parent UpperCAmelCase_ : int = self.calculate_heuristic() UpperCAmelCase_ : Any = self.g_cost + self.h_cost def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : Tuple = self.pos_x - self.goal_x UpperCAmelCase_ : Union[str, Any] = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(lowercase_ ) + abs(lowercase_ ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self , lowercase_ ): """simple docstring""" return self.f_cost < other.f_cost class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Tuple = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , lowercase_ ) UpperCAmelCase_ : List[Any] = Node(goal[1] , goal[0] , goal[1] , goal[0] , 9_9999 , lowercase_ ) UpperCAmelCase_ : str = [self.start] UpperCAmelCase_ : list[Node] = [] UpperCAmelCase_ : int = False def UpperCamelCase__ ( self ): """simple docstring""" while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() UpperCAmelCase_ : List[str] = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(lowercase_ ) self.closed_nodes.append(lowercase_ ) UpperCAmelCase_ : str = self.get_successors(lowercase_ ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(lowercase_ ) else: # retrieve the best current path UpperCAmelCase_ : Union[str, Any] = self.open_nodes.pop(self.open_nodes.index(lowercase_ ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(lowercase_ ) else: self.open_nodes.append(lowercase_ ) return [self.start.pos] def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Any = [] for action in delta: UpperCAmelCase_ : str = parent.pos_x + action[1] UpperCAmelCase_ : int = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(lowercase_ ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( lowercase_ , lowercase_ , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , lowercase_ , ) ) return successors def UpperCamelCase__ ( self , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Optional[int] = node UpperCAmelCase_ : int = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) UpperCAmelCase_ : Optional[int] = current_node.parent path.reverse() return path class A_ : '''simple docstring''' def __init__( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : Dict = AStar(lowercase_ , lowercase_ ) UpperCAmelCase_ : Optional[Any] = AStar(lowercase_ , lowercase_ ) UpperCAmelCase_ : Tuple = False def UpperCamelCase__ ( self ): """simple docstring""" while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() UpperCAmelCase_ : List[str] = self.fwd_astar.open_nodes.pop(0 ) UpperCAmelCase_ : List[Any] = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( lowercase_ , lowercase_ ) self.fwd_astar.closed_nodes.append(lowercase_ ) self.bwd_astar.closed_nodes.append(lowercase_ ) UpperCAmelCase_ : Tuple = current_bwd_node UpperCAmelCase_ : str = current_fwd_node UpperCAmelCase_ : Dict = { self.fwd_astar: self.fwd_astar.get_successors(lowercase_ ), self.bwd_astar: self.bwd_astar.get_successors(lowercase_ ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(lowercase_ ) else: # retrieve the best current path UpperCAmelCase_ : List[Any] = astar.open_nodes.pop( astar.open_nodes.index(lowercase_ ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(lowercase_ ) else: astar.open_nodes.append(lowercase_ ) return [self.fwd_astar.start.pos] def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" UpperCAmelCase_ : List[Any] = self.fwd_astar.retrace_path(lowercase_ ) UpperCAmelCase_ : int = self.bwd_astar.retrace_path(lowercase_ ) bwd_path.pop() bwd_path.reverse() UpperCAmelCase_ : Any = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] _a = (0, 0) _a = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) _a = time.time() _a = AStar(init, goal) _a = a_star.search() _a = time.time() - start_time print(f"""AStar execution time = {end_time:f} seconds""") _a = time.time() _a = BidirectionalAStar(init, goal) _a = time.time() - bd_start_time print(f"""BidirectionalAStar execution time = {bd_end_time:f} seconds""")
23
0