code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import warnings from ...utils import logging from .image_processing_perceiver import PerceiverImageProcessor UpperCAmelCase__ = logging.get_logger(__name__) class lowercase_ ( snake_case_ ): '''simple docstring''' def __init__( self : List[Any] , *__UpperCAmelCase : Dict , **__UpperCAmelCase : Tuple ) ->Optional[Any]: """simple docstring""" warnings.warn( '''The class PerceiverFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use PerceiverImageProcessor instead.''' , _a , ) super().__init__(*_a , **_a )
117
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer import diffusers from diffusers import ( AutoencoderKL, EulerDiscreteScheduler, StableDiffusionLatentUpscalePipeline, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.schedulers import KarrasDiffusionSchedulers from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> List[Any]: snake_case__ = [tensor.shape for tensor in tensor_list] return all(shape == shapes[0] for shape in shapes[1:] ) class __magic_name__ (snake_case_ ,snake_case_ ,snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : Dict = StableDiffusionLatentUpscalePipeline __lowercase : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { 'height', 'width', 'cross_attention_kwargs', 'negative_prompt_embeds', 'prompt_embeds', } __lowercase : List[Any] = PipelineTesterMixin.required_optional_params - {'num_images_per_prompt'} __lowercase : Any = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS __lowercase : int = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __lowercase : List[Any] = frozenset([] ) __lowercase : Any = True @property def SCREAMING_SNAKE_CASE__ ( self:List[str] ): snake_case__ = 1 snake_case__ = 4 snake_case__ = (16, 16) snake_case__ = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(_a ) return image def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): torch.manual_seed(0 ) snake_case__ = UNetaDConditionModel( act_fn='''gelu''' , attention_head_dim=8 , norm_num_groups=_a , block_out_channels=[32, 32, 64, 64] , time_cond_proj_dim=1_60 , conv_in_kernel=1 , conv_out_kernel=1 , cross_attention_dim=32 , down_block_types=( '''KDownBlock2D''', '''KCrossAttnDownBlock2D''', '''KCrossAttnDownBlock2D''', '''KCrossAttnDownBlock2D''', ) , in_channels=8 , mid_block_type=_a , only_cross_attention=_a , out_channels=5 , resnet_time_scale_shift='''scale_shift''' , time_embedding_type='''fourier''' , timestep_post_act='''gelu''' , up_block_types=('''KCrossAttnUpBlock2D''', '''KCrossAttnUpBlock2D''', '''KCrossAttnUpBlock2D''', '''KUpBlock2D''') , ) snake_case__ = AutoencoderKL( block_out_channels=[32, 32, 64, 64] , in_channels=3 , out_channels=3 , down_block_types=[ '''DownEncoderBlock2D''', '''DownEncoderBlock2D''', '''DownEncoderBlock2D''', '''DownEncoderBlock2D''', ] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D''', '''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) snake_case__ = EulerDiscreteScheduler(prediction_type='''sample''' ) snake_case__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act='''quick_gelu''' , projection_dim=5_12 , ) snake_case__ = CLIPTextModel(_a ) snake_case__ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) snake_case__ = { '''unet''': model.eval(), '''vae''': vae.eval(), '''scheduler''': scheduler, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def SCREAMING_SNAKE_CASE__ ( self:List[Any] , _a:Optional[Any] , _a:List[str]=0 ): if str(_a ).startswith('''mps''' ): snake_case__ = torch.manual_seed(_a ) else: snake_case__ = torch.Generator(device=_a ).manual_seed(_a ) snake_case__ = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': self.dummy_image.cpu(), '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = '''cpu''' snake_case__ = self.get_dummy_components() snake_case__ = self.pipeline_class(**_a ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) snake_case__ = self.get_dummy_inputs(_a ) snake_case__ = pipe(**_a ).images snake_case__ = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 2_56, 2_56, 3) ) snake_case__ = np.array( [0.47222412, 0.41921633, 0.44717434, 0.46874192, 0.42588258, 0.46150726, 0.4677534, 0.45583832, 0.48579055] ) snake_case__ = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_a , 1e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): super().test_attention_slicing_forward_pass(expected_max_diff=7e-3 ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): super().test_cpu_offload_forward_pass(expected_max_diff=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:str ): super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Any ): super().test_inference_batch_single_identical(expected_max_diff=7e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): super().test_pt_np_pil_outputs_equivalent(expected_max_diff=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): super().test_save_load_local(expected_max_difference=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:str ): super().test_save_load_optional_components(expected_max_difference=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = [ '''DDIMScheduler''', '''DDPMScheduler''', '''PNDMScheduler''', '''HeunDiscreteScheduler''', '''EulerAncestralDiscreteScheduler''', '''KDPM2DiscreteScheduler''', '''KDPM2AncestralDiscreteScheduler''', '''DPMSolverSDEScheduler''', ] snake_case__ = self.get_dummy_components() snake_case__ = self.pipeline_class(**_a ) # make sure that PNDM does not need warm-up pipe.scheduler.register_to_config(skip_prk_steps=_a ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) snake_case__ = self.get_dummy_inputs(_a ) snake_case__ = 2 snake_case__ = [] for scheduler_enum in KarrasDiffusionSchedulers: if scheduler_enum.name in skip_schedulers: # no sigma schedulers are not supported # no schedulers continue snake_case__ = getattr(_a , scheduler_enum.name ) snake_case__ = scheduler_cls.from_config(pipe.scheduler.config ) snake_case__ = pipe(**_a )[0] outputs.append(_a ) assert check_same_shape(_a ) @require_torch_gpu @slow class __magic_name__ (unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = torch.manual_seed(33 ) snake_case__ = StableDiffusionPipeline.from_pretrained('''CompVis/stable-diffusion-v1-4''' , torch_dtype=torch.floataa ) pipe.to('''cuda''' ) snake_case__ = StableDiffusionLatentUpscalePipeline.from_pretrained( '''stabilityai/sd-x2-latent-upscaler''' , torch_dtype=torch.floataa ) upscaler.to('''cuda''' ) snake_case__ = '''a photo of an astronaut high resolution, unreal engine, ultra realistic''' snake_case__ = pipe(_a , generator=_a , output_type='''latent''' ).images snake_case__ = upscaler( prompt=_a , image=_a , num_inference_steps=20 , guidance_scale=0 , generator=_a , output_type='''np''' , ).images[0] snake_case__ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/astronaut_1024.npy''' ) assert np.abs((expected_image - image).mean() ) < 5e-2 def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): snake_case__ = torch.manual_seed(33 ) snake_case__ = StableDiffusionLatentUpscalePipeline.from_pretrained( '''stabilityai/sd-x2-latent-upscaler''' , torch_dtype=torch.floataa ) upscaler.to('''cuda''' ) snake_case__ = '''the temple of fire by Ross Tran and Gerardo Dottori, oil on canvas''' snake_case__ = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_512.png''' ) snake_case__ = upscaler( prompt=_a , image=_a , num_inference_steps=20 , guidance_scale=0 , generator=_a , output_type='''np''' , ).images[0] snake_case__ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_1024.npy''' ) assert np.abs((expected_image - image).max() ) < 5e-2
33
0
import gc import unittest import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTextModelWithProjection, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DDPMScheduler, PriorTransformer, StableUnCLIPPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.stable_unclip_image_normalizer import StableUnCLIPImageNormalizer from diffusers.utils.testing_utils import enable_full_determinism, load_numpy, require_torch_gpu, slow, torch_device from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_IMAGE_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, assert_mean_pixel_difference, ) enable_full_determinism() class lowerCAmelCase__ ( snake_case_ , snake_case_ , snake_case_ , unittest.TestCase ): '''simple docstring''' lowercase_ = StableUnCLIPPipeline lowercase_ = TEXT_TO_IMAGE_PARAMS lowercase_ = TEXT_TO_IMAGE_BATCH_PARAMS lowercase_ = TEXT_TO_IMAGE_IMAGE_PARAMS lowercase_ = TEXT_TO_IMAGE_IMAGE_PARAMS # TODO(will) Expected attn_bias.stride(1) == 0 to be true, but got false lowercase_ = False def __UpperCamelCase ( self ): '''simple docstring''' __A =3_2 __A =embedder_hidden_size # prior components torch.manual_seed(0 ) __A =CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) torch.manual_seed(0 ) __A =CLIPTextModelWithProjection( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=_a , projection_dim=_a , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) ) torch.manual_seed(0 ) __A =PriorTransformer( num_attention_heads=2 , attention_head_dim=1_2 , embedding_dim=_a , num_layers=1 , ) torch.manual_seed(0 ) __A =DDPMScheduler( variance_type='''fixed_small_log''' , prediction_type='''sample''' , num_train_timesteps=1_0_0_0 , clip_sample=_a , clip_sample_range=5.0 , beta_schedule='''squaredcos_cap_v2''' , ) # regular denoising components torch.manual_seed(0 ) __A =StableUnCLIPImageNormalizer(embedding_dim=_a ) __A =DDPMScheduler(beta_schedule='''squaredcos_cap_v2''' ) torch.manual_seed(0 ) __A =CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) torch.manual_seed(0 ) __A =CLIPTextModel( CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=_a , projection_dim=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) ) torch.manual_seed(0 ) __A =UNetaDConditionModel( sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') , up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') , block_out_channels=(3_2, 6_4) , attention_head_dim=(2, 4) , class_embed_type='''projection''' , projection_class_embeddings_input_dim=embedder_projection_dim * 2 , cross_attention_dim=_a , layers_per_block=1 , upcast_attention=_a , use_linear_projection=_a , ) torch.manual_seed(0 ) __A =DDIMScheduler( beta_schedule='''scaled_linear''' , beta_start=0.0_0085 , beta_end=0.012 , prediction_type='''v_prediction''' , set_alpha_to_one=_a , steps_offset=1 , ) torch.manual_seed(0 ) __A =AutoencoderKL() __A ={ # prior components '''prior_tokenizer''': prior_tokenizer, '''prior_text_encoder''': prior_text_encoder, '''prior''': prior, '''prior_scheduler''': prior_scheduler, # image noising components '''image_normalizer''': image_normalizer, '''image_noising_scheduler''': image_noising_scheduler, # regular denoising components '''tokenizer''': tokenizer, '''text_encoder''': text_encoder, '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, } return components def __UpperCamelCase ( self , lowercase__ , lowercase__=0 ): '''simple docstring''' if str(_a ).startswith('''mps''' ): __A =torch.manual_seed(_a ) else: __A =torch.Generator(device=_a ).manual_seed(_a ) __A ={ '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''prior_num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def __UpperCamelCase ( self ): '''simple docstring''' __A =torch_device == '''cpu''' self._test_attention_slicing_forward_pass(test_max_difference=_a ) def __UpperCamelCase ( self ): '''simple docstring''' __A =torch_device in ['''cpu''', '''mps'''] self._test_inference_batch_single_identical(test_max_difference=_a ) @slow @require_torch_gpu class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def __UpperCamelCase ( self ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def __UpperCamelCase ( self ): '''simple docstring''' __A =load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/stable_unclip/stable_unclip_2_1_l_anime_turtle_fp16.npy''' ) __A =StableUnCLIPPipeline.from_pretrained('''fusing/stable-unclip-2-1-l''' , torch_dtype=torch.floataa ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) # stable unclip will oom when integration tests are run on a V100, # so turn on memory savings pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __A =torch.Generator(device='''cpu''' ).manual_seed(0 ) __A =pipe('''anime turle''' , generator=_a , output_type='''np''' ) __A =output.images[0] assert image.shape == (7_6_8, 7_6_8, 3) assert_mean_pixel_difference(_a , _a ) def __UpperCamelCase ( self ): '''simple docstring''' torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() __A =StableUnCLIPPipeline.from_pretrained('''fusing/stable-unclip-2-1-l''' , torch_dtype=torch.floataa ) __A =pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) pipe.enable_attention_slicing() pipe.enable_sequential_cpu_offload() __A =pipe( '''anime turtle''' , prior_num_inference_steps=2 , num_inference_steps=2 , output_type='''np''' , ) __A =torch.cuda.max_memory_allocated() # make sure that less than 7 GB is allocated assert mem_bytes < 7 * 1_0**9
184
import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class __magic_name__ (unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = '''ZinengTang/tvlt-base''' snake_case__ = tempfile.mkdtemp() def SCREAMING_SNAKE_CASE__ ( self:Dict , **_a:List[Any] ): return TvltImageProcessor.from_pretrained(self.checkpoint , **_a ) def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] , **_a:Tuple ): return TvltFeatureExtractor.from_pretrained(self.checkpoint , **_a ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): shutil.rmtree(self.tmpdirname ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) processor.save_pretrained(self.tmpdirname ) snake_case__ = TvltProcessor.from_pretrained(self.tmpdirname ) self.assertIsInstance(processor.feature_extractor , _a ) self.assertIsInstance(processor.image_processor , _a ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) snake_case__ = np.ones([1_20_00] ) snake_case__ = feature_extractor(_a , return_tensors='''np''' ) snake_case__ = processor(audio=_a , return_tensors='''np''' ) for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) snake_case__ = np.ones([3, 2_24, 2_24] ) snake_case__ = image_processor(_a , return_tensors='''np''' ) snake_case__ = processor(images=_a , return_tensors='''np''' ) for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) snake_case__ = np.ones([1_20_00] ) snake_case__ = np.ones([3, 2_24, 2_24] ) snake_case__ = processor(audio=_a , images=_a ) self.assertListEqual(list(inputs.keys() ) , ['''audio_values''', '''audio_mask''', '''pixel_values''', '''pixel_mask'''] ) # test if it raises when no input is passed with pytest.raises(_a ): processor() def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg='''`processor` and `image_processor`+`feature_extractor` model input names do not match''' , )
33
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE_ : Tuple = { """configuration_roberta""": ["""ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """RobertaConfig""", """RobertaOnnxConfig"""], """tokenization_roberta""": ["""RobertaTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ : Tuple = ["""RobertaTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ : Optional[int] = [ """ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """RobertaForCausalLM""", """RobertaForMaskedLM""", """RobertaForMultipleChoice""", """RobertaForQuestionAnswering""", """RobertaForSequenceClassification""", """RobertaForTokenClassification""", """RobertaModel""", """RobertaPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ : List[str] = [ """TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFRobertaForCausalLM""", """TFRobertaForMaskedLM""", """TFRobertaForMultipleChoice""", """TFRobertaForQuestionAnswering""", """TFRobertaForSequenceClassification""", """TFRobertaForTokenClassification""", """TFRobertaMainLayer""", """TFRobertaModel""", """TFRobertaPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ : str = [ """FlaxRobertaForCausalLM""", """FlaxRobertaForMaskedLM""", """FlaxRobertaForMultipleChoice""", """FlaxRobertaForQuestionAnswering""", """FlaxRobertaForSequenceClassification""", """FlaxRobertaForTokenClassification""", """FlaxRobertaModel""", """FlaxRobertaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_roberta import ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaConfig, RobertaOnnxConfig from .tokenization_roberta import RobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roberta_fast import RobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta import ( ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaForCausalLM, RobertaForMaskedLM, RobertaForMultipleChoice, RobertaForQuestionAnswering, RobertaForSequenceClassification, RobertaForTokenClassification, RobertaModel, RobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta import ( TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForMultipleChoice, TFRobertaForQuestionAnswering, TFRobertaForSequenceClassification, TFRobertaForTokenClassification, TFRobertaMainLayer, TFRobertaModel, TFRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, FlaxRobertaPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE_ : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
375
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase__ : List[Any] = logging.get_logger(__name__) lowerCamelCase__ : Optional[int] = { """facebook/data2vec-vision-base-ft""": ( """https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json""" ), } class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : Optional[int] = 'data2vec-vision' def __init__( self:int , _a:Tuple=7_68 , _a:int=12 , _a:Any=12 , _a:Optional[int]=30_72 , _a:Optional[int]="gelu" , _a:Any=0.0 , _a:Any=0.0 , _a:List[str]=0.02 , _a:Dict=1e-12 , _a:Tuple=2_24 , _a:Any=16 , _a:str=3 , _a:str=False , _a:Union[str, Any]=False , _a:Optional[int]=False , _a:Any=False , _a:Dict=0.1 , _a:Dict=0.1 , _a:str=True , _a:str=[3, 5, 7, 11] , _a:List[str]=[1, 2, 3, 6] , _a:List[str]=True , _a:Any=0.4 , _a:str=2_56 , _a:Union[str, Any]=1 , _a:int=False , _a:Optional[int]=2_55 , **_a:Dict , ): super().__init__(**_a ) snake_case__ = hidden_size snake_case__ = num_hidden_layers snake_case__ = num_attention_heads snake_case__ = intermediate_size snake_case__ = hidden_act snake_case__ = hidden_dropout_prob snake_case__ = attention_probs_dropout_prob snake_case__ = initializer_range snake_case__ = layer_norm_eps snake_case__ = image_size snake_case__ = patch_size snake_case__ = num_channels snake_case__ = use_mask_token snake_case__ = use_absolute_position_embeddings snake_case__ = use_relative_position_bias snake_case__ = use_shared_relative_position_bias snake_case__ = layer_scale_init_value snake_case__ = drop_path_rate snake_case__ = use_mean_pooling # decode head attributes (semantic segmentation) snake_case__ = out_indices snake_case__ = pool_scales # auxiliary head attributes (semantic segmentation) snake_case__ = use_auxiliary_head snake_case__ = auxiliary_loss_weight snake_case__ = auxiliary_channels snake_case__ = auxiliary_num_convs snake_case__ = auxiliary_concat_input snake_case__ = semantic_loss_ignore_index class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : Any = version.parse('1.11' ) @property def SCREAMING_SNAKE_CASE__ ( self:List[str] ): return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def SCREAMING_SNAKE_CASE__ ( self:Tuple ): return 1e-4
33
0
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowercase : Tuple = logging.get_logger(__name__) lowercase : str = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } lowercase : List[str] = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } lowercase : str = {"""facebook/blenderbot_small-90M""": 5_12} def lowerCAmelCase__ ( _a : Tuple ): snake_case_ : Any = set() snake_case_ : Tuple = word[0] for char in word[1:]: pairs.add((prev_char, char) ) snake_case_ : Optional[Any] = char snake_case_ : Dict = set(__lowerCAmelCase ) return pairs class UpperCAmelCase_ ( snake_case_ ): '''simple docstring''' A : Tuple = VOCAB_FILES_NAMES A : Optional[int] = PRETRAINED_VOCAB_FILES_MAP A : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A : Union[str, Any] = ['input_ids', 'attention_mask'] def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="__start__" , _SCREAMING_SNAKE_CASE="__end__" , _SCREAMING_SNAKE_CASE="__unk__" , _SCREAMING_SNAKE_CASE="__null__" , **_SCREAMING_SNAKE_CASE , ) -> str: super().__init__(unk_token=_a , bos_token=_a , eos_token=_a , pad_token=_a , **_a ) with open(_a , encoding="utf-8" ) as vocab_handle: snake_case_ : Any = json.load(_a ) snake_case_ : Tuple = {v: k for k, v in self.encoder.items()} with open(_a , encoding="utf-8" ) as merges_handle: snake_case_ : Optional[Any] = merges_handle.read().split("\n" )[1:-1] snake_case_ : int = [tuple(merge.split() ) for merge in merges] snake_case_ : Tuple = dict(zip(_a , range(len(_a ) ) ) ) snake_case_ : Any = {} @property def _lowerCAmelCase ( self ) -> Optional[int]: return len(self.encoder ) def _lowerCAmelCase ( self ) -> Dict: return dict(self.encoder , **self.added_tokens_encoder ) def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE ) -> Optional[Any]: if token in self.cache: return self.cache[token] snake_case_ : List[Any] = re.sub("([.,!?()])" , r" \1" , _a ) snake_case_ : Optional[int] = re.sub("(\')" , r" \1 " , _a ) snake_case_ : int = re.sub(r"\s{2,}" , " " , _a ) if "\n" in token: snake_case_ : Optional[Any] = token.replace("\n" , " __newln__" ) snake_case_ : Any = token.split(" " ) snake_case_ : Optional[Any] = [] for token in tokens: if not len(_a ): continue snake_case_ : List[Any] = token.lower() snake_case_ : Dict = tuple(_a ) snake_case_ : int = tuple(list(word[:-1] ) + [word[-1] + "</w>"] ) snake_case_ : int = get_pairs(_a ) if not pairs: words.append(_a ) continue while True: snake_case_ : str = min(_a , key=lambda _SCREAMING_SNAKE_CASE : self.bpe_ranks.get(_a , float("inf" ) ) ) if bigram not in self.bpe_ranks: break snake_case_ , snake_case_ : Any = bigram snake_case_ : Optional[Any] = [] snake_case_ : Dict = 0 while i < len(_a ): try: snake_case_ : Union[str, Any] = word.index(_a , _a ) new_word.extend(word[i:j] ) snake_case_ : Any = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(_a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 snake_case_ : Optional[int] = tuple(_a ) snake_case_ : str = new_word if len(_a ) == 1: break else: snake_case_ : Any = get_pairs(_a ) snake_case_ : List[str] = "@@ ".join(_a ) snake_case_ : Tuple = word[:-4] snake_case_ : Dict = word words.append(_a ) return " ".join(_a ) def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE ) -> Tuple: snake_case_ : List[Any] = [] snake_case_ : List[str] = re.findall(r"\S+\n?" , _a ) for token in words: split_tokens.extend(list(self.bpe(_a ).split(" " ) ) ) return split_tokens def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE ) -> Dict: snake_case_ : List[Any] = token.lower() return self.encoder.get(_a , self.encoder.get(self.unk_token ) ) def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE ) -> Optional[Any]: return self.decoder.get(_a , self.unk_token ) def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE ) -> Any: snake_case_ : Dict = " ".join(_a ).replace("@@ " , "" ).strip() return out_string def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = None ) -> List[Any]: if not os.path.isdir(_a ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return snake_case_ : Tuple = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) snake_case_ : Optional[int] = os.path.join( _a , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(_a , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=_a , ensure_ascii=_a ) + "\n" ) snake_case_ : List[Any] = 0 with open(_a , "w" , encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _SCREAMING_SNAKE_CASE : kv[1] ): if index != token_index: logger.warning( f'''Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.''' " Please check that the tokenizer is not corrupted!" ) snake_case_ : Union[str, Any] = token_index writer.write(" ".join(_a ) + "\n" ) index += 1 return vocab_file, merge_file
568
import os import sys lowerCamelCase__ : Optional[int] = os.path.join(os.path.dirname(__file__), """src""") sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) lowerCamelCase__ : Optional[int] = [ """torch""", """numpy""", """tokenizers""", """filelock""", """requests""", """tqdm""", """regex""", """sentencepiece""", """sacremoses""", """importlib_metadata""", """huggingface_hub""", ] @add_start_docstrings(AutoConfig.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Any: return AutoConfig.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoTokenizer.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> List[str]: return AutoTokenizer.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModel.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Tuple: return AutoModel.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Union[str, Any]: return AutoModelForCausalLM.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> List[Any]: return AutoModelForMaskedLM.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> List[str]: return AutoModelForSequenceClassification.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Union[str, Any]: return AutoModelForQuestionAnswering.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase )
33
0
# Imports import numpy as np class lowerCamelCase_ : def __init__( self , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=None ) -> Any: """simple docstring""" self.set_matricies(red=_a , green=_a , blue=_a , red_edge=_a , nir=_a ) def lowercase ( self , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=None ) -> Any: """simple docstring""" if red is not None: _UpperCamelCase = red if green is not None: _UpperCamelCase = green if blue is not None: _UpperCamelCase = blue if red_edge is not None: _UpperCamelCase = red_edge if nir is not None: _UpperCamelCase = nir return True def lowercase ( self , lowerCamelCase_="" , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=None ) -> int: """simple docstring""" self.set_matricies(red=_a , green=_a , blue=_a , red_edge=_a , nir=_a ) _UpperCamelCase = { "ARVI2": self.arvaa, "CCCI": self.ccci, "CVI": self.cvi, "GLI": self.gli, "NDVI": self.ndvi, "BNDVI": self.bndvi, "redEdgeNDVI": self.red_edge_ndvi, "GNDVI": self.gndvi, "GBNDVI": self.gbndvi, "GRNDVI": self.grndvi, "RBNDVI": self.rbndvi, "PNDVI": self.pndvi, "ATSAVI": self.atsavi, "BWDRVI": self.bwdrvi, "CIgreen": self.ci_green, "CIrededge": self.ci_rededge, "CI": self.ci, "CTVI": self.ctvi, "GDVI": self.gdvi, "EVI": self.evi, "GEMI": self.gemi, "GOSAVI": self.gosavi, "GSAVI": self.gsavi, "Hue": self.hue, "IVI": self.ivi, "IPVI": self.ipvi, "I": self.i, "RVI": self.rvi, "MRVI": self.mrvi, "MSAVI": self.m_savi, "NormG": self.norm_g, "NormNIR": self.norm_nir, "NormR": self.norm_r, "NGRDI": self.ngrdi, "RI": self.ri, "S": self.s, "IF": self._if, "DVI": self.dvi, "TVI": self.tvi, "NDRE": self.ndre, } try: return funcs[index]() except KeyError: print("Index not in the list!" ) return False def lowercase ( self ) -> Dict: """simple docstring""" return -0.18 + (1.17 * ((self.nir - self.red) / (self.nir + self.red))) def lowercase ( self ) -> Dict: """simple docstring""" return ((self.nir - self.redEdge) / (self.nir + self.redEdge)) / ( (self.nir - self.red) / (self.nir + self.red) ) def lowercase ( self ) -> List[Any]: """simple docstring""" return self.nir * (self.red / (self.green**2)) def lowercase ( self ) -> Dict: """simple docstring""" return (2 * self.green - self.red - self.blue) / ( 2 * self.green + self.red + self.blue ) def lowercase ( self ) -> int: """simple docstring""" return (self.nir - self.red) / (self.nir + self.red) def lowercase ( self ) -> Tuple: """simple docstring""" return (self.nir - self.blue) / (self.nir + self.blue) def lowercase ( self ) -> Tuple: """simple docstring""" return (self.redEdge - self.red) / (self.redEdge + self.red) def lowercase ( self ) -> Any: """simple docstring""" return (self.nir - self.green) / (self.nir + self.green) def lowercase ( self ) -> int: """simple docstring""" return (self.nir - (self.green + self.blue)) / ( self.nir + (self.green + self.blue) ) def lowercase ( self ) -> List[str]: """simple docstring""" return (self.nir - (self.green + self.red)) / ( self.nir + (self.green + self.red) ) def lowercase ( self ) -> Optional[int]: """simple docstring""" return (self.nir - (self.blue + self.red)) / (self.nir + (self.blue + self.red)) def lowercase ( self ) -> Tuple: """simple docstring""" return (self.nir - (self.green + self.red + self.blue)) / ( self.nir + (self.green + self.red + self.blue) ) def lowercase ( self , lowerCamelCase_=0.08 , lowerCamelCase_=1.22 , lowerCamelCase_=0.03 ) -> Optional[Any]: """simple docstring""" return a * ( (self.nir - a * self.red - b) / (a * self.nir + self.red - a * b + x * (1 + a**2)) ) def lowercase ( self ) -> str: """simple docstring""" return (0.1 * self.nir - self.blue) / (0.1 * self.nir + self.blue) def lowercase ( self ) -> str: """simple docstring""" return (self.nir / self.green) - 1 def lowercase ( self ) -> List[Any]: """simple docstring""" return (self.nir / self.redEdge) - 1 def lowercase ( self ) -> Union[str, Any]: """simple docstring""" return (self.red - self.blue) / self.red def lowercase ( self ) -> List[str]: """simple docstring""" _UpperCamelCase = self.ndvi() return ((ndvi + 0.5) / (abs(ndvi + 0.5 ))) * (abs(ndvi + 0.5 ) ** (1 / 2)) def lowercase ( self ) -> str: """simple docstring""" return self.nir - self.green def lowercase ( self ) -> Dict: """simple docstring""" return 2.5 * ( (self.nir - self.red) / (self.nir + 6 * self.red - 7.5 * self.blue + 1) ) def lowercase ( self ) -> str: """simple docstring""" _UpperCamelCase = (2 * (self.nir**2 - self.red**2) + 1.5 * self.nir + 0.5 * self.red) / ( self.nir + self.red + 0.5 ) return n * (1 - 0.25 * n) - (self.red - 0.1_25) / (1 - self.red) def lowercase ( self , lowerCamelCase_=0.16 ) -> int: """simple docstring""" return (self.nir - self.green) / (self.nir + self.green + y) def lowercase ( self , lowerCamelCase_=0.5 ) -> str: """simple docstring""" return ((self.nir - self.green) / (self.nir + self.green + n)) * (1 + n) def lowercase ( self ) -> List[str]: """simple docstring""" return np.arctan( ((2 * self.red - self.green - self.blue) / 30.5) * (self.green - self.blue) ) def lowercase ( self , lowerCamelCase_=None , lowerCamelCase_=None ) -> Optional[int]: """simple docstring""" return (self.nir - b) / (a * self.red) def lowercase ( self ) -> Union[str, Any]: """simple docstring""" return (self.nir / ((self.nir + self.red) / 2)) * (self.ndvi() + 1) def lowercase ( self ) -> Tuple: """simple docstring""" return (self.red + self.green + self.blue) / 30.5 def lowercase ( self ) -> Union[str, Any]: """simple docstring""" return self.nir / self.red def lowercase ( self ) -> Optional[int]: """simple docstring""" return (self.rvi() - 1) / (self.rvi() + 1) def lowercase ( self ) -> Optional[Any]: """simple docstring""" return ( (2 * self.nir + 1) - ((2 * self.nir + 1) ** 2 - 8 * (self.nir - self.red)) ** (1 / 2) ) / 2 def lowercase ( self ) -> str: """simple docstring""" return self.green / (self.nir + self.red + self.green) def lowercase ( self ) -> str: """simple docstring""" return self.nir / (self.nir + self.red + self.green) def lowercase ( self ) -> str: """simple docstring""" return self.red / (self.nir + self.red + self.green) def lowercase ( self ) -> Any: """simple docstring""" return (self.green - self.red) / (self.green + self.red) def lowercase ( self ) -> Dict: """simple docstring""" return (self.red - self.green) / (self.red + self.green) def lowercase ( self ) -> List[Any]: """simple docstring""" _UpperCamelCase = np.max([np.max(self.red ), np.max(self.green ), np.max(self.blue )] ) _UpperCamelCase = np.min([np.min(self.red ), np.min(self.green ), np.min(self.blue )] ) return (max_value - min_value) / max_value def lowercase ( self ) -> Dict: """simple docstring""" return (2 * self.red - self.green - self.blue) / (self.green - self.blue) def lowercase ( self ) -> Dict: """simple docstring""" return self.nir / self.red def lowercase ( self ) -> Union[str, Any]: """simple docstring""" return (self.ndvi() + 0.5) ** (1 / 2) def lowercase ( self ) -> Dict: """simple docstring""" return (self.nir - self.redEdge) / (self.nir + self.redEdge)
147
import torch from diffusers import CMStochasticIterativeScheduler from .test_schedulers import SchedulerCommonTest class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : str = (CMStochasticIterativeScheduler,) __lowercase : List[str] = 10 def SCREAMING_SNAKE_CASE__ ( self:int , **_a:Optional[int] ): snake_case__ = { '''num_train_timesteps''': 2_01, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } config.update(**_a ) return config def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): snake_case__ = 10 snake_case__ = self.get_scheduler_config() snake_case__ = self.scheduler_classes[0](**_a ) scheduler.set_timesteps(_a ) snake_case__ = scheduler.timesteps[0] snake_case__ = scheduler.timesteps[1] snake_case__ = self.dummy_sample snake_case__ = 0.1 * sample snake_case__ = scheduler.step(_a , _a , _a ).prev_sample snake_case__ = scheduler.step(_a , _a , _a ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def SCREAMING_SNAKE_CASE__ ( self:Any ): for timesteps in [10, 50, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=_a ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): for clip_denoised in [True, False]: self.check_over_configs(clip_denoised=_a ) def SCREAMING_SNAKE_CASE__ ( self:int ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = 1 scheduler.set_timesteps(_a ) snake_case__ = scheduler.timesteps snake_case__ = torch.manual_seed(0 ) snake_case__ = self.dummy_model() snake_case__ = self.dummy_sample_deter * scheduler.init_noise_sigma for i, t in enumerate(_a ): # 1. scale model input snake_case__ = scheduler.scale_model_input(_a , _a ) # 2. predict noise residual snake_case__ = model(_a , _a ) # 3. predict previous sample x_t-1 snake_case__ = scheduler.step(_a , _a , _a , generator=_a ).prev_sample snake_case__ = pred_prev_sample snake_case__ = torch.sum(torch.abs(_a ) ) snake_case__ = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 192.7614 ) < 1e-2 assert abs(result_mean.item() - 0.2510 ) < 1e-3 def SCREAMING_SNAKE_CASE__ ( self:Tuple ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = [1_06, 0] scheduler.set_timesteps(timesteps=_a ) snake_case__ = scheduler.timesteps snake_case__ = torch.manual_seed(0 ) snake_case__ = self.dummy_model() snake_case__ = self.dummy_sample_deter * scheduler.init_noise_sigma for t in timesteps: # 1. scale model input snake_case__ = scheduler.scale_model_input(_a , _a ) # 2. predict noise residual snake_case__ = model(_a , _a ) # 3. predict previous sample x_t-1 snake_case__ = scheduler.step(_a , _a , _a , generator=_a ).prev_sample snake_case__ = pred_prev_sample snake_case__ = torch.sum(torch.abs(_a ) ) snake_case__ = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 347.6357 ) < 1e-2 assert abs(result_mean.item() - 0.4527 ) < 1e-3 def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = [39, 30, 12, 15, 0] with self.assertRaises(_a , msg='''`timesteps` must be in descending order.''' ): scheduler.set_timesteps(timesteps=_a ) def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = [39, 30, 12, 1, 0] snake_case__ = len(_a ) with self.assertRaises(_a , msg='''Can only pass one of `num_inference_steps` or `timesteps`.''' ): scheduler.set_timesteps(num_inference_steps=_a , timesteps=_a ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = [scheduler.config.num_train_timesteps] with self.assertRaises( _a , msg='''`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}''' , ): scheduler.set_timesteps(timesteps=_a )
33
0
from __future__ import annotations import unittest from transformers import AutoTokenizer, MBartConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFMBartForConditionalGeneration, TFMBartModel @require_tf class A : __snake_case = MBartConfig __snake_case = {} __snake_case = 'gelu' def __init__( self, UpperCamelCase__, UpperCamelCase__=13, UpperCamelCase__=7, UpperCamelCase__=True, UpperCamelCase__=False, UpperCamelCase__=99, UpperCamelCase__=32, UpperCamelCase__=2, UpperCamelCase__=4, UpperCamelCase__=37, UpperCamelCase__=0.1, UpperCamelCase__=0.1, UpperCamelCase__=20, UpperCamelCase__=2, UpperCamelCase__=1, UpperCamelCase__=0, ): """simple docstring""" lowerCAmelCase_ = parent lowerCAmelCase_ = batch_size lowerCAmelCase_ = seq_length lowerCAmelCase_ = is_training lowerCAmelCase_ = use_labels lowerCAmelCase_ = vocab_size lowerCAmelCase_ = hidden_size lowerCAmelCase_ = num_hidden_layers lowerCAmelCase_ = num_attention_heads lowerCAmelCase_ = intermediate_size lowerCAmelCase_ = hidden_dropout_prob lowerCAmelCase_ = attention_probs_dropout_prob lowerCAmelCase_ = max_position_embeddings lowerCAmelCase_ = eos_token_id lowerCAmelCase_ = pad_token_id lowerCAmelCase_ = bos_token_id def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = ids_tensor([self.batch_size, self.seq_length - 1], self.vocab_size ) lowerCAmelCase_ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ), 1 ) lowerCAmelCase_ = tf.concat([input_ids, eos_tensor], axis=1 ) lowerCAmelCase_ = ids_tensor([self.batch_size, self.seq_length], self.vocab_size ) lowerCAmelCase_ = self.config_cls( vocab_size=self.vocab_size, d_model=self.hidden_size, encoder_layers=self.num_hidden_layers, decoder_layers=self.num_hidden_layers, encoder_attention_heads=self.num_attention_heads, decoder_attention_heads=self.num_attention_heads, encoder_ffn_dim=self.intermediate_size, decoder_ffn_dim=self.intermediate_size, dropout=self.hidden_dropout_prob, attention_dropout=self.attention_probs_dropout_prob, max_position_embeddings=self.max_position_embeddings, eos_token_ids=[2], bos_token_id=self.bos_token_id, pad_token_id=self.pad_token_id, decoder_start_token_id=self.pad_token_id, **self.config_updates, ) lowerCAmelCase_ = prepare_mbart_inputs_dict(_a, _a, _a ) return config, inputs_dict def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__, UpperCamelCase__ ): """simple docstring""" lowerCAmelCase_ = TFMBartModel(config=_a ).get_decoder() lowerCAmelCase_ = inputs_dict['''input_ids'''] lowerCAmelCase_ = input_ids[:1, :] lowerCAmelCase_ = inputs_dict['''attention_mask'''][:1, :] lowerCAmelCase_ = inputs_dict['''head_mask'''] lowerCAmelCase_ = 1 # first forward pass lowerCAmelCase_ = model(_a, attention_mask=_a, head_mask=_a, use_cache=_a ) lowerCAmelCase_ , lowerCAmelCase_ = outputs.to_tuple() lowerCAmelCase_ = past_key_values[1] def __UpperCamelCase ( _A , _A , _A , _A=None , _A=None , _A=None , _A=None , _A=None , ): if attention_mask is None: lowerCAmelCase_ = tf.cast(tf.math.not_equal(__lowerCAmelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: lowerCAmelCase_ = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: lowerCAmelCase_ = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: lowerCAmelCase_ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: lowerCAmelCase_ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class A ( snake_case_ , snake_case_ , unittest.TestCase ): __snake_case = (TFMBartForConditionalGeneration, TFMBartModel) if is_tf_available() else () __snake_case = (TFMBartForConditionalGeneration,) if is_tf_available() else () __snake_case = ( { 'conversational': TFMBartForConditionalGeneration, 'feature-extraction': TFMBartModel, 'summarization': TFMBartForConditionalGeneration, 'text2text-generation': TFMBartForConditionalGeneration, 'translation': TFMBartForConditionalGeneration, } if is_tf_available() else {} ) __snake_case = True __snake_case = False __snake_case = False def SCREAMING_SNAKE_CASE__ ( self, UpperCamelCase__, UpperCamelCase__, UpperCamelCase__, UpperCamelCase__, UpperCamelCase__ ): """simple docstring""" if pipeline_test_casse_name != "FeatureExtractionPipelineTests": # Exception encountered when calling layer '...' return True return False def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = TFMBartModelTester(self ) lowerCAmelCase_ = ConfigTester(self, config_class=_a ) def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_a ) @require_sentencepiece @require_tokenizers @require_tf class A ( unittest.TestCase ): __snake_case = [ ' UN Chief Says There Is No Military Solution in Syria', ] __snake_case = [ 'Şeful ONU declară că nu există o soluţie militară în Siria', ] __snake_case = 'facebook/mbart-large-en-ro' @cached_property def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" return AutoTokenizer.from_pretrained(self.model_name ) @cached_property def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model def SCREAMING_SNAKE_CASE__ ( self, **UpperCamelCase__ ): """simple docstring""" lowerCAmelCase_ = self.translate_src_text(**_a ) self.assertListEqual(self.expected_text, _a ) def SCREAMING_SNAKE_CASE__ ( self, **UpperCamelCase__ ): """simple docstring""" lowerCAmelCase_ = self.tokenizer(self.src_text, **_a, return_tensors='''tf''' ) lowerCAmelCase_ = self.model.generate( model_inputs.input_ids, attention_mask=model_inputs.attention_mask, num_beams=2 ) lowerCAmelCase_ = self.tokenizer.batch_decode(_a, skip_special_tokens=_a ) return generated_words @slow def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" self._assert_generated_batch_equal_expected()
431
import numpy as np def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> np.ndarray: return 1 / (1 + np.exp(-vector )) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> np.ndarray: return vector * sigmoid(__lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
33
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A__ : List[Any] = { """configuration_x_clip""": [ """XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XCLIPConfig""", """XCLIPTextConfig""", """XCLIPVisionConfig""", ], """processing_x_clip""": ["""XCLIPProcessor"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : Optional[Any] = [ """XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """XCLIPModel""", """XCLIPPreTrainedModel""", """XCLIPTextModel""", """XCLIPVisionModel""", ] if TYPE_CHECKING: from .configuration_x_clip import ( XCLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, XCLIPConfig, XCLIPTextConfig, XCLIPVisionConfig, ) from .processing_x_clip import XCLIPProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_x_clip import ( XCLIP_PRETRAINED_MODEL_ARCHIVE_LIST, XCLIPModel, XCLIPPreTrainedModel, XCLIPTextModel, XCLIPVisionModel, ) else: import sys A__ : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
171
def SCREAMING_SNAKE_CASE ( __lowerCAmelCase = 100 ) -> int: snake_case__ = set() snake_case__ = 0 snake_case__ = n + 1 # maximum limit for a in range(2 , __lowerCAmelCase ): for b in range(2 , __lowerCAmelCase ): snake_case__ = a**b # calculates the current power collect_powers.add(__lowerCAmelCase ) # adds the result to the set return len(__lowerCAmelCase ) if __name__ == "__main__": print("""Number of terms """, solution(int(str(input()).strip())))
33
0
from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def SCREAMING_SNAKE_CASE_ ( UpperCAmelCase_ : Optional[Any] ) -> Dict: if not is_accelerate_available(): return method SCREAMING_SNAKE_CASE_ : List[str] =version.parse(accelerate.__version__ ).base_version if version.parse(__lowerCAmelCase ) < version.parse('''0.17.0''' ): return method def wrapper(self : Union[str, Any] , *UpperCAmelCase_ : Optional[int] , **UpperCAmelCase_ : List[Any] ): if hasattr(self , '''_hf_hook''' ) and hasattr(self._hf_hook , '''pre_forward''' ): self._hf_hook.pre_forward(self ) return method(self , *__lowerCAmelCase , **__lowerCAmelCase ) return wrapper
443
from copy import deepcopy class __magic_name__ : '''simple docstring''' def __init__( self:int , _a:list[int] | None = None , _a:int | None = None ): if arr is None and size is not None: snake_case__ = size snake_case__ = [0] * size elif arr is not None: self.init(_a ) else: raise ValueError('''Either arr or size must be specified''' ) def SCREAMING_SNAKE_CASE__ ( self:Any , _a:list[int] ): snake_case__ = len(_a ) snake_case__ = deepcopy(_a ) for i in range(1 , self.size ): snake_case__ = self.next_(_a ) if j < self.size: self.tree[j] += self.tree[i] def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = self.tree[:] for i in range(self.size - 1 , 0 , -1 ): snake_case__ = self.next_(_a ) if j < self.size: arr[j] -= arr[i] return arr @staticmethod def SCREAMING_SNAKE_CASE__ ( _a:int ): return index + (index & (-index)) @staticmethod def SCREAMING_SNAKE_CASE__ ( _a:int ): return index - (index & (-index)) def SCREAMING_SNAKE_CASE__ ( self:List[Any] , _a:int , _a:int ): if index == 0: self.tree[0] += value return while index < self.size: self.tree[index] += value snake_case__ = self.next_(_a ) def SCREAMING_SNAKE_CASE__ ( self:Dict , _a:int , _a:int ): self.add(_a , value - self.get(_a ) ) def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] , _a:int ): if right == 0: return 0 snake_case__ = self.tree[0] right -= 1 # make right inclusive while right > 0: result += self.tree[right] snake_case__ = self.prev(_a ) return result def SCREAMING_SNAKE_CASE__ ( self:Dict , _a:int , _a:int ): return self.prefix(_a ) - self.prefix(_a ) def SCREAMING_SNAKE_CASE__ ( self:str , _a:int ): return self.query(_a , index + 1 ) def SCREAMING_SNAKE_CASE__ ( self:str , _a:int ): value -= self.tree[0] if value < 0: return -1 snake_case__ = 1 # Largest power of 2 <= size while j * 2 < self.size: j *= 2 snake_case__ = 0 while j > 0: if i + j < self.size and self.tree[i + j] <= value: value -= self.tree[i + j] i += j j //= 2 return i if __name__ == "__main__": import doctest doctest.testmod()
33
0
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def _lowerCamelCase ( __a ): SCREAMING_SNAKE_CASE_ = args.pruning_method SCREAMING_SNAKE_CASE_ = args.threshold SCREAMING_SNAKE_CASE_ = args.model_name_or_path.rstrip('''/''' ) SCREAMING_SNAKE_CASE_ = args.target_model_path print(F'Load fine-pruned model from {model_name_or_path}' ) SCREAMING_SNAKE_CASE_ = torch.load(os.path.join(__lowerCAmelCase, '''pytorch_model.bin''' ) ) SCREAMING_SNAKE_CASE_ = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: SCREAMING_SNAKE_CASE_ = tensor print(F'Copied layer {name}' ) elif "classifier" in name or "qa_output" in name: SCREAMING_SNAKE_CASE_ = tensor print(F'Copied layer {name}' ) elif "bias" in name: SCREAMING_SNAKE_CASE_ = tensor print(F'Copied layer {name}' ) else: if pruning_method == "magnitude": SCREAMING_SNAKE_CASE_ = MagnitudeBinarizer.apply(inputs=__lowerCAmelCase, threshold=__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = tensor * mask print(F'Pruned layer {name}' ) elif pruning_method == "topK": if "mask_scores" in name: continue SCREAMING_SNAKE_CASE_ = name[:-6] SCREAMING_SNAKE_CASE_ = model[F'{prefix_}mask_scores'] SCREAMING_SNAKE_CASE_ = TopKBinarizer.apply(__lowerCAmelCase, __lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = tensor * mask print(F'Pruned layer {name}' ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue SCREAMING_SNAKE_CASE_ = name[:-6] SCREAMING_SNAKE_CASE_ = model[F'{prefix_}mask_scores'] SCREAMING_SNAKE_CASE_ = ThresholdBinarizer.apply(__lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = tensor * mask print(F'Pruned layer {name}' ) elif pruning_method == "l0": if "mask_scores" in name: continue SCREAMING_SNAKE_CASE_ = name[:-6] SCREAMING_SNAKE_CASE_ = model[F'{prefix_}mask_scores'] SCREAMING_SNAKE_CASE_ ,SCREAMING_SNAKE_CASE_ = -0.1, 1.1 SCREAMING_SNAKE_CASE_ = torch.sigmoid(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = s * (r - l) + l SCREAMING_SNAKE_CASE_ = s_bar.clamp(min=0.0, max=1.0 ) SCREAMING_SNAKE_CASE_ = tensor * mask print(F'Pruned layer {name}' ) else: raise ValueError('''Unknown pruning method''' ) if target_model_path is None: SCREAMING_SNAKE_CASE_ = os.path.join( os.path.dirname(__lowerCAmelCase ), F'bertarized_{os.path.basename(__lowerCAmelCase )}' ) if not os.path.isdir(__lowerCAmelCase ): shutil.copytree(__lowerCAmelCase, __lowerCAmelCase ) print(F'\nCreated folder {target_model_path}' ) torch.save(__lowerCAmelCase, os.path.join(__lowerCAmelCase, '''pytorch_model.bin''' ) ) print('''\nPruned model saved! See you later!''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '--pruning_method', choices=['l0', 'magnitude', 'topK', 'sigmoied_threshold'], type=str, required=True, help=( 'Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,' ' sigmoied_threshold = Soft movement pruning)' ), ) parser.add_argument( '--threshold', type=float, required=False, help=( 'For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.' 'For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.' 'Not needed for `l0`' ), ) parser.add_argument( '--model_name_or_path', type=str, required=True, help='Folder containing the model that was previously fine-pruned', ) parser.add_argument( '--target_model_path', default=None, type=str, required=False, help='Folder containing the model that was previously fine-pruned', ) lowerCAmelCase__ = parser.parse_args() main(args)
626
from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class __magic_name__ : '''simple docstring''' __lowercase : int = BlenderbotConfig __lowercase : Any = {} __lowercase : Optional[Any] = 'gelu' def __init__( self:Tuple , _a:Optional[Any] , _a:Optional[Any]=13 , _a:Tuple=7 , _a:Union[str, Any]=True , _a:int=False , _a:int=99 , _a:Optional[int]=32 , _a:List[str]=2 , _a:List[str]=4 , _a:List[Any]=37 , _a:Any=0.1 , _a:int=0.1 , _a:List[Any]=20 , _a:List[str]=2 , _a:int=1 , _a:Dict=0 , ): snake_case__ = parent snake_case__ = batch_size snake_case__ = seq_length snake_case__ = is_training snake_case__ = use_labels snake_case__ = vocab_size snake_case__ = hidden_size snake_case__ = num_hidden_layers snake_case__ = num_attention_heads snake_case__ = intermediate_size snake_case__ = hidden_dropout_prob snake_case__ = attention_probs_dropout_prob snake_case__ = max_position_embeddings snake_case__ = eos_token_id snake_case__ = pad_token_id snake_case__ = bos_token_id def SCREAMING_SNAKE_CASE__ ( self:int ): snake_case__ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) snake_case__ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) snake_case__ = tf.concat([input_ids, eos_tensor] , axis=1 ) snake_case__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case__ = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) snake_case__ = prepare_blenderbot_inputs_dict(_a , _a , _a ) return config, inputs_dict def SCREAMING_SNAKE_CASE__ ( self:int , _a:Optional[Any] , _a:int ): snake_case__ = TFBlenderbotModel(config=_a ).get_decoder() snake_case__ = inputs_dict['''input_ids'''] snake_case__ = input_ids[:1, :] snake_case__ = inputs_dict['''attention_mask'''][:1, :] snake_case__ = inputs_dict['''head_mask'''] snake_case__ = 1 # first forward pass snake_case__ = model(_a , attention_mask=_a , head_mask=_a , use_cache=_a ) snake_case__ , snake_case__ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids snake_case__ = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case__ = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and snake_case__ = tf.concat([input_ids, next_tokens] , axis=-1 ) snake_case__ = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) snake_case__ = model(_a , attention_mask=_a )[0] snake_case__ = model(_a , attention_mask=_a , past_key_values=_a )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice snake_case__ = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) snake_case__ = output_from_no_past[:, -3:, random_slice_idx] snake_case__ = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_a , _a , rtol=1e-3 ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , ) -> Tuple: if attention_mask is None: snake_case__ = tf.cast(tf.math.not_equal(__lowerCAmelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: snake_case__ = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: snake_case__ = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: snake_case__ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: snake_case__ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class __magic_name__ (snake_case_ ,snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : List[str] = (TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () __lowercase : Any = (TFBlenderbotForConditionalGeneration,) if is_tf_available() else () __lowercase : Tuple = ( { 'conversational': TFBlenderbotForConditionalGeneration, 'feature-extraction': TFBlenderbotModel, 'summarization': TFBlenderbotForConditionalGeneration, 'text2text-generation': TFBlenderbotForConditionalGeneration, 'translation': TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) __lowercase : Any = True __lowercase : int = False __lowercase : int = False def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): snake_case__ = TFBlenderbotModelTester(self ) snake_case__ = ConfigTester(self , config_class=_a ) def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): snake_case__ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_a ) @require_tokenizers @require_tf class __magic_name__ (unittest.TestCase ): '''simple docstring''' __lowercase : Optional[int] = ['My friends are cool but they eat too many carbs.'] __lowercase : Optional[int] = 'facebook/blenderbot-400M-distill' @cached_property def SCREAMING_SNAKE_CASE__ ( self:Tuple ): return BlenderbotTokenizer.from_pretrained(self.model_name ) @cached_property def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): snake_case__ = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = self.tokenizer(self.src_text , return_tensors='''tf''' ) snake_case__ = self.model.generate( model_inputs.input_ids , ) snake_case__ = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=_a )[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
33
0
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class __lowerCamelCase ( unittest.TestCase ): @slow def UpperCAmelCase__ ( self ): lowerCamelCase_ = FlaxMTaForConditionalGeneration.from_pretrained('''google/mt5-small''' ) lowerCamelCase_ = AutoTokenizer.from_pretrained('''google/mt5-small''' ) lowerCamelCase_ = tokenizer('''Hello there''' , return_tensors='''np''' ).input_ids lowerCamelCase_ = tokenizer('''Hi I am''' , return_tensors='''np''' ).input_ids lowerCamelCase_ = shift_tokens_right(_a , model.config.pad_token_id , model.config.decoder_start_token_id ) lowerCamelCase_ = model(_a , decoder_input_ids=_a ).logits lowerCamelCase_ = optax.softmax_cross_entropy(_a , onehot(_a , logits.shape[-1] ) ).mean() lowerCamelCase_ = -(labels.shape[-1] * loss.item()) lowerCamelCase_ = -8_4.9_1_2_7 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1e-4 )
29
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class __magic_name__ (unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = 0 def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): snake_case__ = AutoImageProcessor.from_pretrained('''openai/clip-vit-base-patch32''' ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:str ): with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = Path(_a ) / '''preprocessor_config.json''' snake_case__ = Path(_a ) / '''config.json''' json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) json.dump({'''model_type''': '''clip'''} , open(_a , '''w''' ) ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = Path(_a ) / '''preprocessor_config.json''' snake_case__ = Path(_a ) / '''config.json''' json.dump( {'''feature_extractor_type''': '''CLIPFeatureExtractor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) json.dump({'''model_type''': '''clip'''} , open(_a , '''w''' ) ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = CLIPConfig() # Create a dummy config file with image_proceesor_type snake_case__ = Path(_a ) / '''preprocessor_config.json''' snake_case__ = Path(_a ) / '''config.json''' json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) json.dump({'''model_type''': '''clip'''} , open(_a , '''w''' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally snake_case__ = AutoImageProcessor.from_pretrained(_a ).to_dict() config_dict.pop('''image_processor_type''' ) snake_case__ = CLIPImageProcessor(**_a ) # save in new folder model_config.save_pretrained(_a ) config.save_pretrained(_a ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) # make sure private variable is not incorrectly saved snake_case__ = json.loads(config.to_json_string() ) self.assertTrue('''_processor_class''' not in dict_as_saved ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = Path(_a ) / '''preprocessor_config.json''' json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): with self.assertRaisesRegex( _a , '''clip-base is not a local folder and is not a valid model identifier''' ): snake_case__ = AutoImageProcessor.from_pretrained('''clip-base''' ) def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): with self.assertRaisesRegex( _a , r'''aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)''' ): snake_case__ = AutoImageProcessor.from_pretrained(_a , revision='''aaaaaa''' ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): with self.assertRaisesRegex( _a , '''hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.''' , ): snake_case__ = AutoImageProcessor.from_pretrained('''hf-internal-testing/config-no-model''' ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(_a ): snake_case__ = AutoImageProcessor.from_pretrained('''hf-internal-testing/test_dynamic_image_processor''' ) # If remote code is disabled, we can't load this config. with self.assertRaises(_a ): snake_case__ = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=_a ) snake_case__ = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=_a ) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_a ) snake_case__ = AutoImageProcessor.from_pretrained(_a , trust_remote_code=_a ) self.assertEqual(reloaded_image_processor.__class__.__name__ , '''NewImageProcessor''' ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): try: AutoConfig.register('''custom''' , _a ) AutoImageProcessor.register(_a , _a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_a ): AutoImageProcessor.register(_a , _a ) with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = Path(_a ) / '''preprocessor_config.json''' snake_case__ = Path(_a ) / '''config.json''' json.dump( {'''feature_extractor_type''': '''CLIPFeatureExtractor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) json.dump({'''model_type''': '''clip'''} , open(_a , '''w''' ) ) snake_case__ = CustomImageProcessor.from_pretrained(_a ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_a ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : List[str] = True try: AutoConfig.register('''custom''' , _a ) AutoImageProcessor.register(_a , _a ) # If remote code is not set, the default is to use local snake_case__ = AutoImageProcessor.from_pretrained('''hf-internal-testing/test_dynamic_image_processor''' ) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. snake_case__ = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=_a ) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub snake_case__ = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=_a ) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''' ) self.assertTrue(not hasattr(_a , '''is_local''' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
33
0
"""simple docstring""" def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> list: """simple docstring""" _UpperCamelCase = len(__lowerCAmelCase ) _UpperCamelCase = [[0] * n for i in range(__lowerCAmelCase )] for i in range(__lowerCAmelCase ): _UpperCamelCase = y_points[i] for i in range(2, __lowerCAmelCase ): for j in range(__lowerCAmelCase, __lowerCAmelCase ): _UpperCamelCase = ( (xa - x_points[j - i + 1]) * q[j][i - 1] - (xa - x_points[j]) * q[j - 1][i - 1] ) / (x_points[j] - x_points[j - i + 1]) return [q[n - 1][n - 1], q] if __name__ == "__main__": import doctest doctest.testmod()
19
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase__ : int = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase=False ) -> int: snake_case__ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" snake_case__ = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False ) -> Dict: for i in range(config.num_hidden_layers ): if base_model: snake_case__ = '''''' else: snake_case__ = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) snake_case__ = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) snake_case__ = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict snake_case__ = in_proj_weight[ : config.hidden_size, : ] snake_case__ = in_proj_bias[: config.hidden_size] snake_case__ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] snake_case__ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] snake_case__ = in_proj_weight[ -config.hidden_size :, : ] snake_case__ = in_proj_bias[-config.hidden_size :] def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> Optional[Any]: snake_case__ = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Union[str, Any]: snake_case__ = dct.pop(__lowerCAmelCase ) snake_case__ = val def SCREAMING_SNAKE_CASE ( ) -> str: snake_case__ = '''http://images.cocodataset.org/val2017/000000039769.jpg''' snake_case__ = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> Dict: snake_case__ = ViTConfig() snake_case__ = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": snake_case__ = True snake_case__ = int(vit_name[-12:-10] ) snake_case__ = int(vit_name[-9:-6] ) else: snake_case__ = 1000 snake_case__ = '''huggingface/label-files''' snake_case__ = '''imagenet-1k-id2label.json''' snake_case__ = json.load(open(hf_hub_download(__lowerCAmelCase , __lowerCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) snake_case__ = {int(__lowerCAmelCase ): v for k, v in idalabel.items()} snake_case__ = idalabel snake_case__ = {v: k for k, v in idalabel.items()} snake_case__ = int(vit_name[-6:-4] ) snake_case__ = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith('''tiny''' ): snake_case__ = 192 snake_case__ = 768 snake_case__ = 12 snake_case__ = 3 elif vit_name[9:].startswith('''small''' ): snake_case__ = 384 snake_case__ = 1536 snake_case__ = 12 snake_case__ = 6 else: pass else: if vit_name[4:].startswith('''small''' ): snake_case__ = 768 snake_case__ = 2304 snake_case__ = 8 snake_case__ = 8 elif vit_name[4:].startswith('''base''' ): pass elif vit_name[4:].startswith('''large''' ): snake_case__ = 1024 snake_case__ = 4096 snake_case__ = 24 snake_case__ = 16 elif vit_name[4:].startswith('''huge''' ): snake_case__ = 1280 snake_case__ = 5120 snake_case__ = 32 snake_case__ = 16 # load original model from timm snake_case__ = timm.create_model(__lowerCAmelCase , pretrained=__lowerCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys snake_case__ = timm_model.state_dict() if base_model: remove_classification_head_(__lowerCAmelCase ) snake_case__ = create_rename_keys(__lowerCAmelCase , __lowerCAmelCase ) for src, dest in rename_keys: rename_key(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) read_in_q_k_v(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # load HuggingFace model if vit_name[-5:] == "in21k": snake_case__ = ViTModel(__lowerCAmelCase ).eval() else: snake_case__ = ViTForImageClassification(__lowerCAmelCase ).eval() model.load_state_dict(__lowerCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: snake_case__ = DeiTImageProcessor(size=config.image_size ) else: snake_case__ = ViTImageProcessor(size=config.image_size ) snake_case__ = image_processor(images=prepare_img() , return_tensors='''pt''' ) snake_case__ = encoding['''pixel_values'''] snake_case__ = model(__lowerCAmelCase ) if base_model: snake_case__ = timm_model.forward_features(__lowerCAmelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__lowerCAmelCase , outputs.pooler_output , atol=1e-3 ) else: snake_case__ = timm_model(__lowerCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__lowerCAmelCase , outputs.logits , atol=1e-3 ) Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) print(F"""Saving model {vit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCAmelCase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": lowerCamelCase__ : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--vit_name""", default="""vit_base_patch16_224""", type=str, help="""Name of the ViT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) lowerCamelCase__ : str = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
33
0
import unittest from transformers import is_tf_available from transformers.testing_utils import require_tf if is_tf_available(): import tensorflow as tf from tensorflow.python.eager import context from tensorflow.python.framework import ops from transformers import GradientAccumulator, create_optimizer @require_tf class lowercase_ ( unittest.TestCase ): '''simple docstring''' def __lowerCAmelCase ( self : Optional[int] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : List[Any] ) ->Dict: """simple docstring""" self.assertEqual(len(_a ) , len(_a ) ) for a, b in zip(_a , _a ): self.assertAlmostEqual(_a , _a , delta=_a ) def __lowerCAmelCase ( self : str ) ->Optional[Any]: """simple docstring""" a = GradientAccumulator() accumulator([tf.constant([1.0, 2.0] )] ) accumulator([tf.constant([-2.0, 1.0] )] ) accumulator([tf.constant([-1.0, 2.0] )] ) with self.assertRaises(_a ): accumulator([tf.constant([1.0, 1.0] ), tf.constant([2.0, 2.0] )] ) self.assertEqual(accumulator.step , 3 ) self.assertEqual(len(accumulator.gradients ) , 1 ) self.assertListAlmostEqual(accumulator.gradients[0].numpy().tolist() , [-2.0, 5.0] , tol=1e-2 ) accumulator.reset() self.assertEqual(accumulator.step , 0 ) self.assertListAlmostEqual(accumulator.gradients[0].numpy().tolist() , [0.0, 0.0] , tol=1e-2 ) def __lowerCAmelCase ( self : Union[str, Any] ) ->Dict: """simple docstring""" a = None ops.enable_eager_execution_internal() a = tf.config.list_physical_devices('''CPU''' ) if len(_a ) == 1: tf.config.set_logical_device_configuration( physical_devices[0] , [tf.config.LogicalDeviceConfiguration(), tf.config.LogicalDeviceConfiguration()] ) a = tf.config.list_logical_devices(device_type='''CPU''' ) a = tf.distribute.MirroredStrategy(devices=devices[:2] ) with strategy.scope(): a = GradientAccumulator() a = tf.Variable([4.0, 3.0] ) a , a = create_optimizer(5e-5 , 10 , 5 ) a = tf.Variable([0.0, 0.0] , trainable=_a ) def accumulate_on_replica(__UpperCAmelCase : Any ): accumulator([gradient] ) def apply_on_replica(): optimizer.apply_gradients(list(zip(accumulator.gradients , [variable] ) ) ) @tf.function def accumulate(__UpperCAmelCase : Optional[int] , __UpperCAmelCase : List[str] ): with strategy.scope(): a = strategy.experimental_local_results(_a ) local_variables[0].assign(_a ) local_variables[1].assign(_a ) strategy.run(_a , args=(gradient_placeholder,) ) @tf.function def apply_grad(): with strategy.scope(): strategy.run(_a ) def _check_local_values(__UpperCAmelCase : Any , __UpperCAmelCase : Any ): a = strategy.experimental_local_results(accumulator._gradients[0] ) self.assertListAlmostEqual(values[0].value() , _a , tol=1e-2 ) self.assertListAlmostEqual(values[1].value() , _a , tol=1e-2 ) accumulate([1.0, 2.0] , [-1.0, 1.0] ) accumulate([3.0, -1.0] , [-1.0, -1.0] ) accumulate([-2.0, 2.0] , [3.0, -2.0] ) self.assertEqual(accumulator.step , 3 ) _check_local_values([2.0, 3.0] , [1.0, -2.0] ) apply_grad() self.assertListAlmostEqual(variable.value() , [4.0, 3.0] , tol=1e-2 ) accumulator.reset() self.assertEqual(accumulator.step , 0 ) _check_local_values([0.0, 0.0] , [0.0, 0.0] )
117
import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : List[str] = ['image_processor', 'tokenizer'] __lowercase : str = 'AutoImageProcessor' __lowercase : Dict = 'AutoTokenizer' def __init__( self:int , _a:List[str]=None , _a:Optional[Any]=None , **_a:List[str] ): snake_case__ = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _a , ) snake_case__ = kwargs.pop('''feature_extractor''' ) snake_case__ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_a , _a ) snake_case__ = self.image_processor snake_case__ = False def __call__( self:Optional[int] , *_a:str , **_a:int ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*_a , **_a ) snake_case__ = kwargs.pop('''images''' , _a ) snake_case__ = kwargs.pop('''text''' , _a ) if len(_a ) > 0: snake_case__ = args[0] snake_case__ = args[1:] if images is None and text is None: raise ValueError('''You need to specify either an `images` or `text` input to process.''' ) if images is not None: snake_case__ = self.image_processor(_a , *_a , **_a ) if text is not None: snake_case__ = self.tokenizer(_a , **_a ) if text is None: return inputs elif images is None: return encodings else: snake_case__ = encodings['''input_ids'''] return inputs def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] , *_a:Union[str, Any] , **_a:Any ): return self.tokenizer.batch_decode(*_a , **_a ) def SCREAMING_SNAKE_CASE__ ( self:Tuple , *_a:Union[str, Any] , **_a:Optional[int] ): return self.tokenizer.decode(*_a , **_a ) @contextmanager def SCREAMING_SNAKE_CASE__ ( self:Tuple ): warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your images inputs, or in a separate call.''' ) snake_case__ = True snake_case__ = self.tokenizer yield snake_case__ = self.image_processor snake_case__ = False def SCREAMING_SNAKE_CASE__ ( self:List[str] , _a:Dict , _a:Dict=False , _a:Optional[int]=None ): if added_vocab is None: snake_case__ = self.tokenizer.get_added_vocab() snake_case__ = {} while tokens: snake_case__ = re.search(r'''<s_(.*?)>''' , _a , re.IGNORECASE ) if start_token is None: break snake_case__ = start_token.group(1 ) snake_case__ = re.search(rF"""</s_{key}>""" , _a , re.IGNORECASE ) snake_case__ = start_token.group() if end_token is None: snake_case__ = tokens.replace(_a , '''''' ) else: snake_case__ = end_token.group() snake_case__ = re.escape(_a ) snake_case__ = re.escape(_a ) snake_case__ = re.search(F"""{start_token_escaped}(.*?){end_token_escaped}""" , _a , re.IGNORECASE ) if content is not None: snake_case__ = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node snake_case__ = self.tokenajson(_a , is_inner_value=_a , added_vocab=_a ) if value: if len(_a ) == 1: snake_case__ = value[0] snake_case__ = value else: # leaf nodes snake_case__ = [] for leaf in content.split(r'''<sep/>''' ): snake_case__ = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": snake_case__ = leaf[1:-2] # for categorical special tokens output[key].append(_a ) if len(output[key] ) == 1: snake_case__ = output[key][0] snake_case__ = tokens[tokens.find(_a ) + len(_a ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:] , is_inner_value=_a , added_vocab=_a ) if len(_a ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , _a , ) return self.image_processor_class @property def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , _a , ) return self.image_processor
33
0
import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer _lowerCamelCase : Tuple = """bart""" _lowerCamelCase : Dict = True @st.cache(allow_output_mutation=__lowerCAmelCase ) def A__ ( ) ->Optional[Any]: if LOAD_DENSE_INDEX: __A =AutoTokenizer.from_pretrained('''yjernite/retribert-base-uncased''' ) __A =AutoModel.from_pretrained('''yjernite/retribert-base-uncased''' ).to('''cuda:0''' ) __A =qar_model.eval() else: __A , __A =(None, None) if MODEL_TYPE == "bart": __A =AutoTokenizer.from_pretrained('''yjernite/bart_eli5''' ) __A =AutoModelForSeqaSeqLM.from_pretrained('''yjernite/bart_eli5''' ).to('''cuda:0''' ) __A =torch.load('''seq2seq_models/eli5_bart_model_blm_2.pth''' ) sas_model.load_state_dict(save_dict['''model'''] ) __A =sas_model.eval() else: __A , __A =make_qa_sas_model( model_name='''t5-small''' , from_file='''seq2seq_models/eli5_t5_model_1024_4.pth''' , device='''cuda:0''' ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=__lowerCAmelCase ) def A__ ( ) ->int: if LOAD_DENSE_INDEX: __A =faiss.StandardGpuResources() __A =datasets.load_dataset(path='''wiki_snippets''' , name='''wiki40b_en_100_0''' )['''train'''] __A =np.memmap( '''wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat''' , dtype='''float32''' , mode='''r''' , shape=(wikiaab_passages.num_rows, 1_28) , ) __A =faiss.IndexFlatIP(1_28 ) __A =faiss.index_cpu_to_gpu(__lowerCAmelCase , 1 , __lowerCAmelCase ) wikiaab_gpu_index_flat.add(__lowerCAmelCase ) # TODO fix for larger GPU else: __A , __A =(None, None) __A =Elasticsearch([{'''host''': '''localhost''', '''port''': '''9200'''}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=__lowerCAmelCase ) def A__ ( ) ->List[str]: __A =datasets.load_dataset('''eli5''' , name='''LFQA_reddit''' ) __A =elia['''train_eli5'''] __A =np.memmap( '''eli5_questions_reps.dat''' , dtype='''float32''' , mode='''r''' , shape=(elia_train.num_rows, 1_28) ) __A =faiss.IndexFlatIP(1_28 ) eli5_train_q_index.add(__lowerCAmelCase ) return (elia_train, eli5_train_q_index) _lowerCamelCase : Optional[int] = load_indexes() _lowerCamelCase : List[str] = load_models() _lowerCamelCase : Union[str, Any] = load_train_data() def A__ ( __A : str , __A : Tuple=10 ) ->List[Any]: __A =embed_questions_for_retrieval([question] , __lowerCAmelCase , __lowerCAmelCase ) __A , __A =eli5_train_q_index.search(__lowerCAmelCase , __lowerCAmelCase ) __A =[elia_train[int(__lowerCAmelCase )] for i in I[0]] return nn_examples def A__ ( __A : List[Any] , __A : str="wiki40b" , __A : Optional[int]="dense" , __A : Any=10 ) ->int: if source == "none": __A , __A =(''' <P> '''.join(['''''' for _ in range(11 )] ).strip(), []) else: if method == "dense": __A , __A =query_qa_dense_index( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) else: __A , __A =query_es_index( __lowerCAmelCase , __lowerCAmelCase , index_name='''english_wiki40b_snippets_100w''' , n_results=__lowerCAmelCase , ) __A =[ (res['''article_title'''], res['''section_title'''].strip(), res['''score'''], res['''passage_text''']) for res in hit_lst ] __A ='''question: {} context: {}'''.format(__lowerCAmelCase , __lowerCAmelCase ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda __A : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda __A : None), } ) def A__ ( __A : int , __A : List[str] , __A : str , __A : int=64 , __A : int=2_56 , __A : Any=False , __A : Union[str, Any]=2 , __A : Union[str, Any]=0.95 , __A : Optional[int]=0.8 ) ->Dict: with torch.no_grad(): __A =qa_sas_generate( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , num_answers=1 , num_beams=__lowerCAmelCase , min_len=__lowerCAmelCase , max_len=__lowerCAmelCase , do_sample=__lowerCAmelCase , temp=__lowerCAmelCase , top_p=__lowerCAmelCase , top_k=__lowerCAmelCase , max_input_length=10_24 , device='''cuda:0''' , )[0] return (answer, support_list) st.title('''Long Form Question Answering with ELI5''') # Start sidebar _lowerCamelCase : int = """<img src='https://huggingface.co/front/assets/huggingface_logo.svg'>""" _lowerCamelCase : List[Any] = """ <html> <head> <style> .img-container { padding-left: 90px; padding-right: 90px; padding-top: 50px; padding-bottom: 50px; background-color: #f0f3f9; } </style> </head> <body> <span class=\"img-container\"> <!-- Inline parent element --> %s </span> </body> </html> """ % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia _lowerCamelCase : Any = """ This demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html). First, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset, a pre-processed fixed snapshot of Wikipedia. """ st.sidebar.markdown(description, unsafe_allow_html=True) _lowerCamelCase : str = [ """Answer the question""", """View the retrieved document only""", """View the most similar ELI5 question and answer""", """Show me everything, please!""", ] _lowerCamelCase : Tuple = st.sidebar.checkbox('''Demo options''') if demo_options: _lowerCamelCase : List[str] = st.sidebar.selectbox( '''''', action_list, index=3, ) _lowerCamelCase : Union[str, Any] = action_list.index(action_st) _lowerCamelCase : List[Any] = st.sidebar.selectbox( '''''', ['''Show full text of passages''', '''Show passage section titles'''], index=0, ) _lowerCamelCase : int = show_type == """Show full text of passages""" else: _lowerCamelCase : str = 3 _lowerCamelCase : Any = True _lowerCamelCase : Tuple = st.sidebar.checkbox('''Retrieval options''') if retrieval_options: _lowerCamelCase : Optional[int] = """ ### Information retriever options The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs. The answer is then generated by sequence to sequence model which takes the question and retrieved document as input. """ st.sidebar.markdown(retriever_info) _lowerCamelCase : Optional[int] = st.sidebar.selectbox('''Which Wikipedia format should the model use?''', ['''wiki40b''', '''none''']) _lowerCamelCase : int = st.sidebar.selectbox('''Which Wikipedia indexer should the model use?''', ['''dense''', '''sparse''', '''mixed''']) else: _lowerCamelCase : str = """wiki40b""" _lowerCamelCase : str = """dense""" _lowerCamelCase : int = """beam""" _lowerCamelCase : Optional[int] = 2 _lowerCamelCase : int = 64 _lowerCamelCase : List[Any] = 256 _lowerCamelCase : Optional[Any] = None _lowerCamelCase : List[str] = None _lowerCamelCase : str = st.sidebar.checkbox('''Generation options''') if generate_options: _lowerCamelCase : int = """ ### Answer generation options The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large) weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with **beam** search, or **sample** from the decoder's output probabilities. """ st.sidebar.markdown(generate_info) _lowerCamelCase : List[str] = st.sidebar.selectbox('''Would you like to use beam search or sample an answer?''', ['''beam''', '''sampled''']) _lowerCamelCase : Optional[Any] = st.sidebar.slider( '''Minimum generation length''', min_value=8, max_value=256, value=64, step=8, format=None, key=None ) _lowerCamelCase : Optional[int] = st.sidebar.slider( '''Maximum generation length''', min_value=64, max_value=512, value=256, step=16, format=None, key=None ) if sampled == "beam": _lowerCamelCase : Optional[int] = st.sidebar.slider('''Beam size''', min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: _lowerCamelCase : Optional[Any] = st.sidebar.slider( '''Nucleus sampling p''', min_value=0.1, max_value=1.0, value=0.95, step=0.01, format=None, key=None ) _lowerCamelCase : Union[str, Any] = st.sidebar.slider( '''Temperature''', min_value=0.1, max_value=1.0, value=0.7, step=0.01, format=None, key=None ) _lowerCamelCase : Dict = None # start main text _lowerCamelCase : Union[str, Any] = [ """<MY QUESTION>""", """How do people make chocolate?""", """Why do we get a fever when we are sick?""", """How can different animals perceive different colors?""", """What is natural language processing?""", """What's the best way to treat a sunburn?""", """What exactly are vitamins ?""", """How does nuclear energy provide electricity?""", """What's the difference between viruses and bacteria?""", """Why are flutes classified as woodwinds when most of them are made out of metal ?""", """Why do people like drinking coffee even though it tastes so bad?""", """What happens when wine ages? How does it make the wine taste better?""", """If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?""", """How can we set a date to the beginning or end of an artistic period? Doesn't the change happen gradually?""", """How does New Zealand have so many large bird predators?""", ] _lowerCamelCase : Optional[Any] = st.selectbox( '''What would you like to ask? ---- select <MY QUESTION> to enter a new query''', questions_list, index=1, ) if question_s == "<MY QUESTION>": _lowerCamelCase : Dict = st.text_input('''Enter your question here:''', '''''') else: _lowerCamelCase : Union[str, Any] = question_s if st.button('''Show me!'''): if action in [0, 1, 3]: if index_type == "mixed": _lowerCamelCase : List[str] = make_support(question, source=wiki_source, method='''dense''', n_results=10) _lowerCamelCase : Any = make_support(question, source=wiki_source, method='''sparse''', n_results=10) _lowerCamelCase : Optional[Any] = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] _lowerCamelCase : Union[str, Any] = support_list[:10] _lowerCamelCase : Optional[Any] = """<P> """ + """ <P> """.join([res[-1] for res in support_list]) else: _lowerCamelCase : Optional[Any] = make_support(question, source=wiki_source, method=index_type, n_results=10) if action in [0, 3]: _lowerCamelCase : Any = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == '''sampled'''), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown('''### The model generated answer is:''') st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown('''--- \n ### The model is drawing information from the following Wikipedia passages:''') for i, res in enumerate(support_list): _lowerCamelCase : Tuple = """https://en.wikipedia.org/wiki/{}""".format(res[0].replace(''' ''', '''_''')) _lowerCamelCase : List[Any] = res[1].strip() if sec_titles == "": _lowerCamelCase : int = """[{}]({})""".format(res[0], wiki_url) else: _lowerCamelCase : str = sec_titles.split(''' & ''') _lowerCamelCase : int = """ & """.join( ['''[{}]({}#{})'''.format(sec.strip(), wiki_url, sec.strip().replace(''' ''', '''_''')) for sec in sec_list] ) st.markdown( '''{0:02d} - **Article**: {1:<18} <br> _Section_: {2}'''.format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( '''> <span style=\"font-family:arial; font-size:10pt;\">''' + res[-1] + '''</span>''', unsafe_allow_html=True ) if action in [2, 3]: _lowerCamelCase : List[Any] = find_nearest_training(question) _lowerCamelCase : int = nn_train_list[0] st.markdown( '''--- \n ### The most similar question in the ELI5 training set was: \n\n {}'''.format(train_exple['''title''']) ) _lowerCamelCase : Optional[int] = [ """{}. {}""".format(i + 1, ''' \n'''.join([line.strip() for line in ans.split('''\n''') if line.strip() != ''''''])) for i, (ans, sc) in enumerate(zip(train_exple['''answers''']['''text'''], train_exple['''answers''']['''score'''])) if i == 0 or sc > 2 ] st.markdown('''##### Its answers were: \n\n {}'''.format('''\n'''.join(answers_st))) _lowerCamelCase : Union[str, Any] = """ --- **Disclaimer** *The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system. Evaluating biases of such a model and ensuring factual generations are still very much open research problems. Therefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.* """ st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
184
import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class __magic_name__ : '''simple docstring''' def __init__( self:Optional[Any] , _a:int , _a:str=3 , _a:Optional[int]=32 , _a:Optional[Any]=3 , _a:Tuple=10 , _a:List[Any]=[8, 16, 32, 64] , _a:str=[1, 1, 2, 1] , _a:Any=True , _a:List[Any]=True , _a:List[str]="relu" , _a:int=3 , _a:Tuple=None , _a:Tuple=["stage2", "stage3", "stage4"] , _a:List[Any]=[2, 3, 4] , _a:Union[str, Any]=1 , ): snake_case__ = parent snake_case__ = batch_size snake_case__ = image_size snake_case__ = num_channels snake_case__ = embeddings_size snake_case__ = hidden_sizes snake_case__ = depths snake_case__ = is_training snake_case__ = use_labels snake_case__ = hidden_act snake_case__ = num_labels snake_case__ = scope snake_case__ = len(_a ) snake_case__ = out_features snake_case__ = out_indices snake_case__ = num_groups def SCREAMING_SNAKE_CASE__ ( self:int ): snake_case__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case__ = None if self.use_labels: snake_case__ = ids_tensor([self.batch_size] , self.num_labels ) snake_case__ = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def SCREAMING_SNAKE_CASE__ ( self:Any , _a:Optional[int] , _a:Tuple , _a:int ): snake_case__ = BitModel(config=_a ) model.to(_a ) model.eval() snake_case__ = model(_a ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def SCREAMING_SNAKE_CASE__ ( self:int , _a:Tuple , _a:Any , _a:Union[str, Any] ): snake_case__ = self.num_labels snake_case__ = BitForImageClassification(_a ) model.to(_a ) model.eval() snake_case__ = model(_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self:str , _a:str , _a:List[str] , _a:Any ): snake_case__ = BitBackbone(config=_a ) model.to(_a ) model.eval() snake_case__ = model(_a ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None snake_case__ = None snake_case__ = BitBackbone(config=_a ) model.to(_a ) model.eval() snake_case__ = model(_a ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = self.prepare_config_and_inputs() snake_case__ , snake_case__ , snake_case__ = config_and_inputs snake_case__ = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __magic_name__ (snake_case_ ,snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : Any = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () __lowercase : int = ( {'feature-extraction': BitModel, 'image-classification': BitForImageClassification} if is_torch_available() else {} ) __lowercase : Tuple = False __lowercase : Optional[Any] = False __lowercase : str = False __lowercase : Tuple = False __lowercase : Tuple = False def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): snake_case__ = BitModelTester(self ) snake_case__ = ConfigTester(self , config_class=_a , has_text_modality=_a ) def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): return @unittest.skip(reason='''Bit does not output attentions''' ) def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): pass @unittest.skip(reason='''Bit does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): pass @unittest.skip(reason='''Bit does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): pass def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ , snake_case__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case__ = model_class(_a ) snake_case__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case__ = [*signature.parameters.keys()] snake_case__ = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _a ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): snake_case__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_a ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ , snake_case__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case__ = model_class(config=_a ) for name, module in model.named_modules(): if isinstance(_a , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): def check_hidden_states_output(_a:List[Any] , _a:int , _a:Union[str, Any] ): snake_case__ = model_class(_a ) model.to(_a ) model.eval() with torch.no_grad(): snake_case__ = model(**self._prepare_for_class(_a , _a ) ) snake_case__ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states snake_case__ = self.model_tester.num_stages self.assertEqual(len(_a ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) snake_case__ , snake_case__ = self.model_tester.prepare_config_and_inputs_for_common() snake_case__ = ['''preactivation''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: snake_case__ = layer_type snake_case__ = True check_hidden_states_output(_a , _a , _a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case__ = True check_hidden_states_output(_a , _a , _a ) @unittest.skip(reason='''Bit does not use feedforward chunking''' ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): pass def SCREAMING_SNAKE_CASE__ ( self:List[str] ): snake_case__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_a ) @slow def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case__ = BitModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def SCREAMING_SNAKE_CASE ( ) -> Any: snake_case__ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class __magic_name__ (unittest.TestCase ): '''simple docstring''' @cached_property def SCREAMING_SNAKE_CASE__ ( self:Tuple ): return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): snake_case__ = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(_a ) snake_case__ = self.default_image_processor snake_case__ = prepare_img() snake_case__ = image_processor(images=_a , return_tensors='''pt''' ).to(_a ) # forward pass with torch.no_grad(): snake_case__ = model(**_a ) # verify the logits snake_case__ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , _a ) snake_case__ = torch.tensor([[-0.6526, -0.5263, -1.4398]] ).to(_a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _a , atol=1e-4 ) ) @require_torch class __magic_name__ (snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : Optional[Any] = (BitBackbone,) if is_torch_available() else () __lowercase : int = BitConfig __lowercase : Any = False def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = BitModelTester(self )
33
0
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE_ : str = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ : List[Any] = { """asapp/sew-tiny-100k""": """https://huggingface.co/asapp/sew-tiny-100k/resolve/main/config.json""", # See all SEW models at https://huggingface.co/models?filter=sew } class snake_case_ ( snake_case_ ): '''simple docstring''' __UpperCamelCase = 'sew' def __init__( self : Any , __lowerCamelCase : Union[str, Any]=32 , __lowerCamelCase : Optional[int]=768 , __lowerCamelCase : Optional[int]=12 , __lowerCamelCase : Any=12 , __lowerCamelCase : List[Any]=3_072 , __lowerCamelCase : List[str]=2 , __lowerCamelCase : int="gelu" , __lowerCamelCase : Any=0.1 , __lowerCamelCase : Tuple=0.1 , __lowerCamelCase : int=0.1 , __lowerCamelCase : int=0.0 , __lowerCamelCase : Any=0.1 , __lowerCamelCase : Tuple=0.1 , __lowerCamelCase : List[Any]=0.02 , __lowerCamelCase : List[str]=1E-5 , __lowerCamelCase : Union[str, Any]="group" , __lowerCamelCase : Optional[int]="gelu" , __lowerCamelCase : Optional[Any]=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , __lowerCamelCase : Optional[Any]=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , __lowerCamelCase : Optional[int]=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , __lowerCamelCase : Tuple=False , __lowerCamelCase : Any=128 , __lowerCamelCase : Optional[Any]=16 , __lowerCamelCase : str=True , __lowerCamelCase : Dict=0.05 , __lowerCamelCase : Tuple=10 , __lowerCamelCase : Optional[Any]=2 , __lowerCamelCase : str=0.0 , __lowerCamelCase : Union[str, Any]=10 , __lowerCamelCase : Optional[Any]=0 , __lowerCamelCase : Union[str, Any]="mean" , __lowerCamelCase : Tuple=False , __lowerCamelCase : List[str]=False , __lowerCamelCase : Optional[Any]=256 , __lowerCamelCase : Any=0 , __lowerCamelCase : Any=1 , __lowerCamelCase : List[str]=2 , **__lowerCamelCase : Any , ) -> List[str]: '''simple docstring''' super().__init__(**_a , pad_token_id=_a , bos_token_id=_a , eos_token_id=_a ) __lowercase = hidden_size __lowercase = feat_extract_norm __lowercase = feat_extract_activation __lowercase = list(_a ) __lowercase = list(_a ) __lowercase = list(_a ) __lowercase = conv_bias __lowercase = num_conv_pos_embeddings __lowercase = num_conv_pos_embedding_groups __lowercase = len(self.conv_dim ) __lowercase = num_hidden_layers __lowercase = intermediate_size __lowercase = squeeze_factor __lowercase = hidden_act __lowercase = num_attention_heads __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = feat_proj_dropout __lowercase = final_dropout __lowercase = layerdrop __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( 'Configuration for convolutional layers is incorrect.' 'It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,' F"but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)" F"= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`." ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowercase = apply_spec_augment __lowercase = mask_time_prob __lowercase = mask_time_length __lowercase = mask_time_min_masks __lowercase = mask_feature_prob __lowercase = mask_feature_length __lowercase = mask_feature_min_masks # ctc loss __lowercase = ctc_loss_reduction __lowercase = ctc_zero_infinity # sequence classification __lowercase = use_weighted_layer_sum __lowercase = classifier_proj_size @property def UpperCAmelCase ( self : Optional[int] ) -> Union[str, Any]: '''simple docstring''' return functools.reduce(operator.mul , self.conv_stride , 1 )
375
import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging lowerCamelCase__ : Any = """\ """ lowerCamelCase__ : List[str] = """ Perplexity (PPL) is one of the most common metrics for evaluating language models. It is defined as the exponentiated average negative log-likelihood of a sequence. For more information, see https://huggingface.co/docs/transformers/perplexity """ lowerCamelCase__ : Any = """ Args: model_id (str): model used for calculating Perplexity NOTE: Perplexity can only be calculated for causal language models. This includes models such as gpt2, causal variations of bert, causal versions of t5, and more (the full list can be found in the AutoModelForCausalLM documentation here: https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM ) input_texts (list of str): input text, each separate text snippet is one list entry. batch_size (int): the batch size to run texts through the model. Defaults to 16. add_start_token (bool): whether to add the start token to the texts, so the perplexity can include the probability of the first word. Defaults to True. device (str): device to run on, defaults to 'cuda' when available Returns: perplexity: dictionary containing the perplexity scores for the texts in the input list, as well as the mean perplexity. If one of the input texts is longer than the max input length of the model, then it is truncated to the max length for the perplexity computation. Examples: Example 1: >>> perplexity = datasets.load_metric(\"perplexity\") >>> input_texts = [\"lorem ipsum\", \"Happy Birthday!\", \"Bienvenue\"] >>> results = perplexity.compute(model_id='gpt2', ... add_start_token=False, ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) ['perplexities', 'mean_perplexity'] >>> print(round(results[\"mean_perplexity\"], 2)) 78.22 >>> print(round(results[\"perplexities\"][0], 2)) 11.11 Example 2: >>> perplexity = datasets.load_metric(\"perplexity\") >>> input_texts = datasets.load_dataset(\"wikitext\", ... \"wikitext-2-raw-v1\", ... split=\"test\")[\"text\"][:50] # doctest:+ELLIPSIS [...] >>> input_texts = [s for s in input_texts if s!=''] >>> results = perplexity.compute(model_id='gpt2', ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) ['perplexities', 'mean_perplexity'] >>> print(round(results[\"mean_perplexity\"], 2)) 60.35 >>> print(round(results[\"perplexities\"][0], 2)) 81.12 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class __magic_name__ (datasets.Metric ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self:int ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''input_texts''': datasets.Value('''string''' ), } ) , reference_urls=['''https://huggingface.co/docs/transformers/perplexity'''] , ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] , _a:int , _a:List[Any] , _a:int = 16 , _a:bool = True , _a:Any=None ): if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": snake_case__ = '''cuda''' else: snake_case__ = '''cuda''' if torch.cuda.is_available() else '''cpu''' snake_case__ = AutoModelForCausalLM.from_pretrained(_a ) snake_case__ = model.to(_a ) snake_case__ = AutoTokenizer.from_pretrained(_a ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: snake_case__ = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(_a ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({'''pad_token''': existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" snake_case__ = model.config.max_length - 1 else: snake_case__ = model.config.max_length snake_case__ = tokenizer( _a , add_special_tokens=_a , padding=_a , truncation=_a , max_length=_a , return_tensors='''pt''' , return_attention_mask=_a , ).to(_a ) snake_case__ = encodings['''input_ids'''] snake_case__ = encodings['''attention_mask'''] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." snake_case__ = [] snake_case__ = CrossEntropyLoss(reduction='''none''' ) for start_index in logging.tqdm(range(0 , len(_a ) , _a ) ): snake_case__ = min(start_index + batch_size , len(_a ) ) snake_case__ = encoded_texts[start_index:end_index] snake_case__ = attn_masks[start_index:end_index] if add_start_token: snake_case__ = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(_a ) snake_case__ = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) snake_case__ = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(_a ), attn_mask] , dim=1 ) snake_case__ = encoded_batch with torch.no_grad(): snake_case__ = model(_a , attention_mask=_a ).logits snake_case__ = out_logits[..., :-1, :].contiguous() snake_case__ = labels[..., 1:].contiguous() snake_case__ = attn_mask[..., 1:].contiguous() snake_case__ = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , _a ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(_a )}
33
0
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class UpperCAmelCase_ ( snake_case_ , snake_case_ , unittest.TestCase ): '''simple docstring''' A : Union[str, Any] = IFInpaintingSuperResolutionPipeline A : Optional[int] = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'width', 'height'} A : List[str] = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({'original_image'} ) A : Union[str, Any] = PipelineTesterMixin.required_optional_params - {'latents'} def _lowerCAmelCase ( self ) -> int: return self._get_superresolution_dummy_components() def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=0 ) -> int: if str(_a ).startswith("mps" ): snake_case_ : Tuple = torch.manual_seed(_a ) else: snake_case_ : List[str] = torch.Generator(device=_a ).manual_seed(_a ) snake_case_ : str = floats_tensor((1, 3, 16, 16) , rng=random.Random(_a ) ).to(_a ) snake_case_ : str = floats_tensor((1, 3, 32, 32) , rng=random.Random(_a ) ).to(_a ) snake_case_ : str = floats_tensor((1, 3, 32, 32) , rng=random.Random(_a ) ).to(_a ) snake_case_ : Optional[int] = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def _lowerCAmelCase ( self ) -> str: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def _lowerCAmelCase ( self ) -> Dict: self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def _lowerCAmelCase ( self ) -> Union[str, Any]: # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1 ) def _lowerCAmelCase ( self ) -> int: self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def _lowerCAmelCase ( self ) -> Optional[Any]: self._test_save_load_local() def _lowerCAmelCase ( self ) -> str: self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
568
import os from datetime import datetime as dt from github import Github lowerCamelCase__ : int = [ """good first issue""", """good second issue""", """good difficult issue""", """enhancement""", """new pipeline/model""", """new scheduler""", """wip""", ] def SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: snake_case__ = Github(os.environ['''GITHUB_TOKEN'''] ) snake_case__ = g.get_repo('''huggingface/diffusers''' ) snake_case__ = repo.get_issues(state='''open''' ) for issue in open_issues: snake_case__ = sorted(issue.get_comments() , key=lambda __lowerCAmelCase : i.created_at , reverse=__lowerCAmelCase ) snake_case__ = comments[0] if len(__lowerCAmelCase ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Closes the issue after 7 days of inactivity since the Stalebot notification. issue.edit(state='''closed''' ) elif ( "stale" in issue.get_labels() and last_comment is not None and last_comment.user.login != "github-actions[bot]" ): # Opens the issue if someone other than Stalebot commented. issue.edit(state='''open''' ) issue.remove_from_labels('''stale''' ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Post a Stalebot notification after 23 days of inactivity. issue.create_comment( '''This issue has been automatically marked as stale because it has not had ''' '''recent activity. If you think this still needs to be addressed ''' '''please comment on this thread.\n\nPlease note that issues that do not follow the ''' '''[contributing guidelines](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md) ''' '''are likely to be ignored.''' ) issue.add_to_labels('''stale''' ) if __name__ == "__main__": main()
33
0
import unittest from huggingface_hub import hf_hub_download from transformers import MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING, VideoMAEFeatureExtractor from transformers.pipelines import VideoClassificationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_decord, require_tf, require_torch, require_torch_or_tf, require_vision, ) from .test_pipelines_common import ANY @is_pipeline_test @require_torch_or_tf @require_vision @require_decord class lowerCamelCase_ ( unittest.TestCase ): __lowercase : Tuple = MODEL_FOR_VIDEO_CLASSIFICATION_MAPPING def lowercase ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ ) -> str: """simple docstring""" _UpperCamelCase = hf_hub_download( repo_id="nateraw/video-demo" , filename="archery.mp4" , repo_type="dataset" ) _UpperCamelCase = VideoClassificationPipeline(model=_a , image_processor=_a , top_k=2 ) _UpperCamelCase = [ example_video_filepath, "https://huggingface.co/datasets/nateraw/video-demo/resolve/main/archery.mp4", ] return video_classifier, examples def lowercase ( self , lowerCamelCase_ , lowerCamelCase_ ) -> Tuple: """simple docstring""" for example in examples: _UpperCamelCase = video_classifier(_a ) self.assertEqual( _a , [ {"score": ANY(_a ), "label": ANY(_a )}, {"score": ANY(_a ), "label": ANY(_a )}, ] , ) @require_torch def lowercase ( self ) -> Union[str, Any]: """simple docstring""" _UpperCamelCase = "hf-internal-testing/tiny-random-VideoMAEForVideoClassification" _UpperCamelCase = VideoMAEFeatureExtractor( size={"shortest_edge": 10} , crop_size={"height": 10, "width": 10} ) _UpperCamelCase = pipeline( "video-classification" , model=_a , feature_extractor=_a , frame_sampling_rate=4 ) _UpperCamelCase = hf_hub_download(repo_id="nateraw/video-demo" , filename="archery.mp4" , repo_type="dataset" ) _UpperCamelCase = video_classifier(_a , top_k=2 ) self.assertEqual( nested_simplify(_a , decimals=4 ) , [{"score": 0.51_99, "label": "LABEL_0"}, {"score": 0.48_01, "label": "LABEL_1"}] , ) _UpperCamelCase = video_classifier( [ video_file_path, video_file_path, ] , top_k=2 , ) self.assertEqual( nested_simplify(_a , decimals=4 ) , [ [{"score": 0.51_99, "label": "LABEL_0"}, {"score": 0.48_01, "label": "LABEL_1"}], [{"score": 0.51_99, "label": "LABEL_0"}, {"score": 0.48_01, "label": "LABEL_1"}], ] , ) @require_tf def lowercase ( self ) -> Union[str, Any]: """simple docstring""" pass
147
import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( '''kwargs, expected''' , [ ({'''num_shards''': 0, '''max_num_jobs''': 1}, []), ({'''num_shards''': 10, '''max_num_jobs''': 1}, [range(10 )]), ({'''num_shards''': 10, '''max_num_jobs''': 10}, [range(__lowerCAmelCase , i + 1 ) for i in range(10 )]), ({'''num_shards''': 1, '''max_num_jobs''': 10}, [range(1 )]), ({'''num_shards''': 10, '''max_num_jobs''': 3}, [range(0 , 4 ), range(4 , 7 ), range(7 , 10 )]), ({'''num_shards''': 3, '''max_num_jobs''': 10}, [range(0 , 1 ), range(1 , 2 ), range(2 , 3 )]), ] , ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> Optional[int]: snake_case__ = _distribute_shards(**__lowerCAmelCase ) assert out == expected @pytest.mark.parametrize( '''gen_kwargs, max_num_jobs, expected''' , [ ({'''foo''': 0}, 10, [{'''foo''': 0}]), ({'''shards''': [0, 1, 2, 3]}, 1, [{'''shards''': [0, 1, 2, 3]}]), ({'''shards''': [0, 1, 2, 3]}, 4, [{'''shards''': [0]}, {'''shards''': [1]}, {'''shards''': [2]}, {'''shards''': [3]}]), ({'''shards''': [0, 1]}, 4, [{'''shards''': [0]}, {'''shards''': [1]}]), ({'''shards''': [0, 1, 2, 3]}, 2, [{'''shards''': [0, 1]}, {'''shards''': [2, 3]}]), ] , ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Dict: snake_case__ = _split_gen_kwargs(__lowerCAmelCase , __lowerCAmelCase ) assert out == expected @pytest.mark.parametrize( '''gen_kwargs, expected''' , [ ({'''foo''': 0}, 1), ({'''shards''': [0]}, 1), ({'''shards''': [0, 1, 2, 3]}, 4), ({'''shards''': [0, 1, 2, 3], '''foo''': 0}, 4), ({'''shards''': [0, 1, 2, 3], '''other''': (0, 1)}, 4), ({'''shards''': [0, 1, 2, 3], '''shards2''': [0, 1]}, RuntimeError), ] , ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> List[Any]: if expected is RuntimeError: with pytest.raises(__lowerCAmelCase ): _number_of_shards_in_gen_kwargs(__lowerCAmelCase ) else: snake_case__ = _number_of_shards_in_gen_kwargs(__lowerCAmelCase ) assert out == expected
33
0
import argparse import json import os import pickle import shutil import numpy as np import torch from distiller import Distiller from lm_seqs_dataset import LmSeqsDataset from transformers import ( BertConfig, BertForMaskedLM, BertTokenizer, DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer, GPTaConfig, GPTaLMHeadModel, GPTaTokenizer, RobertaConfig, RobertaForMaskedLM, RobertaTokenizer, ) from utils import git_log, init_gpu_params, logger, set_seed _A = { """distilbert""": (DistilBertConfig, DistilBertForMaskedLM, DistilBertTokenizer), """roberta""": (RobertaConfig, RobertaForMaskedLM, RobertaTokenizer), """bert""": (BertConfig, BertForMaskedLM, BertTokenizer), """gpt2""": (GPTaConfig, GPTaLMHeadModel, GPTaTokenizer), } def __UpperCamelCase ( _A ): assert (args.mlm and args.alpha_mlm > 0.0) or (not args.mlm and args.alpha_mlm == 0.0) assert (args.alpha_mlm > 0.0 and args.alpha_clm == 0.0) or (args.alpha_mlm == 0.0 and args.alpha_clm > 0.0) if args.mlm: assert os.path.isfile(args.token_counts ) assert (args.student_type in ["roberta", "distilbert"]) and (args.teacher_type in ["roberta", "bert"]) else: assert (args.student_type in ["gpt2"]) and (args.teacher_type in ["gpt2"]) assert args.teacher_type == args.student_type or ( args.student_type == "distilbert" and args.teacher_type == "bert" ) assert os.path.isfile(args.student_config ) if args.student_pretrained_weights is not None: assert os.path.isfile(args.student_pretrained_weights ) if args.freeze_token_type_embds: assert args.student_type in ["roberta"] assert args.alpha_ce >= 0.0 assert args.alpha_mlm >= 0.0 assert args.alpha_clm >= 0.0 assert args.alpha_mse >= 0.0 assert args.alpha_cos >= 0.0 assert args.alpha_ce + args.alpha_mlm + args.alpha_clm + args.alpha_mse + args.alpha_cos > 0.0 def __UpperCamelCase ( _A , _A ): if args.student_type == "roberta": lowerCAmelCase_ = False elif args.student_type == "gpt2": lowerCAmelCase_ = False def __UpperCamelCase ( _A , _A ): if args.student_type == "roberta": lowerCAmelCase_ = False def __UpperCamelCase ( ): lowerCAmelCase_ = argparse.ArgumentParser(description='''Training''' ) parser.add_argument('''--force''' , action='''store_true''' , help='''Overwrite dump_path if it already exists.''' ) parser.add_argument( '''--dump_path''' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='''The output directory (log, checkpoints, parameters, etc.)''' ) parser.add_argument( '''--data_file''' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='''The binarized file (tokenized + tokens_to_ids) and grouped by sequence.''' , ) parser.add_argument( '''--student_type''' , type=__lowerCAmelCase , choices=['''distilbert''', '''roberta''', '''gpt2'''] , required=__lowerCAmelCase , help='''The student type (DistilBERT, RoBERTa).''' , ) parser.add_argument('''--student_config''' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='''Path to the student configuration.''' ) parser.add_argument( '''--student_pretrained_weights''' , default=__lowerCAmelCase , type=__lowerCAmelCase , help='''Load student initialization checkpoint.''' ) parser.add_argument( '''--teacher_type''' , choices=['''bert''', '''roberta''', '''gpt2'''] , required=__lowerCAmelCase , help='''Teacher type (BERT, RoBERTa).''' ) parser.add_argument('''--teacher_name''' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='''The teacher model.''' ) parser.add_argument('''--temperature''' , default=2.0 , type=__lowerCAmelCase , help='''Temperature for the softmax temperature.''' ) parser.add_argument( '''--alpha_ce''' , default=0.5 , type=__lowerCAmelCase , help='''Linear weight for the distillation loss. Must be >=0.''' ) parser.add_argument( '''--alpha_mlm''' , default=0.0 , type=__lowerCAmelCase , help='''Linear weight for the MLM loss. Must be >=0. Should be used in conjunction with `mlm` flag.''' , ) parser.add_argument('''--alpha_clm''' , default=0.5 , type=__lowerCAmelCase , help='''Linear weight for the CLM loss. Must be >=0.''' ) parser.add_argument('''--alpha_mse''' , default=0.0 , type=__lowerCAmelCase , help='''Linear weight of the MSE loss. Must be >=0.''' ) parser.add_argument( '''--alpha_cos''' , default=0.0 , type=__lowerCAmelCase , help='''Linear weight of the cosine embedding loss. Must be >=0.''' ) parser.add_argument( '''--mlm''' , action='''store_true''' , help='''The LM step: MLM or CLM. If `mlm` is True, the MLM is used over CLM.''' ) parser.add_argument( '''--mlm_mask_prop''' , default=0.1_5 , type=__lowerCAmelCase , help='''Proportion of tokens for which we need to make a prediction.''' , ) parser.add_argument('''--word_mask''' , default=0.8 , type=__lowerCAmelCase , help='''Proportion of tokens to mask out.''' ) parser.add_argument('''--word_keep''' , default=0.1 , type=__lowerCAmelCase , help='''Proportion of tokens to keep.''' ) parser.add_argument('''--word_rand''' , default=0.1 , type=__lowerCAmelCase , help='''Proportion of tokens to randomly replace.''' ) parser.add_argument( '''--mlm_smoothing''' , default=0.7 , type=__lowerCAmelCase , help='''Smoothing parameter to emphasize more rare tokens (see XLM, similar to word2vec).''' , ) parser.add_argument('''--token_counts''' , type=__lowerCAmelCase , help='''The token counts in the data_file for MLM.''' ) parser.add_argument( '''--restrict_ce_to_mask''' , action='''store_true''' , help='''If true, compute the distillation loss only the [MLM] prediction distribution.''' , ) parser.add_argument( '''--freeze_pos_embs''' , action='''store_true''' , help='''Freeze positional embeddings during distillation. For student_type in [\'roberta\', \'gpt2\'] only.''' , ) parser.add_argument( '''--freeze_token_type_embds''' , action='''store_true''' , help='''Freeze token type embeddings during distillation if existent. For student_type in [\'roberta\'] only.''' , ) parser.add_argument('''--n_epoch''' , type=__lowerCAmelCase , default=3 , help='''Number of pass on the whole dataset.''' ) parser.add_argument('''--batch_size''' , type=__lowerCAmelCase , default=5 , help='''Batch size (for each process).''' ) parser.add_argument( '''--group_by_size''' , action='''store_false''' , help='''If true, group sequences that have similar length into the same batch. Default is true.''' , ) parser.add_argument( '''--gradient_accumulation_steps''' , type=__lowerCAmelCase , default=50 , help='''Gradient accumulation for larger training batches.''' , ) parser.add_argument('''--warmup_prop''' , default=0.0_5 , type=__lowerCAmelCase , help='''Linear warmup proportion.''' ) parser.add_argument('''--weight_decay''' , default=0.0 , type=__lowerCAmelCase , help='''Weight decay if we apply some.''' ) parser.add_argument('''--learning_rate''' , default=5E-4 , type=__lowerCAmelCase , help='''The initial learning rate for Adam.''' ) parser.add_argument('''--adam_epsilon''' , default=1E-6 , type=__lowerCAmelCase , help='''Epsilon for Adam optimizer.''' ) parser.add_argument('''--max_grad_norm''' , default=5.0 , type=__lowerCAmelCase , help='''Max gradient norm.''' ) parser.add_argument('''--initializer_range''' , default=0.0_2 , type=__lowerCAmelCase , help='''Random initialization range.''' ) parser.add_argument( '''--fp16''' , action='''store_true''' , help='''Whether to use 16-bit (mixed) precision (through NVIDIA apex) instead of 32-bit''' , ) parser.add_argument( '''--fp16_opt_level''' , type=__lowerCAmelCase , default='''O1''' , help=( '''For fp16: Apex AMP optimization level selected in [\'O0\', \'O1\', \'O2\', and \'O3\'].''' '''See details at https://nvidia.github.io/apex/amp.html''' ) , ) parser.add_argument('''--n_gpu''' , type=__lowerCAmelCase , default=1 , help='''Number of GPUs in the node.''' ) parser.add_argument('''--local_rank''' , type=__lowerCAmelCase , default=-1 , help='''Distributed training - Local rank''' ) parser.add_argument('''--seed''' , type=__lowerCAmelCase , default=56 , help='''Random seed''' ) parser.add_argument('''--log_interval''' , type=__lowerCAmelCase , default=500 , help='''Tensorboard logging interval.''' ) parser.add_argument('''--checkpoint_interval''' , type=__lowerCAmelCase , default=4000 , help='''Checkpoint interval.''' ) lowerCAmelCase_ = parser.parse_args() sanity_checks(__lowerCAmelCase ) # ARGS # init_gpu_params(__lowerCAmelCase ) set_seed(__lowerCAmelCase ) if args.is_master: if os.path.exists(args.dump_path ): if not args.force: raise ValueError( f"Serialization dir {args.dump_path} already exists, but you have not precised wheter to overwrite" ''' itUse `--force` if you want to overwrite it''' ) else: shutil.rmtree(args.dump_path ) if not os.path.exists(args.dump_path ): os.makedirs(args.dump_path ) logger.info(f"Experiment will be dumped and logged in {args.dump_path}" ) # SAVE PARAMS # logger.info(f"Param: {args}" ) with open(os.path.join(args.dump_path , '''parameters.json''' ) , '''w''' ) as f: json.dump(vars(__lowerCAmelCase ) , __lowerCAmelCase , indent=4 ) git_log(args.dump_path ) lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = MODEL_CLASSES[args.student_type] lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ = MODEL_CLASSES[args.teacher_type] # TOKENIZER # lowerCAmelCase_ = teacher_tokenizer_class.from_pretrained(args.teacher_name ) lowerCAmelCase_ = {} for tok_name, tok_symbol in tokenizer.special_tokens_map.items(): lowerCAmelCase_ = tokenizer.all_special_tokens.index(__lowerCAmelCase ) lowerCAmelCase_ = tokenizer.all_special_ids[idx] logger.info(f"Special tokens {special_tok_ids}" ) lowerCAmelCase_ = special_tok_ids lowerCAmelCase_ = tokenizer.max_model_input_sizes[args.teacher_name] # DATA LOADER # logger.info(f"Loading data from {args.data_file}" ) with open(args.data_file , '''rb''' ) as fp: lowerCAmelCase_ = pickle.load(__lowerCAmelCase ) if args.mlm: logger.info(f"Loading token counts from {args.token_counts} (already pre-computed)" ) with open(args.token_counts , '''rb''' ) as fp: lowerCAmelCase_ = pickle.load(__lowerCAmelCase ) lowerCAmelCase_ = np.maximum(__lowerCAmelCase , 1 ) ** -args.mlm_smoothing for idx in special_tok_ids.values(): lowerCAmelCase_ = 0.0 # do not predict special tokens lowerCAmelCase_ = torch.from_numpy(__lowerCAmelCase ) else: lowerCAmelCase_ = None lowerCAmelCase_ = LmSeqsDataset(params=__lowerCAmelCase , data=__lowerCAmelCase ) logger.info('''Data loader created.''' ) # STUDENT # logger.info(f"Loading student config from {args.student_config}" ) lowerCAmelCase_ = student_config_class.from_pretrained(args.student_config ) lowerCAmelCase_ = True if args.student_pretrained_weights is not None: logger.info(f"Loading pretrained weights from {args.student_pretrained_weights}" ) lowerCAmelCase_ = student_model_class.from_pretrained(args.student_pretrained_weights , config=__lowerCAmelCase ) else: lowerCAmelCase_ = student_model_class(__lowerCAmelCase ) if args.n_gpu > 0: student.to(f"cuda:{args.local_rank}" ) logger.info('''Student loaded.''' ) # TEACHER # lowerCAmelCase_ = teacher_model_class.from_pretrained(args.teacher_name , output_hidden_states=__lowerCAmelCase ) if args.n_gpu > 0: teacher.to(f"cuda:{args.local_rank}" ) logger.info(f"Teacher loaded from {args.teacher_name}." ) # FREEZING # if args.freeze_pos_embs: freeze_pos_embeddings(__lowerCAmelCase , __lowerCAmelCase ) if args.freeze_token_type_embds: freeze_token_type_embeddings(__lowerCAmelCase , __lowerCAmelCase ) # SANITY CHECKS # assert student.config.vocab_size == teacher.config.vocab_size assert student.config.hidden_size == teacher.config.hidden_size assert student.config.max_position_embeddings == teacher.config.max_position_embeddings if args.mlm: assert token_probs.size(0 ) == stu_architecture_config.vocab_size # DISTILLER # torch.cuda.empty_cache() lowerCAmelCase_ = Distiller( params=__lowerCAmelCase , dataset=__lowerCAmelCase , token_probs=__lowerCAmelCase , student=__lowerCAmelCase , teacher=__lowerCAmelCase ) distiller.train() logger.info('''Let\'s go get some drinks.''' ) if __name__ == "__main__": main()
431
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __magic_name__ (snake_case_ ,snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : str = IFImgaImgSuperResolutionPipeline __lowercase : Optional[Any] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'width', 'height'} __lowercase : int = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'original_image'} ) __lowercase : List[str] = PipelineTesterMixin.required_optional_params - {'latents'} def SCREAMING_SNAKE_CASE__ ( self:Dict ): return self._get_superresolution_dummy_components() def SCREAMING_SNAKE_CASE__ ( self:Dict , _a:int , _a:Optional[Any]=0 ): if str(_a ).startswith('''mps''' ): snake_case__ = torch.manual_seed(_a ) else: snake_case__ = torch.Generator(device=_a ).manual_seed(_a ) snake_case__ = floats_tensor((1, 3, 32, 32) , rng=random.Random(_a ) ).to(_a ) snake_case__ = floats_tensor((1, 3, 16, 16) , rng=random.Random(_a ) ).to(_a ) snake_case__ = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1 ) def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def SCREAMING_SNAKE_CASE__ ( self:str ): self._test_save_load_local() def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
33
0
A__ : List[str] = """Alexander Joslin""" import operator as op from .stack import Stack def UpperCamelCase( __UpperCamelCase : int ): lowerCAmelCase_ : List[Any] = {'''*''': op.mul, '''/''': op.truediv, '''+''': op.add, '''-''': op.sub} lowerCAmelCase_ : Dict = Stack() lowerCAmelCase_ : Optional[Any] = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(__lowerCAmelCase ) ) elif i in operators: # RULE 2 operator_stack.push(__lowerCAmelCase ) elif i == ")": # RULE 4 lowerCAmelCase_ : List[Any] = operator_stack.peek() operator_stack.pop() lowerCAmelCase_ : List[str] = operand_stack.peek() operand_stack.pop() lowerCAmelCase_ : Optional[Any] = operand_stack.peek() operand_stack.pop() lowerCAmelCase_ : int = operators[opr](__lowerCAmelCase ,__lowerCAmelCase ) operand_stack.push(__lowerCAmelCase ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": A__ : Optional[Any] = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F'''{equation} = {dijkstras_two_stack_algorithm(equation)}''')
171
import math class __magic_name__ : '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self:Optional[int] , _a:list[list[float]] , _a:list[int] ): snake_case__ = 0.0 snake_case__ = 0.0 for i in range(len(_a ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def SCREAMING_SNAKE_CASE__ ( self:Tuple , _a:list[list[int | float]] , _a:list[int] , _a:int , _a:float ): for i in range(len(_a ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def SCREAMING_SNAKE_CASE ( ) -> None: # Training Examples ( m, n ) snake_case__ = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) snake_case__ = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training snake_case__ = SelfOrganizingMap() snake_case__ = 3 snake_case__ = 0.5 for _ in range(__lowerCAmelCase ): for j in range(len(__lowerCAmelCase ) ): # training sample snake_case__ = training_samples[j] # Compute the winning vector snake_case__ = self_organizing_map.get_winner(__lowerCAmelCase , __lowerCAmelCase ) # Update the winning vector snake_case__ = self_organizing_map.update(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # classify test sample snake_case__ = [0, 0, 0, 1] snake_case__ = self_organizing_map.get_winner(__lowerCAmelCase , __lowerCAmelCase ) # results print(F"""Clusters that the test sample belongs to : {winner}""" ) print(F"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
33
0
import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor _lowercase = logging.get_logger(__name__) class lowercase_ ( snake_case_ ): def __init__( self , *__A , **__A ) -> Any: warnings.warn( '''The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please''' ''' use GLPNImageProcessor instead.''' , _a , ) super().__init__(*_a , **_a )
443
from __future__ import annotations from statistics import mean def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> list[int]: snake_case__ = [0] * no_of_processes snake_case__ = [0] * no_of_processes # Initialize remaining_time to waiting_time. for i in range(__lowerCAmelCase ): snake_case__ = burst_time[i] snake_case__ = [] snake_case__ = 0 snake_case__ = 0 # When processes are not completed, # A process whose arrival time has passed \ # and has remaining execution time is put into the ready_process. # The shortest process in the ready_process, target_process is executed. while completed != no_of_processes: snake_case__ = [] snake_case__ = -1 for i in range(__lowerCAmelCase ): if (arrival_time[i] <= total_time) and (remaining_time[i] > 0): ready_process.append(__lowerCAmelCase ) if len(__lowerCAmelCase ) > 0: snake_case__ = ready_process[0] for i in ready_process: if remaining_time[i] < remaining_time[target_process]: snake_case__ = i total_time += burst_time[target_process] completed += 1 snake_case__ = 0 snake_case__ = ( total_time - arrival_time[target_process] - burst_time[target_process] ) else: total_time += 1 return waiting_time def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> list[int]: snake_case__ = [0] * no_of_processes for i in range(__lowerCAmelCase ): snake_case__ = burst_time[i] + waiting_time[i] return turn_around_time if __name__ == "__main__": print("""[TEST CASE 01]""") lowerCamelCase__ : Tuple = 4 lowerCamelCase__ : Union[str, Any] = [2, 5, 3, 7] lowerCamelCase__ : Optional[Any] = [0, 0, 0, 0] lowerCamelCase__ : Dict = calculate_waitingtime(arrival_time, burst_time, no_of_processes) lowerCamelCase__ : Union[str, Any] = calculate_turnaroundtime( burst_time, no_of_processes, waiting_time ) # Printing the Result print("""PID\tBurst Time\tArrival Time\tWaiting Time\tTurnaround Time""") for i, process_id in enumerate(list(range(1, 5))): print( F"""{process_id}\t{burst_time[i]}\t\t\t{arrival_time[i]}\t\t\t\t""" F"""{waiting_time[i]}\t\t\t\t{turn_around_time[i]}""" ) print(F"""\nAverage waiting time = {mean(waiting_time):.5f}""") print(F"""Average turnaround time = {mean(turn_around_time):.5f}""")
33
0
"""simple docstring""" import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class snake_case ( unittest.TestCase ): def __init__(self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=7 , SCREAMING_SNAKE_CASE_=3 , SCREAMING_SNAKE_CASE_=18 , SCREAMING_SNAKE_CASE_=30 , SCREAMING_SNAKE_CASE_=4_00 , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=True , ): """simple docstring""" SCREAMING_SNAKE_CASE_ = size if size is not None else {'''height''': 18, '''width''': 18} SCREAMING_SNAKE_CASE_ = parent SCREAMING_SNAKE_CASE_ = batch_size SCREAMING_SNAKE_CASE_ = num_channels SCREAMING_SNAKE_CASE_ = image_size SCREAMING_SNAKE_CASE_ = min_resolution SCREAMING_SNAKE_CASE_ = max_resolution SCREAMING_SNAKE_CASE_ = do_resize SCREAMING_SNAKE_CASE_ = size SCREAMING_SNAKE_CASE_ = do_normalize def _lowercase (self ): """simple docstring""" return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.88_66_44_36_34_03_32_03, 0.66_18_82_93_69_54_49_83, 0.38_91_74_64_01_78_68_04], [-0.60_42_55_91_46_88_11_04, -0.0_22_95_00_88_60_52_84_69, 0.54_23_79_73_69_00_32_96], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class snake_case ( snake_case_ , unittest.TestCase ): UpperCAmelCase__ = ImageGPTImageProcessor if is_vision_available() else None def _lowercase (self ): """simple docstring""" SCREAMING_SNAKE_CASE_ = ImageGPTImageProcessingTester(self ) @property def _lowercase (self ): """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def _lowercase (self ): """simple docstring""" SCREAMING_SNAKE_CASE_ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(_a , '''clusters''' ) ) self.assertTrue(hasattr(_a , '''do_resize''' ) ) self.assertTrue(hasattr(_a , '''size''' ) ) self.assertTrue(hasattr(_a , '''do_normalize''' ) ) def _lowercase (self ): """simple docstring""" SCREAMING_SNAKE_CASE_ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'''height''': 18, '''width''': 18} ) SCREAMING_SNAKE_CASE_ = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {'''height''': 42, '''width''': 42} ) def _lowercase (self ): """simple docstring""" SCREAMING_SNAKE_CASE_ = self.image_processing_class(**self.image_processor_dict ) SCREAMING_SNAKE_CASE_ = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(_a , obj[key] ) ) else: self.assertEqual(obj[key] , _a ) def _lowercase (self ): """simple docstring""" SCREAMING_SNAKE_CASE_ = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: SCREAMING_SNAKE_CASE_ = os.path.join(_a , '''image_processor.json''' ) image_processor_first.to_json_file(_a ) SCREAMING_SNAKE_CASE_ = self.image_processing_class.from_json_file(_a ).to_dict() SCREAMING_SNAKE_CASE_ = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(_a , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , _a ) def _lowercase (self ): """simple docstring""" SCREAMING_SNAKE_CASE_ = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(_a ) SCREAMING_SNAKE_CASE_ = self.image_processing_class.from_pretrained(_a ).to_dict() SCREAMING_SNAKE_CASE_ = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(_a , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , _a ) @unittest.skip('''ImageGPT requires clusters at initialization''' ) def _lowercase (self ): """simple docstring""" pass def _lowerCamelCase ( ): SCREAMING_SNAKE_CASE_ = load_dataset('''hf-internal-testing/fixtures_image_utils''', split='''test''' ) SCREAMING_SNAKE_CASE_ = Image.open(dataset[4]['''file'''] ) SCREAMING_SNAKE_CASE_ = Image.open(dataset[5]['''file'''] ) SCREAMING_SNAKE_CASE_ = [imagea, imagea] return images @require_vision @require_torch class snake_case ( unittest.TestCase ): @slow def _lowercase (self ): """simple docstring""" SCREAMING_SNAKE_CASE_ = ImageGPTImageProcessor.from_pretrained('''openai/imagegpt-small''' ) SCREAMING_SNAKE_CASE_ = prepare_images() # test non-batched SCREAMING_SNAKE_CASE_ = image_processing(images[0] , return_tensors='''pt''' ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (1, 10_24) ) SCREAMING_SNAKE_CASE_ = [3_06, 1_91, 1_91] self.assertEqual(encoding.input_ids[0, :3].tolist() , _a ) # test batched SCREAMING_SNAKE_CASE_ = image_processing(_a , return_tensors='''pt''' ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (2, 10_24) ) SCREAMING_SNAKE_CASE_ = [3_03, 13, 13] self.assertEqual(encoding.input_ids[1, -3:].tolist() , _a )
626
lowerCamelCase__ : List[str] = """Alexander Joslin""" import operator as op from .stack import Stack def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> int: snake_case__ = {'''*''': op.mul, '''/''': op.truediv, '''+''': op.add, '''-''': op.sub} snake_case__ = Stack() snake_case__ = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(__lowerCAmelCase ) ) elif i in operators: # RULE 2 operator_stack.push(__lowerCAmelCase ) elif i == ")": # RULE 4 snake_case__ = operator_stack.peek() operator_stack.pop() snake_case__ = operand_stack.peek() operand_stack.pop() snake_case__ = operand_stack.peek() operand_stack.pop() snake_case__ = operators[opr](__lowerCAmelCase , __lowerCAmelCase ) operand_stack.push(__lowerCAmelCase ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ : Optional[Any] = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
33
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) A_ = { """configuration_deberta""": ["""DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """DebertaConfig""", """DebertaOnnxConfig"""], """tokenization_deberta""": ["""DebertaTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ["""DebertaTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ """DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """DebertaForMaskedLM""", """DebertaForQuestionAnswering""", """DebertaForSequenceClassification""", """DebertaForTokenClassification""", """DebertaModel""", """DebertaPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ """TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFDebertaForMaskedLM""", """TFDebertaForQuestionAnswering""", """TFDebertaForSequenceClassification""", """TFDebertaForTokenClassification""", """TFDebertaModel""", """TFDebertaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_deberta import DEBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, DebertaConfig, DebertaOnnxConfig from .tokenization_deberta import DebertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_deberta_fast import DebertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_deberta import ( DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, DebertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_deberta import ( TF_DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFDebertaForMaskedLM, TFDebertaForQuestionAnswering, TFDebertaForSequenceClassification, TFDebertaForTokenClassification, TFDebertaModel, TFDebertaPreTrainedModel, ) else: import sys A_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
29
import warnings from ...utils import logging from .image_processing_perceiver import PerceiverImageProcessor lowerCamelCase__ : int = logging.get_logger(__name__) class __magic_name__ (snake_case_ ): '''simple docstring''' def __init__( self:List[Any] , *_a:Dict , **_a:Tuple ): warnings.warn( '''The class PerceiverFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use PerceiverImageProcessor instead.''' , _a , ) super().__init__(*_a , **_a )
33
0
"""simple docstring""" _a = """0.18.2""" from .configuration_utils import ConfigMixin from .utils import ( OptionalDependencyNotAvailable, is_flax_available, is_inflect_available, is_invisible_watermark_available, is_k_diffusion_available, is_k_diffusion_version, is_librosa_available, is_note_seq_available, is_onnx_available, is_scipy_available, is_torch_available, is_torchsde_available, is_transformers_available, is_transformers_version, is_unidecode_available, logging, ) try: if not is_onnx_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_onnx_objects import * # noqa F403 else: from .pipelines import OnnxRuntimeModel try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_pt_objects import * # noqa F403 else: from .models import ( AutoencoderKL, ControlNetModel, ModelMixin, PriorTransformer, TaFilmDecoder, TransformeraDModel, UNetaDModel, UNetaDConditionModel, UNetaDModel, UNetaDConditionModel, VQModel, ) from .optimization import ( get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, get_scheduler, ) from .pipelines import ( AudioPipelineOutput, ConsistencyModelPipeline, DanceDiffusionPipeline, DDIMPipeline, DDPMPipeline, DiffusionPipeline, DiTPipeline, ImagePipelineOutput, KarrasVePipeline, LDMPipeline, LDMSuperResolutionPipeline, PNDMPipeline, RePaintPipeline, ScoreSdeVePipeline, ) from .schedulers import ( CMStochasticIterativeScheduler, DDIMInverseScheduler, DDIMParallelScheduler, DDIMScheduler, DDPMParallelScheduler, DDPMScheduler, DEISMultistepScheduler, DPMSolverMultistepInverseScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, HeunDiscreteScheduler, IPNDMScheduler, KarrasVeScheduler, KDPMaAncestralDiscreteScheduler, KDPMaDiscreteScheduler, PNDMScheduler, RePaintScheduler, SchedulerMixin, ScoreSdeVeScheduler, UnCLIPScheduler, UniPCMultistepScheduler, VQDiffusionScheduler, ) from .training_utils import EMAModel try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .schedulers import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .schedulers import DPMSolverSDEScheduler try: if not (is_torch_available() and is_transformers_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipelines import ( AltDiffusionImgaImgPipeline, AltDiffusionPipeline, AudioLDMPipeline, CycleDiffusionPipeline, IFImgaImgPipeline, IFImgaImgSuperResolutionPipeline, IFInpaintingPipeline, IFInpaintingSuperResolutionPipeline, IFPipeline, IFSuperResolutionPipeline, ImageTextPipelineOutput, KandinskyImgaImgPipeline, KandinskyInpaintPipeline, KandinskyPipeline, KandinskyPriorPipeline, KandinskyVaaControlnetImgaImgPipeline, KandinskyVaaControlnetPipeline, KandinskyVaaImgaImgPipeline, KandinskyVaaInpaintPipeline, KandinskyVaaPipeline, KandinskyVaaPriorEmbaEmbPipeline, KandinskyVaaPriorPipeline, LDMTextToImagePipeline, PaintByExamplePipeline, SemanticStableDiffusionPipeline, ShapEImgaImgPipeline, ShapEPipeline, StableDiffusionAttendAndExcitePipeline, StableDiffusionControlNetImgaImgPipeline, StableDiffusionControlNetInpaintPipeline, StableDiffusionControlNetPipeline, StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionImageVariationPipeline, StableDiffusionImgaImgPipeline, StableDiffusionInpaintPipeline, StableDiffusionInpaintPipelineLegacy, StableDiffusionInstructPixaPixPipeline, StableDiffusionLatentUpscalePipeline, StableDiffusionLDMaDPipeline, StableDiffusionModelEditingPipeline, StableDiffusionPanoramaPipeline, StableDiffusionParadigmsPipeline, StableDiffusionPipeline, StableDiffusionPipelineSafe, StableDiffusionPixaPixZeroPipeline, StableDiffusionSAGPipeline, StableDiffusionUpscalePipeline, StableUnCLIPImgaImgPipeline, StableUnCLIPPipeline, TextToVideoSDPipeline, TextToVideoZeroPipeline, UnCLIPImageVariationPipeline, UnCLIPPipeline, UniDiffuserModel, UniDiffuserPipeline, UniDiffuserTextDecoder, VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, VideoToVideoSDPipeline, VQDiffusionPipeline, ) try: if not (is_torch_available() and is_transformers_available() and is_invisible_watermark_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_invisible_watermark_objects import * # noqa F403 else: from .pipelines import StableDiffusionXLImgaImgPipeline, StableDiffusionXLPipeline try: if not (is_torch_available() and is_transformers_available() and is_k_diffusion_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipelines import StableDiffusionKDiffusionPipeline try: if not (is_torch_available() and is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_transformers_and_onnx_objects import * # noqa F403 else: from .pipelines import ( OnnxStableDiffusionImgaImgPipeline, OnnxStableDiffusionInpaintPipeline, OnnxStableDiffusionInpaintPipelineLegacy, OnnxStableDiffusionPipeline, OnnxStableDiffusionUpscalePipeline, StableDiffusionOnnxPipeline, ) try: if not (is_torch_available() and is_librosa_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_torch_and_librosa_objects import * # noqa F403 else: from .pipelines import AudioDiffusionPipeline, Mel try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .pipelines import SpectrogramDiffusionPipeline try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_flax_objects import * # noqa F403 else: from .models.controlnet_flax import FlaxControlNetModel from .models.modeling_flax_utils import FlaxModelMixin from .models.unet_ad_condition_flax import FlaxUNetaDConditionModel from .models.vae_flax import FlaxAutoencoderKL from .pipelines import FlaxDiffusionPipeline from .schedulers import ( FlaxDDIMScheduler, FlaxDDPMScheduler, FlaxDPMSolverMultistepScheduler, FlaxKarrasVeScheduler, FlaxLMSDiscreteScheduler, FlaxPNDMScheduler, FlaxSchedulerMixin, FlaxScoreSdeVeScheduler, ) try: if not (is_flax_available() and is_transformers_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_flax_and_transformers_objects import * # noqa F403 else: from .pipelines import ( FlaxStableDiffusionControlNetPipeline, FlaxStableDiffusionImgaImgPipeline, FlaxStableDiffusionInpaintPipeline, FlaxStableDiffusionPipeline, ) try: if not (is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from .utils.dummy_note_seq_objects import * # noqa F403 else: from .pipelines import MidiProcessor
19
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase__ : Tuple = { """configuration_roberta""": ["""ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """RobertaConfig""", """RobertaOnnxConfig"""], """tokenization_roberta""": ["""RobertaTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : Tuple = ["""RobertaTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : Optional[int] = [ """ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """RobertaForCausalLM""", """RobertaForMaskedLM""", """RobertaForMultipleChoice""", """RobertaForQuestionAnswering""", """RobertaForSequenceClassification""", """RobertaForTokenClassification""", """RobertaModel""", """RobertaPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : List[str] = [ """TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFRobertaForCausalLM""", """TFRobertaForMaskedLM""", """TFRobertaForMultipleChoice""", """TFRobertaForQuestionAnswering""", """TFRobertaForSequenceClassification""", """TFRobertaForTokenClassification""", """TFRobertaMainLayer""", """TFRobertaModel""", """TFRobertaPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : str = [ """FlaxRobertaForCausalLM""", """FlaxRobertaForMaskedLM""", """FlaxRobertaForMultipleChoice""", """FlaxRobertaForQuestionAnswering""", """FlaxRobertaForSequenceClassification""", """FlaxRobertaForTokenClassification""", """FlaxRobertaModel""", """FlaxRobertaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_roberta import ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaConfig, RobertaOnnxConfig from .tokenization_roberta import RobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roberta_fast import RobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta import ( ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaForCausalLM, RobertaForMaskedLM, RobertaForMultipleChoice, RobertaForQuestionAnswering, RobertaForSequenceClassification, RobertaForTokenClassification, RobertaModel, RobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta import ( TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForMultipleChoice, TFRobertaForQuestionAnswering, TFRobertaForSequenceClassification, TFRobertaForTokenClassification, TFRobertaMainLayer, TFRobertaModel, TFRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, FlaxRobertaPreTrainedModel, ) else: import sys lowerCamelCase__ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
33
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging UpperCAmelCase__ = logging.get_logger(__name__) if is_vision_available(): import PIL class lowercase_ ( snake_case_ ): '''simple docstring''' __snake_case = ['pixel_values'] def __init__( self : Union[str, Any] , __UpperCAmelCase : bool = True , __UpperCAmelCase : Dict[str, int] = None , __UpperCAmelCase : PILImageResampling = PILImageResampling.BICUBIC , __UpperCAmelCase : bool = True , __UpperCAmelCase : Dict[str, int] = None , __UpperCAmelCase : bool = True , __UpperCAmelCase : Union[int, float] = 1 / 255 , __UpperCAmelCase : bool = True , __UpperCAmelCase : Optional[Union[float, List[float]]] = None , __UpperCAmelCase : Optional[Union[float, List[float]]] = None , __UpperCAmelCase : bool = True , **__UpperCAmelCase : Tuple , ) ->List[str]: """simple docstring""" super().__init__(**_a ) a = size if size is not None else {'''shortest_edge''': 224} a = get_size_dict(_a , default_to_square=_a ) a = crop_size if crop_size is not None else {'''height''': 224, '''width''': 224} a = get_size_dict(_a , default_to_square=_a , param_name='''crop_size''' ) a = do_resize a = size a = resample a = do_center_crop a = crop_size a = do_rescale a = rescale_factor a = do_normalize a = image_mean if image_mean is not None else OPENAI_CLIP_MEAN a = image_std if image_std is not None else OPENAI_CLIP_STD a = do_convert_rgb def __lowerCAmelCase ( self : Optional[int] , __UpperCAmelCase : np.ndarray , __UpperCAmelCase : Dict[str, int] , __UpperCAmelCase : PILImageResampling = PILImageResampling.BICUBIC , __UpperCAmelCase : Optional[Union[str, ChannelDimension]] = None , **__UpperCAmelCase : Dict , ) ->Dict: """simple docstring""" a = get_size_dict(_a , default_to_square=_a ) if "shortest_edge" not in size: raise ValueError(F"""The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}""" ) a = get_resize_output_image_size(_a , size=size['''shortest_edge'''] , default_to_square=_a ) return resize(_a , size=_a , resample=_a , data_format=_a , **_a ) def __lowerCAmelCase ( self : Dict , __UpperCAmelCase : np.ndarray , __UpperCAmelCase : Dict[str, int] , __UpperCAmelCase : Optional[Union[str, ChannelDimension]] = None , **__UpperCAmelCase : Optional[int] , ) ->Optional[int]: """simple docstring""" a = get_size_dict(_a ) if "height" not in size or "width" not in size: raise ValueError(F"""The `size` parameter must contain the keys (height, width). Got {size.keys()}""" ) return center_crop(_a , size=(size['''height'''], size['''width''']) , data_format=_a , **_a ) def __lowerCAmelCase ( self : Union[str, Any] , __UpperCAmelCase : np.ndarray , __UpperCAmelCase : Union[int, float] , __UpperCAmelCase : Optional[Union[str, ChannelDimension]] = None , **__UpperCAmelCase : str , ) ->List[Any]: """simple docstring""" return rescale(_a , scale=_a , data_format=_a , **_a ) def __lowerCAmelCase ( self : Dict , __UpperCAmelCase : np.ndarray , __UpperCAmelCase : Union[float, List[float]] , __UpperCAmelCase : Union[float, List[float]] , __UpperCAmelCase : Optional[Union[str, ChannelDimension]] = None , **__UpperCAmelCase : Optional[Any] , ) ->Union[str, Any]: """simple docstring""" return normalize(_a , mean=_a , std=_a , data_format=_a , **_a ) def __lowerCAmelCase ( self : Tuple , __UpperCAmelCase : ImageInput , __UpperCAmelCase : bool = None , __UpperCAmelCase : Dict[str, int] = None , __UpperCAmelCase : PILImageResampling = None , __UpperCAmelCase : bool = None , __UpperCAmelCase : int = None , __UpperCAmelCase : bool = None , __UpperCAmelCase : float = None , __UpperCAmelCase : bool = None , __UpperCAmelCase : Optional[Union[float, List[float]]] = None , __UpperCAmelCase : Optional[Union[float, List[float]]] = None , __UpperCAmelCase : bool = None , __UpperCAmelCase : Optional[Union[str, TensorType]] = None , __UpperCAmelCase : Optional[ChannelDimension] = ChannelDimension.FIRST , **__UpperCAmelCase : int , ) ->List[Any]: """simple docstring""" a = do_resize if do_resize is not None else self.do_resize a = size if size is not None else self.size a = get_size_dict(_a , param_name='''size''' , default_to_square=_a ) a = resample if resample is not None else self.resample a = do_center_crop if do_center_crop is not None else self.do_center_crop a = crop_size if crop_size is not None else self.crop_size a = get_size_dict(_a , param_name='''crop_size''' , default_to_square=_a ) a = do_rescale if do_rescale is not None else self.do_rescale a = rescale_factor if rescale_factor is not None else self.rescale_factor a = do_normalize if do_normalize is not None else self.do_normalize a = image_mean if image_mean is not None else self.image_mean a = image_std if image_std is not None else self.image_std a = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb a = make_list_of_images(_a ) if not valid_images(_a ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # PIL RGBA images are converted to RGB if do_convert_rgb: a = [convert_to_rgb(_a ) for image in images] # All transformations expect numpy arrays. a = [to_numpy_array(_a ) for image in images] if do_resize: a = [self.resize(image=_a , size=_a , resample=_a ) for image in images] if do_center_crop: a = [self.center_crop(image=_a , size=_a ) for image in images] if do_rescale: a = [self.rescale(image=_a , scale=_a ) for image in images] if do_normalize: a = [self.normalize(image=_a , mean=_a , std=_a ) for image in images] a = [to_channel_dimension_format(_a , _a ) for image in images] a = {'''pixel_values''': images} return BatchFeature(data=_a , tensor_type=_a )
117
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer import diffusers from diffusers import ( AutoencoderKL, EulerDiscreteScheduler, StableDiffusionLatentUpscalePipeline, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.schedulers import KarrasDiffusionSchedulers from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> List[Any]: snake_case__ = [tensor.shape for tensor in tensor_list] return all(shape == shapes[0] for shape in shapes[1:] ) class __magic_name__ (snake_case_ ,snake_case_ ,snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : Dict = StableDiffusionLatentUpscalePipeline __lowercase : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { 'height', 'width', 'cross_attention_kwargs', 'negative_prompt_embeds', 'prompt_embeds', } __lowercase : List[Any] = PipelineTesterMixin.required_optional_params - {'num_images_per_prompt'} __lowercase : Any = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS __lowercase : int = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __lowercase : List[Any] = frozenset([] ) __lowercase : Any = True @property def SCREAMING_SNAKE_CASE__ ( self:List[str] ): snake_case__ = 1 snake_case__ = 4 snake_case__ = (16, 16) snake_case__ = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(_a ) return image def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): torch.manual_seed(0 ) snake_case__ = UNetaDConditionModel( act_fn='''gelu''' , attention_head_dim=8 , norm_num_groups=_a , block_out_channels=[32, 32, 64, 64] , time_cond_proj_dim=1_60 , conv_in_kernel=1 , conv_out_kernel=1 , cross_attention_dim=32 , down_block_types=( '''KDownBlock2D''', '''KCrossAttnDownBlock2D''', '''KCrossAttnDownBlock2D''', '''KCrossAttnDownBlock2D''', ) , in_channels=8 , mid_block_type=_a , only_cross_attention=_a , out_channels=5 , resnet_time_scale_shift='''scale_shift''' , time_embedding_type='''fourier''' , timestep_post_act='''gelu''' , up_block_types=('''KCrossAttnUpBlock2D''', '''KCrossAttnUpBlock2D''', '''KCrossAttnUpBlock2D''', '''KUpBlock2D''') , ) snake_case__ = AutoencoderKL( block_out_channels=[32, 32, 64, 64] , in_channels=3 , out_channels=3 , down_block_types=[ '''DownEncoderBlock2D''', '''DownEncoderBlock2D''', '''DownEncoderBlock2D''', '''DownEncoderBlock2D''', ] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D''', '''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) snake_case__ = EulerDiscreteScheduler(prediction_type='''sample''' ) snake_case__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act='''quick_gelu''' , projection_dim=5_12 , ) snake_case__ = CLIPTextModel(_a ) snake_case__ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) snake_case__ = { '''unet''': model.eval(), '''vae''': vae.eval(), '''scheduler''': scheduler, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def SCREAMING_SNAKE_CASE__ ( self:List[Any] , _a:Optional[Any] , _a:List[str]=0 ): if str(_a ).startswith('''mps''' ): snake_case__ = torch.manual_seed(_a ) else: snake_case__ = torch.Generator(device=_a ).manual_seed(_a ) snake_case__ = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': self.dummy_image.cpu(), '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = '''cpu''' snake_case__ = self.get_dummy_components() snake_case__ = self.pipeline_class(**_a ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) snake_case__ = self.get_dummy_inputs(_a ) snake_case__ = pipe(**_a ).images snake_case__ = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 2_56, 2_56, 3) ) snake_case__ = np.array( [0.47222412, 0.41921633, 0.44717434, 0.46874192, 0.42588258, 0.46150726, 0.4677534, 0.45583832, 0.48579055] ) snake_case__ = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_a , 1e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): super().test_attention_slicing_forward_pass(expected_max_diff=7e-3 ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): super().test_cpu_offload_forward_pass(expected_max_diff=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:str ): super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Any ): super().test_inference_batch_single_identical(expected_max_diff=7e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): super().test_pt_np_pil_outputs_equivalent(expected_max_diff=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): super().test_save_load_local(expected_max_difference=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:str ): super().test_save_load_optional_components(expected_max_difference=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = [ '''DDIMScheduler''', '''DDPMScheduler''', '''PNDMScheduler''', '''HeunDiscreteScheduler''', '''EulerAncestralDiscreteScheduler''', '''KDPM2DiscreteScheduler''', '''KDPM2AncestralDiscreteScheduler''', '''DPMSolverSDEScheduler''', ] snake_case__ = self.get_dummy_components() snake_case__ = self.pipeline_class(**_a ) # make sure that PNDM does not need warm-up pipe.scheduler.register_to_config(skip_prk_steps=_a ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) snake_case__ = self.get_dummy_inputs(_a ) snake_case__ = 2 snake_case__ = [] for scheduler_enum in KarrasDiffusionSchedulers: if scheduler_enum.name in skip_schedulers: # no sigma schedulers are not supported # no schedulers continue snake_case__ = getattr(_a , scheduler_enum.name ) snake_case__ = scheduler_cls.from_config(pipe.scheduler.config ) snake_case__ = pipe(**_a )[0] outputs.append(_a ) assert check_same_shape(_a ) @require_torch_gpu @slow class __magic_name__ (unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = torch.manual_seed(33 ) snake_case__ = StableDiffusionPipeline.from_pretrained('''CompVis/stable-diffusion-v1-4''' , torch_dtype=torch.floataa ) pipe.to('''cuda''' ) snake_case__ = StableDiffusionLatentUpscalePipeline.from_pretrained( '''stabilityai/sd-x2-latent-upscaler''' , torch_dtype=torch.floataa ) upscaler.to('''cuda''' ) snake_case__ = '''a photo of an astronaut high resolution, unreal engine, ultra realistic''' snake_case__ = pipe(_a , generator=_a , output_type='''latent''' ).images snake_case__ = upscaler( prompt=_a , image=_a , num_inference_steps=20 , guidance_scale=0 , generator=_a , output_type='''np''' , ).images[0] snake_case__ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/astronaut_1024.npy''' ) assert np.abs((expected_image - image).mean() ) < 5e-2 def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): snake_case__ = torch.manual_seed(33 ) snake_case__ = StableDiffusionLatentUpscalePipeline.from_pretrained( '''stabilityai/sd-x2-latent-upscaler''' , torch_dtype=torch.floataa ) upscaler.to('''cuda''' ) snake_case__ = '''the temple of fire by Ross Tran and Gerardo Dottori, oil on canvas''' snake_case__ = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_512.png''' ) snake_case__ = upscaler( prompt=_a , image=_a , num_inference_steps=20 , guidance_scale=0 , generator=_a , output_type='''np''' , ).images[0] snake_case__ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_1024.npy''' ) assert np.abs((expected_image - image).max() ) < 5e-2
33
0
import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGeneration, MusicgenProcessor, TaEncoderModel, ) from transformers.models.musicgen.modeling_musicgen import MusicgenForCausalLM from transformers.utils import logging logging.set_verbosity_info() _lowerCamelCase : Dict = logging.get_logger(__name__) _lowerCamelCase : str = ["""model.decoder.embed_positions.weights"""] def A__ ( __A : Optional[Any] ) ->Union[str, Any]: if "emb" in name: __A =name.replace('''emb''' , '''model.decoder.embed_tokens''' ) if "transformer" in name: __A =name.replace('''transformer''' , '''model.decoder''' ) if "cross_attention" in name: __A =name.replace('''cross_attention''' , '''encoder_attn''' ) if "linear1" in name: __A =name.replace('''linear1''' , '''fc1''' ) if "linear2" in name: __A =name.replace('''linear2''' , '''fc2''' ) if "norm1" in name: __A =name.replace('''norm1''' , '''self_attn_layer_norm''' ) if "norm_cross" in name: __A =name.replace('''norm_cross''' , '''encoder_attn_layer_norm''' ) if "norm2" in name: __A =name.replace('''norm2''' , '''final_layer_norm''' ) if "out_norm" in name: __A =name.replace('''out_norm''' , '''model.decoder.layer_norm''' ) if "linears" in name: __A =name.replace('''linears''' , '''lm_heads''' ) if "condition_provider.conditioners.description.output_proj" in name: __A =name.replace('''condition_provider.conditioners.description.output_proj''' , '''enc_to_dec_proj''' ) return name def A__ ( __A : Union[str, Any] , __A : Optional[int] ) ->Tuple[Dict, Dict]: __A =list(state_dict.keys() ) __A ={} for key in keys: __A =state_dict.pop(__lowerCAmelCase ) __A =rename_keys(__lowerCAmelCase ) if "in_proj_weight" in key: # split fused qkv proj __A =val[:hidden_size, :] __A =val[hidden_size : 2 * hidden_size, :] __A =val[-hidden_size:, :] elif "enc_to_dec_proj" in key: __A =val else: __A =val return state_dict, enc_dec_proj_state_dict def A__ ( __A : Optional[int] ) ->MusicgenDecoderConfig: if checkpoint == "small": # default config values __A =10_24 __A =24 __A =16 elif checkpoint == "medium": __A =15_36 __A =48 __A =24 elif checkpoint == "large": __A =20_48 __A =48 __A =32 else: raise ValueError(F'''Checkpoint should be one of `[\'small\', \'medium\', \'large\']`, got {checkpoint}.''' ) __A =MusicgenDecoderConfig( hidden_size=__lowerCAmelCase , ffn_dim=hidden_size * 4 , num_hidden_layers=__lowerCAmelCase , num_attention_heads=__lowerCAmelCase , ) return config @torch.no_grad() def A__ ( __A : Dict , __A : List[Any]=None , __A : Any=None , __A : Tuple="cpu" ) ->Union[str, Any]: __A =MusicGen.get_pretrained(__lowerCAmelCase , device=__lowerCAmelCase ) __A =decoder_config_from_checkpoint(__lowerCAmelCase ) __A =fairseq_model.lm.state_dict() __A , __A =rename_state_dict( __lowerCAmelCase , hidden_size=decoder_config.hidden_size ) __A =TaEncoderModel.from_pretrained('''t5-base''' ) __A =EncodecModel.from_pretrained('''facebook/encodec_32khz''' ) __A =MusicgenForCausalLM(__lowerCAmelCase ).eval() # load all decoder weights - expect that we'll be missing embeddings and enc-dec projection __A , __A =decoder.load_state_dict(__lowerCAmelCase , strict=__lowerCAmelCase ) for key in missing_keys.copy(): if key.startswith(('''text_encoder''', '''audio_encoder''') ) or key in EXPECTED_MISSING_KEYS: missing_keys.remove(__lowerCAmelCase ) if len(__lowerCAmelCase ) > 0: raise ValueError(F'''Missing key(s) in state_dict: {missing_keys}''' ) if len(__lowerCAmelCase ) > 0: raise ValueError(F'''Unexpected key(s) in state_dict: {unexpected_keys}''' ) # init the composite model __A =MusicgenForConditionalGeneration(text_encoder=__lowerCAmelCase , audio_encoder=__lowerCAmelCase , decoder=__lowerCAmelCase ) # load the pre-trained enc-dec projection (from the decoder state dict) model.enc_to_dec_proj.load_state_dict(__lowerCAmelCase ) # check we can do a forward pass __A =torch.arange(0 , 8 , dtype=torch.long ).reshape(2 , -1 ) __A =input_ids.reshape(2 * 4 , -1 ) with torch.no_grad(): __A =model(input_ids=__lowerCAmelCase , decoder_input_ids=__lowerCAmelCase ).logits if logits.shape != (8, 1, 20_48): raise ValueError('''Incorrect shape for logits''' ) # now construct the processor __A =AutoTokenizer.from_pretrained('''t5-base''' ) __A =AutoFeatureExtractor.from_pretrained('''facebook/encodec_32khz''' , padding_side='''left''' ) __A =MusicgenProcessor(feature_extractor=__lowerCAmelCase , tokenizer=__lowerCAmelCase ) # set the appropriate bos/pad token ids __A =20_48 __A =20_48 # set other default generation config params __A =int(30 * audio_encoder.config.frame_rate ) __A =True __A =3.0 if pytorch_dump_folder is not None: Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) logger.info(F'''Saving model {checkpoint} to {pytorch_dump_folder}''' ) model.save_pretrained(__lowerCAmelCase ) processor.save_pretrained(__lowerCAmelCase ) if repo_id: logger.info(F'''Pushing model {checkpoint} to {repo_id}''' ) model.push_to_hub(__lowerCAmelCase ) processor.push_to_hub(__lowerCAmelCase ) if __name__ == "__main__": _lowerCamelCase : Any = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint''', default='''small''', type=str, help='''Checkpoint size of the MusicGen model you\'d like to convert. Can be one of: `[\'small\', \'medium\', \'large\']`.''', ) parser.add_argument( '''--pytorch_dump_folder''', required=True, default=None, type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', default=None, type=str, help='''Where to upload the converted model on the 🤗 hub.''' ) parser.add_argument( '''--device''', default='''cpu''', type=str, help='''Torch device to run the conversion, either cpu or cuda.''' ) _lowerCamelCase : Optional[int] = parser.parse_args() convert_musicgen_checkpoint(args.checkpoint, args.pytorch_dump_folder, args.push_to_hub)
184
import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class __magic_name__ (unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = '''ZinengTang/tvlt-base''' snake_case__ = tempfile.mkdtemp() def SCREAMING_SNAKE_CASE__ ( self:Dict , **_a:List[Any] ): return TvltImageProcessor.from_pretrained(self.checkpoint , **_a ) def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] , **_a:Tuple ): return TvltFeatureExtractor.from_pretrained(self.checkpoint , **_a ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): shutil.rmtree(self.tmpdirname ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) processor.save_pretrained(self.tmpdirname ) snake_case__ = TvltProcessor.from_pretrained(self.tmpdirname ) self.assertIsInstance(processor.feature_extractor , _a ) self.assertIsInstance(processor.image_processor , _a ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) snake_case__ = np.ones([1_20_00] ) snake_case__ = feature_extractor(_a , return_tensors='''np''' ) snake_case__ = processor(audio=_a , return_tensors='''np''' ) for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) snake_case__ = np.ones([3, 2_24, 2_24] ) snake_case__ = image_processor(_a , return_tensors='''np''' ) snake_case__ = processor(images=_a , return_tensors='''np''' ) for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) snake_case__ = np.ones([1_20_00] ) snake_case__ = np.ones([3, 2_24, 2_24] ) snake_case__ = processor(audio=_a , images=_a ) self.assertListEqual(list(inputs.keys() ) , ['''audio_values''', '''audio_mask''', '''pixel_values''', '''pixel_mask'''] ) # test if it raises when no input is passed with pytest.raises(_a ): processor() def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg='''`processor` and `image_processor`+`feature_extractor` model input names do not match''' , )
33
0
import warnings from ...utils import logging from .image_processing_clip import CLIPImageProcessor SCREAMING_SNAKE_CASE_ : List[str] = logging.get_logger(__name__) class snake_case_ ( snake_case_ ): '''simple docstring''' def __init__( self : Dict , *__lowerCamelCase : Union[str, Any] , **__lowerCamelCase : List[Any] ) -> Any: '''simple docstring''' warnings.warn( 'The class CLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use CLIPImageProcessor instead.' , _a , ) super().__init__(*_a , **_a )
375
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase__ : List[Any] = logging.get_logger(__name__) lowerCamelCase__ : Optional[int] = { """facebook/data2vec-vision-base-ft""": ( """https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json""" ), } class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : Optional[int] = 'data2vec-vision' def __init__( self:int , _a:Tuple=7_68 , _a:int=12 , _a:Any=12 , _a:Optional[int]=30_72 , _a:Optional[int]="gelu" , _a:Any=0.0 , _a:Any=0.0 , _a:List[str]=0.02 , _a:Dict=1e-12 , _a:Tuple=2_24 , _a:Any=16 , _a:str=3 , _a:str=False , _a:Union[str, Any]=False , _a:Optional[int]=False , _a:Any=False , _a:Dict=0.1 , _a:Dict=0.1 , _a:str=True , _a:str=[3, 5, 7, 11] , _a:List[str]=[1, 2, 3, 6] , _a:List[str]=True , _a:Any=0.4 , _a:str=2_56 , _a:Union[str, Any]=1 , _a:int=False , _a:Optional[int]=2_55 , **_a:Dict , ): super().__init__(**_a ) snake_case__ = hidden_size snake_case__ = num_hidden_layers snake_case__ = num_attention_heads snake_case__ = intermediate_size snake_case__ = hidden_act snake_case__ = hidden_dropout_prob snake_case__ = attention_probs_dropout_prob snake_case__ = initializer_range snake_case__ = layer_norm_eps snake_case__ = image_size snake_case__ = patch_size snake_case__ = num_channels snake_case__ = use_mask_token snake_case__ = use_absolute_position_embeddings snake_case__ = use_relative_position_bias snake_case__ = use_shared_relative_position_bias snake_case__ = layer_scale_init_value snake_case__ = drop_path_rate snake_case__ = use_mean_pooling # decode head attributes (semantic segmentation) snake_case__ = out_indices snake_case__ = pool_scales # auxiliary head attributes (semantic segmentation) snake_case__ = use_auxiliary_head snake_case__ = auxiliary_loss_weight snake_case__ = auxiliary_channels snake_case__ = auxiliary_num_convs snake_case__ = auxiliary_concat_input snake_case__ = semantic_loss_ignore_index class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : Any = version.parse('1.11' ) @property def SCREAMING_SNAKE_CASE__ ( self:List[str] ): return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def SCREAMING_SNAKE_CASE__ ( self:Tuple ): return 1e-4
33
0
import enum import warnings from .. import MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING from ..utils import add_end_docstrings, is_tf_available from .base import PIPELINE_INIT_ARGS, Pipeline if is_tf_available(): import tensorflow as tf class UpperCAmelCase_ ( enum.Enum ): '''simple docstring''' A : Dict = 0 A : Tuple = 1 A : List[Any] = 2 @add_end_docstrings(snake_case_ ) class UpperCAmelCase_ ( snake_case_ ): '''simple docstring''' A : List[Any] = '\n In 1991, the remains of Russian Tsar Nicholas II and his family (except for Alexei and Maria) are discovered. The\n voice of Nicholas\'s young son, Tsarevich Alexei Nikolaevich, narrates the remainder of the story. 1883 Western\n Siberia, a young Grigori Rasputin is asked by his father and a group of men to perform magic. Rasputin has a vision\n and denounces one of the men as a horse thief. Although his father initially slaps him for making such an\n accusation, Rasputin watches as the man is chased outside and beaten. Twenty years later, Rasputin sees a vision of\n the Virgin Mary, prompting him to become a priest. Rasputin quickly becomes famous, with people, even a bishop,\n begging for his blessing. <eod> </s> <eos>\n ' def __init__( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> Union[str, Any]: super().__init__(*_a , **_a ) self.check_model_type( TF_MODEL_FOR_CAUSAL_LM_MAPPING if self.framework == "tf" else MODEL_FOR_CAUSAL_LM_MAPPING ) if "prefix" not in self._preprocess_params: # This is very specific. The logic is quite complex and needs to be done # as a "default". # It also defines both some preprocess_kwargs and generate_kwargs # which is why we cannot put them in their respective methods. snake_case_ : Optional[Any] = None if self.model.config.prefix is not None: snake_case_ : Any = self.model.config.prefix if prefix is None and self.model.__class__.__name__ in [ "XLNetLMHeadModel", "TransfoXLLMHeadModel", "TFXLNetLMHeadModel", "TFTransfoXLLMHeadModel", ]: # For XLNet and TransformerXL we add an article to the prompt to give more state to the model. snake_case_ : Dict = self.XL_PREFIX if prefix is not None: # Recalculate some generate_kwargs linked to prefix. snake_case_ , snake_case_ , snake_case_ : List[Any] = self._sanitize_parameters(prefix=_a , **self._forward_params ) snake_case_ : Union[str, Any] = {**self._preprocess_params, **preprocess_params} snake_case_ : int = {**self._forward_params, **forward_params} def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , _SCREAMING_SNAKE_CASE=None , **_SCREAMING_SNAKE_CASE , ) -> int: snake_case_ : Optional[Any] = {} if prefix is not None: snake_case_ : Union[str, Any] = prefix if prefix: snake_case_ : List[str] = self.tokenizer( _a , padding=_a , add_special_tokens=_a , return_tensors=self.framework ) snake_case_ : int = prefix_inputs["input_ids"].shape[-1] if handle_long_generation is not None: if handle_long_generation not in {"hole"}: raise ValueError( f'''{handle_long_generation} is not a valid value for `handle_long_generation` parameter expected''' " [None, \'hole\']" ) snake_case_ : int = handle_long_generation preprocess_params.update(_a ) snake_case_ : List[str] = generate_kwargs snake_case_ : Tuple = {} if return_full_text is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_full_text`" ) if return_tensors is not None: raise ValueError("`return_full_text` is mutually exclusive with `return_tensors`" ) snake_case_ : Tuple = ReturnType.FULL_TEXT if return_full_text else ReturnType.NEW_TEXT if return_tensors is not None and return_type is None: if return_text is not None: raise ValueError("`return_text` is mutually exclusive with `return_tensors`" ) snake_case_ : Tuple = ReturnType.TENSORS if return_type is not None: snake_case_ : Union[str, Any] = return_type if clean_up_tokenization_spaces is not None: snake_case_ : List[Any] = clean_up_tokenization_spaces if stop_sequence is not None: snake_case_ : Tuple = self.tokenizer.encode(_a , add_special_tokens=_a ) if len(_a ) > 1: warnings.warn( "Stopping on a multiple token sequence is not yet supported on transformers. The first token of" " the stop sequence will be used as the stop sequence string in the interim." ) snake_case_ : List[str] = stop_sequence_ids[0] return preprocess_params, forward_params, postprocess_params def _lowerCAmelCase ( self , *_SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> int: # Parse arguments if self.model.__class__.__name__ in ["TransfoXLLMHeadModel"]: kwargs.update({"add_space_before_punct_symbol": True} ) return super()._parse_and_tokenize(*_a , **_a ) def __call__( self , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> Optional[int]: return super().__call__(_a , **_a ) def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE="" , _SCREAMING_SNAKE_CASE=None , **_SCREAMING_SNAKE_CASE ) -> Optional[Any]: snake_case_ : int = self.tokenizer( prefix + prompt_text , padding=_a , add_special_tokens=_a , return_tensors=self.framework ) snake_case_ : Optional[int] = prompt_text if handle_long_generation == "hole": snake_case_ : Tuple = inputs["input_ids"].shape[-1] if "max_new_tokens" in generate_kwargs: snake_case_ : Dict = generate_kwargs["max_new_tokens"] else: snake_case_ : Optional[Any] = generate_kwargs.get("max_length" , self.model.config.max_length ) - cur_len if new_tokens < 0: raise ValueError("We cannot infer how many new tokens are expected" ) if cur_len + new_tokens > self.tokenizer.model_max_length: snake_case_ : Union[str, Any] = self.tokenizer.model_max_length - new_tokens if keep_length <= 0: raise ValueError( "We cannot use `hole` to handle this generation the number of desired tokens exceeds the" " models max length" ) snake_case_ : str = inputs["input_ids"][:, -keep_length:] if "attention_mask" in inputs: snake_case_ : Optional[Any] = inputs["attention_mask"][:, -keep_length:] return inputs def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , **_SCREAMING_SNAKE_CASE ) -> Dict: snake_case_ : Union[str, Any] = model_inputs["input_ids"] snake_case_ : Any = model_inputs.get("attention_mask" , _a ) # Allow empty prompts if input_ids.shape[1] == 0: snake_case_ : Optional[int] = None snake_case_ : Dict = None snake_case_ : Any = 1 else: snake_case_ : Optional[int] = input_ids.shape[0] snake_case_ : int = model_inputs.pop("prompt_text" ) # If there is a prefix, we may need to adjust the generation length. Do so without permanently modifying # generate_kwargs, as some of the parameterization may come from the initialization of the pipeline. snake_case_ : Union[str, Any] = generate_kwargs.pop("prefix_length" , 0 ) if prefix_length > 0: snake_case_ : Optional[int] = "max_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].max_new_tokens is not None ) if not has_max_new_tokens: snake_case_ : int = generate_kwargs.get("max_length" ) or self.model.config.max_length generate_kwargs["max_length"] += prefix_length snake_case_ : Any = "min_new_tokens" in generate_kwargs or ( "generation_config" in generate_kwargs and generate_kwargs["generation_config"].min_new_tokens is not None ) if not has_min_new_tokens and "min_length" in generate_kwargs: generate_kwargs["min_length"] += prefix_length # BS x SL snake_case_ : Tuple = self.model.generate(input_ids=_a , attention_mask=_a , **_a ) snake_case_ : int = generated_sequence.shape[0] if self.framework == "pt": snake_case_ : int = generated_sequence.reshape(_a , out_b // in_b , *generated_sequence.shape[1:] ) elif self.framework == "tf": snake_case_ : str = tf.reshape(_a , (in_b, out_b // in_b, *generated_sequence.shape[1:]) ) return {"generated_sequence": generated_sequence, "input_ids": input_ids, "prompt_text": prompt_text} def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE=ReturnType.FULL_TEXT , _SCREAMING_SNAKE_CASE=True ) -> Union[str, Any]: snake_case_ : Union[str, Any] = model_outputs["generated_sequence"][0] snake_case_ : Optional[Any] = model_outputs["input_ids"] snake_case_ : Dict = model_outputs["prompt_text"] snake_case_ : str = generated_sequence.numpy().tolist() snake_case_ : Dict = [] for sequence in generated_sequence: if return_type == ReturnType.TENSORS: snake_case_ : Tuple = {"generated_token_ids": sequence} elif return_type in {ReturnType.NEW_TEXT, ReturnType.FULL_TEXT}: # Decode text snake_case_ : Optional[int] = self.tokenizer.decode( _a , skip_special_tokens=_a , clean_up_tokenization_spaces=_a , ) # Remove PADDING prompt of the sequence if XLNet or Transfo-XL model is used if input_ids is None: snake_case_ : int = 0 else: snake_case_ : str = len( self.tokenizer.decode( input_ids[0] , skip_special_tokens=_a , clean_up_tokenization_spaces=_a , ) ) if return_type == ReturnType.FULL_TEXT: snake_case_ : Dict = prompt_text + text[prompt_length:] else: snake_case_ : str = text[prompt_length:] snake_case_ : int = {"generated_text": all_text} records.append(_a ) return records
568
import os import sys lowerCamelCase__ : Optional[int] = os.path.join(os.path.dirname(__file__), """src""") sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) lowerCamelCase__ : Optional[int] = [ """torch""", """numpy""", """tokenizers""", """filelock""", """requests""", """tqdm""", """regex""", """sentencepiece""", """sacremoses""", """importlib_metadata""", """huggingface_hub""", ] @add_start_docstrings(AutoConfig.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Any: return AutoConfig.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoTokenizer.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> List[str]: return AutoTokenizer.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModel.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Tuple: return AutoModel.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Union[str, Any]: return AutoModelForCausalLM.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> List[Any]: return AutoModelForMaskedLM.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> List[str]: return AutoModelForSequenceClassification.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Union[str, Any]: return AutoModelForQuestionAnswering.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase )
33
0
def _lowercase ( a__ : int ) -> int: """simple docstring""" _UpperCamelCase = 0 while num > 0: digit_sum += num % 10 num //= 10 return digit_sum def _lowercase ( a__ : str = 1_00 ) -> int: """simple docstring""" _UpperCamelCase = 1 _UpperCamelCase = 2 for i in range(2 , max_n + 1 ): _UpperCamelCase = pre_numerator _UpperCamelCase = 2 * i // 3 if i % 3 == 0 else 1 _UpperCamelCase = cur_numerator _UpperCamelCase = e_cont * pre_numerator + temp return sum_digits(__lowerCAmelCase ) if __name__ == "__main__": print(F'''{solution() = }''')
147
import torch from diffusers import CMStochasticIterativeScheduler from .test_schedulers import SchedulerCommonTest class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : str = (CMStochasticIterativeScheduler,) __lowercase : List[str] = 10 def SCREAMING_SNAKE_CASE__ ( self:int , **_a:Optional[int] ): snake_case__ = { '''num_train_timesteps''': 2_01, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } config.update(**_a ) return config def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): snake_case__ = 10 snake_case__ = self.get_scheduler_config() snake_case__ = self.scheduler_classes[0](**_a ) scheduler.set_timesteps(_a ) snake_case__ = scheduler.timesteps[0] snake_case__ = scheduler.timesteps[1] snake_case__ = self.dummy_sample snake_case__ = 0.1 * sample snake_case__ = scheduler.step(_a , _a , _a ).prev_sample snake_case__ = scheduler.step(_a , _a , _a ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def SCREAMING_SNAKE_CASE__ ( self:Any ): for timesteps in [10, 50, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=_a ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): for clip_denoised in [True, False]: self.check_over_configs(clip_denoised=_a ) def SCREAMING_SNAKE_CASE__ ( self:int ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = 1 scheduler.set_timesteps(_a ) snake_case__ = scheduler.timesteps snake_case__ = torch.manual_seed(0 ) snake_case__ = self.dummy_model() snake_case__ = self.dummy_sample_deter * scheduler.init_noise_sigma for i, t in enumerate(_a ): # 1. scale model input snake_case__ = scheduler.scale_model_input(_a , _a ) # 2. predict noise residual snake_case__ = model(_a , _a ) # 3. predict previous sample x_t-1 snake_case__ = scheduler.step(_a , _a , _a , generator=_a ).prev_sample snake_case__ = pred_prev_sample snake_case__ = torch.sum(torch.abs(_a ) ) snake_case__ = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 192.7614 ) < 1e-2 assert abs(result_mean.item() - 0.2510 ) < 1e-3 def SCREAMING_SNAKE_CASE__ ( self:Tuple ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = [1_06, 0] scheduler.set_timesteps(timesteps=_a ) snake_case__ = scheduler.timesteps snake_case__ = torch.manual_seed(0 ) snake_case__ = self.dummy_model() snake_case__ = self.dummy_sample_deter * scheduler.init_noise_sigma for t in timesteps: # 1. scale model input snake_case__ = scheduler.scale_model_input(_a , _a ) # 2. predict noise residual snake_case__ = model(_a , _a ) # 3. predict previous sample x_t-1 snake_case__ = scheduler.step(_a , _a , _a , generator=_a ).prev_sample snake_case__ = pred_prev_sample snake_case__ = torch.sum(torch.abs(_a ) ) snake_case__ = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 347.6357 ) < 1e-2 assert abs(result_mean.item() - 0.4527 ) < 1e-3 def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = [39, 30, 12, 15, 0] with self.assertRaises(_a , msg='''`timesteps` must be in descending order.''' ): scheduler.set_timesteps(timesteps=_a ) def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = [39, 30, 12, 1, 0] snake_case__ = len(_a ) with self.assertRaises(_a , msg='''Can only pass one of `num_inference_steps` or `timesteps`.''' ): scheduler.set_timesteps(num_inference_steps=_a , timesteps=_a ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = [scheduler.config.num_train_timesteps] with self.assertRaises( _a , msg='''`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}''' , ): scheduler.set_timesteps(timesteps=_a )
33
0
from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def __UpperCamelCase ( _A ): lowerCAmelCase_ , lowerCAmelCase_ = analyze_text(__lowerCAmelCase ) lowerCAmelCase_ = list(''' ''' + ascii_lowercase ) # what is our total sum of probabilities. lowerCAmelCase_ = sum(single_char_strings.values() ) # one length string lowerCAmelCase_ = 0 # for each alpha we go in our dict and if it is in it we calculate entropy for ch in my_alphas: if ch in single_char_strings: lowerCAmelCase_ = single_char_strings[ch] lowerCAmelCase_ = my_str / all_sum my_fir_sum += prob * math.loga(__lowerCAmelCase ) # entropy formula. # print entropy print(f"{round(-1 * my_fir_sum ):.1f}" ) # two len string lowerCAmelCase_ = sum(two_char_strings.values() ) lowerCAmelCase_ = 0 # for each alpha (two in size) calculate entropy. for cha in my_alphas: for cha in my_alphas: lowerCAmelCase_ = cha + cha if sequence in two_char_strings: lowerCAmelCase_ = two_char_strings[sequence] lowerCAmelCase_ = int(__lowerCAmelCase ) / all_sum my_sec_sum += prob * math.loga(__lowerCAmelCase ) # print second entropy print(f"{round(-1 * my_sec_sum ):.1f}" ) # print the difference between them print(f"{round((-1 * my_sec_sum) - (-1 * my_fir_sum) ):.1f}" ) def __UpperCamelCase ( _A ): lowerCAmelCase_ = Counter() # type: ignore lowerCAmelCase_ = Counter() # type: ignore single_char_strings[text[-1]] += 1 # first case when we have space at start. two_char_strings[" " + text[0]] += 1 for i in range(0 , len(__lowerCAmelCase ) - 1 ): single_char_strings[text[i]] += 1 two_char_strings[text[i : i + 2]] += 1 return single_char_strings, two_char_strings def __UpperCamelCase ( ): import doctest doctest.testmod() # text = ( # "Had repulsive dashwoods suspicion sincerity but advantage now him. Remark " # "easily garret nor nay. Civil those mrs enjoy shy fat merry. You greatest " # "jointure saw horrible. He private he on be imagine suppose. Fertile " # "beloved evident through no service elderly is. Blind there if every no so " # "at. Own neglected you preferred way sincerity delivered his attempted. To " # "of message cottage windows do besides against uncivil. Delightful " # "unreserved impossible few estimating men favourable see entreaties. She " # "propriety immediate was improving. He or entrance humoured likewise " # "moderate. Much nor game son say feel. Fat make met can must form into " # "gate. Me we offending prevailed discovery. " # ) # calculate_prob(text) if __name__ == "__main__": main()
431
import numpy as np def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> np.ndarray: return 1 / (1 + np.exp(-vector )) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> np.ndarray: return vector * sigmoid(__lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
33
0
import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger A__ : Tuple = """<<<<<<< This should probably be modified because it mentions: """ A__ : str = """======= >>>>>>> """ A__ : List[Any] = [ """TextEncoderConfig""", """ByteTextEncoder""", """SubwordTextEncoder""", """encoder_config""", """maybe_build_from_corpus""", """manual_dir""", ] A__ : Dict = [ # (pattern, replacement) # Order is important here for some replacements (r"""tfds\.core""", r"""datasets"""), (r"""tf\.io\.gfile\.GFile""", r"""open"""), (r"""tf\.([\w\d]+)""", r"""datasets.Value('\1')"""), (r"""tfds\.features\.Text\(\)""", r"""datasets.Value('string')"""), (r"""tfds\.features\.Text\(""", r"""datasets.Value('string'),"""), (r"""features\s*=\s*tfds.features.FeaturesDict\(""", r"""features=datasets.Features("""), (r"""tfds\.features\.FeaturesDict\(""", r"""dict("""), (r"""The TensorFlow Datasets Authors""", r"""The TensorFlow Datasets Authors and the HuggingFace Datasets Authors"""), (r"""tfds\.""", r"""datasets."""), (r"""dl_manager\.manual_dir""", r"""self.config.data_dir"""), (r"""self\.builder_config""", r"""self.config"""), ] def UpperCamelCase( __UpperCamelCase : int ): return ConvertCommand(args.tfds_path ,args.datasets_directory ) class __snake_case ( snake_case_ ): @staticmethod def UpperCAmelCase__ ( A_ : ArgumentParser): lowerCAmelCase_ : str = parser.add_parser( '''convert''' , help='''Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.''' , ) train_parser.add_argument( '''--tfds_path''' , type=_a , required=_a , help='''Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.''' , ) train_parser.add_argument( '''--datasets_directory''' , type=_a , required=_a , help='''Path to the HuggingFace Datasets folder.''') train_parser.set_defaults(func=_a) def __init__( self : Any , A_ : str , A_ : str , *A_ : Any): lowerCAmelCase_ : Any = get_logger('''datasets-cli/converting''') lowerCAmelCase_ : Any = tfds_path lowerCAmelCase_ : Any = datasets_directory def UpperCAmelCase__ ( self : List[str]): if os.path.isdir(self._tfds_path): lowerCAmelCase_ : Dict = os.path.abspath(self._tfds_path) elif os.path.isfile(self._tfds_path): lowerCAmelCase_ : str = os.path.dirname(self._tfds_path) else: raise ValueError('''--tfds_path is neither a directory nor a file. Please check path.''') lowerCAmelCase_ : List[Any] = os.path.abspath(self._datasets_directory) self._logger.info(F"""Converting datasets from {abs_tfds_path} to {abs_datasets_path}""") lowerCAmelCase_ : int = [] lowerCAmelCase_ : Optional[int] = [] lowerCAmelCase_ : List[str] = {} if os.path.isdir(self._tfds_path): lowerCAmelCase_ : Union[str, Any] = os.listdir(_a) else: lowerCAmelCase_ : Union[str, Any] = [os.path.basename(self._tfds_path)] for f_name in file_names: self._logger.info(F"""Looking at file {f_name}""") lowerCAmelCase_ : Dict = os.path.join(_a , _a) lowerCAmelCase_ : Tuple = os.path.join(_a , _a) if not os.path.isfile(_a) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info('''Skipping file''') continue with open(_a , encoding='''utf-8''') as f: lowerCAmelCase_ : Any = f.readlines() lowerCAmelCase_ : List[str] = [] lowerCAmelCase_ : Union[str, Any] = False lowerCAmelCase_ : List[Any] = False lowerCAmelCase_ : str = [] for line in lines: lowerCAmelCase_ : Any = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: lowerCAmelCase_ : List[Any] = '''import datasets\n''' elif "import tensorflow" in out_line: # order is important here lowerCAmelCase_ : str = '''''' continue elif "from absl import logging" in out_line: lowerCAmelCase_ : Dict = '''from datasets import logging\n''' elif "getLogger" in out_line: lowerCAmelCase_ : Dict = out_line.replace('''getLogger''' , '''get_logger''') elif any(expression in out_line for expression in TO_HIGHLIGHT): lowerCAmelCase_ : int = True lowerCAmelCase_ : Union[str, Any] = list(filter(lambda A_: e in out_line , _a)) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(_a) + '''\n''') out_lines.append(_a) out_lines.append(_a) continue else: for pattern, replacement in TO_CONVERT: lowerCAmelCase_ : Dict = re.sub(_a , _a , _a) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: lowerCAmelCase_ : Tuple = re.match(r'''from\stensorflow_datasets.*import\s([^\.\r\n]+)''' , _a) tfds_imports.extend(imp.strip() for imp in match.group(1).split(''',''')) lowerCAmelCase_ : str = '''from . import ''' + match.group(1) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(F"""Error converting {out_line.strip()}""") if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: lowerCAmelCase_ : Tuple = True out_lines.append(_a) if is_builder or "wmt" in f_name: # We create a new directory for each dataset lowerCAmelCase_ : Union[str, Any] = f_name.replace('''.py''' , '''''') lowerCAmelCase_ : List[str] = os.path.join(_a , _a) lowerCAmelCase_ : Optional[int] = os.path.join(_a , _a) os.makedirs(_a , exist_ok=_a) self._logger.info(F"""Adding directory {output_dir}""") imports_to_builder_map.update({imp: output_dir for imp in tfds_imports}) else: # Utilities will be moved at the end utils_files.append(_a) if needs_manual_update: with_manual_update.append(_a) with open(_a , '''w''' , encoding='''utf-8''') as f: f.writelines(_a) self._logger.info(F"""Converted in {output_file}""") for utils_file in utils_files: try: lowerCAmelCase_ : Union[str, Any] = os.path.basename(_a) lowerCAmelCase_ : List[str] = imports_to_builder_map[f_name.replace('''.py''' , '''''')] self._logger.info(F"""Moving {dest_folder} to {utils_file}""") shutil.copy(_a , _a) except KeyError: self._logger.error(F"""Cannot find destination folder for {utils_file}. Please copy manually.""") if with_manual_update: for file_path in with_manual_update: self._logger.warning( F"""You need to manually update file {file_path} to remove configurations using 'TextEncoderConfig'.""")
171
def SCREAMING_SNAKE_CASE ( __lowerCAmelCase = 100 ) -> int: snake_case__ = set() snake_case__ = 0 snake_case__ = n + 1 # maximum limit for a in range(2 , __lowerCAmelCase ): for b in range(2 , __lowerCAmelCase ): snake_case__ = a**b # calculates the current power collect_powers.add(__lowerCAmelCase ) # adds the result to the set return len(__lowerCAmelCase ) if __name__ == "__main__": print("""Number of terms """, solution(int(str(input()).strip())))
33
0
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import tensorflow as tf from transformers import AutoTokenizer, TFAutoModelForSeqaSeqLM @require_tf @require_sentencepiece @require_tokenizers class lowercase_ ( unittest.TestCase ): @slow def _snake_case ( self ) -> Dict: SCREAMING_SNAKE_CASE_ : List[str] =TFAutoModelForSeqaSeqLM.from_pretrained('''google/mt5-small''' ) SCREAMING_SNAKE_CASE_ : str =AutoTokenizer.from_pretrained('''google/mt5-small''' ) SCREAMING_SNAKE_CASE_ : List[Any] =tokenizer('''Hello there''' , return_tensors='''tf''' ).input_ids SCREAMING_SNAKE_CASE_ : Optional[int] =tokenizer('''Hi I am''' , return_tensors='''tf''' ).input_ids SCREAMING_SNAKE_CASE_ : Union[str, Any] =model(_a , labels=_a ).loss SCREAMING_SNAKE_CASE_ : Optional[int] =-tf.math.reduce_mean(_a ).numpy() SCREAMING_SNAKE_CASE_ : Tuple =-21.228_168 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 2e-4 )
443
from copy import deepcopy class __magic_name__ : '''simple docstring''' def __init__( self:int , _a:list[int] | None = None , _a:int | None = None ): if arr is None and size is not None: snake_case__ = size snake_case__ = [0] * size elif arr is not None: self.init(_a ) else: raise ValueError('''Either arr or size must be specified''' ) def SCREAMING_SNAKE_CASE__ ( self:Any , _a:list[int] ): snake_case__ = len(_a ) snake_case__ = deepcopy(_a ) for i in range(1 , self.size ): snake_case__ = self.next_(_a ) if j < self.size: self.tree[j] += self.tree[i] def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = self.tree[:] for i in range(self.size - 1 , 0 , -1 ): snake_case__ = self.next_(_a ) if j < self.size: arr[j] -= arr[i] return arr @staticmethod def SCREAMING_SNAKE_CASE__ ( _a:int ): return index + (index & (-index)) @staticmethod def SCREAMING_SNAKE_CASE__ ( _a:int ): return index - (index & (-index)) def SCREAMING_SNAKE_CASE__ ( self:List[Any] , _a:int , _a:int ): if index == 0: self.tree[0] += value return while index < self.size: self.tree[index] += value snake_case__ = self.next_(_a ) def SCREAMING_SNAKE_CASE__ ( self:Dict , _a:int , _a:int ): self.add(_a , value - self.get(_a ) ) def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] , _a:int ): if right == 0: return 0 snake_case__ = self.tree[0] right -= 1 # make right inclusive while right > 0: result += self.tree[right] snake_case__ = self.prev(_a ) return result def SCREAMING_SNAKE_CASE__ ( self:Dict , _a:int , _a:int ): return self.prefix(_a ) - self.prefix(_a ) def SCREAMING_SNAKE_CASE__ ( self:str , _a:int ): return self.query(_a , index + 1 ) def SCREAMING_SNAKE_CASE__ ( self:str , _a:int ): value -= self.tree[0] if value < 0: return -1 snake_case__ = 1 # Largest power of 2 <= size while j * 2 < self.size: j *= 2 snake_case__ = 0 while j > 0: if i + j < self.size and self.tree[i + j] <= value: value -= self.tree[i + j] i += j j //= 2 return i if __name__ == "__main__": import doctest doctest.testmod()
33
0
"""simple docstring""" def _lowerCamelCase ( __a = 100 ): SCREAMING_SNAKE_CASE_ = set() SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = n + 1 # maximum limit for a in range(2, __lowerCAmelCase ): for b in range(2, __lowerCAmelCase ): SCREAMING_SNAKE_CASE_ = a**b # calculates the current power collect_powers.add(__lowerCAmelCase ) # adds the result to the set return len(__lowerCAmelCase ) if __name__ == "__main__": print('Number of terms ', solution(int(str(input()).strip())))
626
from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class __magic_name__ : '''simple docstring''' __lowercase : int = BlenderbotConfig __lowercase : Any = {} __lowercase : Optional[Any] = 'gelu' def __init__( self:Tuple , _a:Optional[Any] , _a:Optional[Any]=13 , _a:Tuple=7 , _a:Union[str, Any]=True , _a:int=False , _a:int=99 , _a:Optional[int]=32 , _a:List[str]=2 , _a:List[str]=4 , _a:List[Any]=37 , _a:Any=0.1 , _a:int=0.1 , _a:List[Any]=20 , _a:List[str]=2 , _a:int=1 , _a:Dict=0 , ): snake_case__ = parent snake_case__ = batch_size snake_case__ = seq_length snake_case__ = is_training snake_case__ = use_labels snake_case__ = vocab_size snake_case__ = hidden_size snake_case__ = num_hidden_layers snake_case__ = num_attention_heads snake_case__ = intermediate_size snake_case__ = hidden_dropout_prob snake_case__ = attention_probs_dropout_prob snake_case__ = max_position_embeddings snake_case__ = eos_token_id snake_case__ = pad_token_id snake_case__ = bos_token_id def SCREAMING_SNAKE_CASE__ ( self:int ): snake_case__ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) snake_case__ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) snake_case__ = tf.concat([input_ids, eos_tensor] , axis=1 ) snake_case__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case__ = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) snake_case__ = prepare_blenderbot_inputs_dict(_a , _a , _a ) return config, inputs_dict def SCREAMING_SNAKE_CASE__ ( self:int , _a:Optional[Any] , _a:int ): snake_case__ = TFBlenderbotModel(config=_a ).get_decoder() snake_case__ = inputs_dict['''input_ids'''] snake_case__ = input_ids[:1, :] snake_case__ = inputs_dict['''attention_mask'''][:1, :] snake_case__ = inputs_dict['''head_mask'''] snake_case__ = 1 # first forward pass snake_case__ = model(_a , attention_mask=_a , head_mask=_a , use_cache=_a ) snake_case__ , snake_case__ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids snake_case__ = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case__ = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and snake_case__ = tf.concat([input_ids, next_tokens] , axis=-1 ) snake_case__ = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) snake_case__ = model(_a , attention_mask=_a )[0] snake_case__ = model(_a , attention_mask=_a , past_key_values=_a )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice snake_case__ = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) snake_case__ = output_from_no_past[:, -3:, random_slice_idx] snake_case__ = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_a , _a , rtol=1e-3 ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , ) -> Tuple: if attention_mask is None: snake_case__ = tf.cast(tf.math.not_equal(__lowerCAmelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: snake_case__ = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: snake_case__ = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: snake_case__ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: snake_case__ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class __magic_name__ (snake_case_ ,snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : List[str] = (TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () __lowercase : Any = (TFBlenderbotForConditionalGeneration,) if is_tf_available() else () __lowercase : Tuple = ( { 'conversational': TFBlenderbotForConditionalGeneration, 'feature-extraction': TFBlenderbotModel, 'summarization': TFBlenderbotForConditionalGeneration, 'text2text-generation': TFBlenderbotForConditionalGeneration, 'translation': TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) __lowercase : Any = True __lowercase : int = False __lowercase : int = False def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): snake_case__ = TFBlenderbotModelTester(self ) snake_case__ = ConfigTester(self , config_class=_a ) def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): snake_case__ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_a ) @require_tokenizers @require_tf class __magic_name__ (unittest.TestCase ): '''simple docstring''' __lowercase : Optional[int] = ['My friends are cool but they eat too many carbs.'] __lowercase : Optional[int] = 'facebook/blenderbot-400M-distill' @cached_property def SCREAMING_SNAKE_CASE__ ( self:Tuple ): return BlenderbotTokenizer.from_pretrained(self.model_name ) @cached_property def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): snake_case__ = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = self.tokenizer(self.src_text , return_tensors='''tf''' ) snake_case__ = self.model.generate( model_inputs.input_ids , ) snake_case__ = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=_a )[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
33
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tensorflow_text_available, is_torch_available A_ = { """configuration_ernie""": ["""ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ErnieConfig""", """ErnieOnnxConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ """ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST""", """ErnieForCausalLM""", """ErnieForMaskedLM""", """ErnieForMultipleChoice""", """ErnieForNextSentencePrediction""", """ErnieForPreTraining""", """ErnieForQuestionAnswering""", """ErnieForSequenceClassification""", """ErnieForTokenClassification""", """ErnieModel""", """ErniePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_ernie import ERNIE_PRETRAINED_CONFIG_ARCHIVE_MAP, ErnieConfig, ErnieOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ernie import ( ERNIE_PRETRAINED_MODEL_ARCHIVE_LIST, ErnieForCausalLM, ErnieForMaskedLM, ErnieForMultipleChoice, ErnieForNextSentencePrediction, ErnieForPreTraining, ErnieForQuestionAnswering, ErnieForSequenceClassification, ErnieForTokenClassification, ErnieModel, ErniePreTrainedModel, ) else: import sys A_ = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
29
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class __magic_name__ (unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = 0 def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): snake_case__ = AutoImageProcessor.from_pretrained('''openai/clip-vit-base-patch32''' ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:str ): with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = Path(_a ) / '''preprocessor_config.json''' snake_case__ = Path(_a ) / '''config.json''' json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) json.dump({'''model_type''': '''clip'''} , open(_a , '''w''' ) ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = Path(_a ) / '''preprocessor_config.json''' snake_case__ = Path(_a ) / '''config.json''' json.dump( {'''feature_extractor_type''': '''CLIPFeatureExtractor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) json.dump({'''model_type''': '''clip'''} , open(_a , '''w''' ) ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = CLIPConfig() # Create a dummy config file with image_proceesor_type snake_case__ = Path(_a ) / '''preprocessor_config.json''' snake_case__ = Path(_a ) / '''config.json''' json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) json.dump({'''model_type''': '''clip'''} , open(_a , '''w''' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally snake_case__ = AutoImageProcessor.from_pretrained(_a ).to_dict() config_dict.pop('''image_processor_type''' ) snake_case__ = CLIPImageProcessor(**_a ) # save in new folder model_config.save_pretrained(_a ) config.save_pretrained(_a ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) # make sure private variable is not incorrectly saved snake_case__ = json.loads(config.to_json_string() ) self.assertTrue('''_processor_class''' not in dict_as_saved ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = Path(_a ) / '''preprocessor_config.json''' json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): with self.assertRaisesRegex( _a , '''clip-base is not a local folder and is not a valid model identifier''' ): snake_case__ = AutoImageProcessor.from_pretrained('''clip-base''' ) def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): with self.assertRaisesRegex( _a , r'''aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)''' ): snake_case__ = AutoImageProcessor.from_pretrained(_a , revision='''aaaaaa''' ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): with self.assertRaisesRegex( _a , '''hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.''' , ): snake_case__ = AutoImageProcessor.from_pretrained('''hf-internal-testing/config-no-model''' ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(_a ): snake_case__ = AutoImageProcessor.from_pretrained('''hf-internal-testing/test_dynamic_image_processor''' ) # If remote code is disabled, we can't load this config. with self.assertRaises(_a ): snake_case__ = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=_a ) snake_case__ = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=_a ) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_a ) snake_case__ = AutoImageProcessor.from_pretrained(_a , trust_remote_code=_a ) self.assertEqual(reloaded_image_processor.__class__.__name__ , '''NewImageProcessor''' ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): try: AutoConfig.register('''custom''' , _a ) AutoImageProcessor.register(_a , _a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_a ): AutoImageProcessor.register(_a , _a ) with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = Path(_a ) / '''preprocessor_config.json''' snake_case__ = Path(_a ) / '''config.json''' json.dump( {'''feature_extractor_type''': '''CLIPFeatureExtractor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) json.dump({'''model_type''': '''clip'''} , open(_a , '''w''' ) ) snake_case__ = CustomImageProcessor.from_pretrained(_a ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_a ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : List[str] = True try: AutoConfig.register('''custom''' , _a ) AutoImageProcessor.register(_a , _a ) # If remote code is not set, the default is to use local snake_case__ = AutoImageProcessor.from_pretrained('''hf-internal-testing/test_dynamic_image_processor''' ) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. snake_case__ = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=_a ) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub snake_case__ = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=_a ) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''' ) self.assertTrue(not hasattr(_a , '''is_local''' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
33
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _a = { """configuration_jukebox""": [ """JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP""", """JukeboxConfig""", """JukeboxPriorConfig""", """JukeboxVQVAEConfig""", ], """tokenization_jukebox""": ["""JukeboxTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a = [ """JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST""", """JukeboxModel""", """JukeboxPreTrainedModel""", """JukeboxVQVAE""", """JukeboxPrior""", ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys _a = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
19
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase__ : int = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase=False ) -> int: snake_case__ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" snake_case__ = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False ) -> Dict: for i in range(config.num_hidden_layers ): if base_model: snake_case__ = '''''' else: snake_case__ = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) snake_case__ = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) snake_case__ = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict snake_case__ = in_proj_weight[ : config.hidden_size, : ] snake_case__ = in_proj_bias[: config.hidden_size] snake_case__ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] snake_case__ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] snake_case__ = in_proj_weight[ -config.hidden_size :, : ] snake_case__ = in_proj_bias[-config.hidden_size :] def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> Optional[Any]: snake_case__ = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Union[str, Any]: snake_case__ = dct.pop(__lowerCAmelCase ) snake_case__ = val def SCREAMING_SNAKE_CASE ( ) -> str: snake_case__ = '''http://images.cocodataset.org/val2017/000000039769.jpg''' snake_case__ = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> Dict: snake_case__ = ViTConfig() snake_case__ = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": snake_case__ = True snake_case__ = int(vit_name[-12:-10] ) snake_case__ = int(vit_name[-9:-6] ) else: snake_case__ = 1000 snake_case__ = '''huggingface/label-files''' snake_case__ = '''imagenet-1k-id2label.json''' snake_case__ = json.load(open(hf_hub_download(__lowerCAmelCase , __lowerCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) snake_case__ = {int(__lowerCAmelCase ): v for k, v in idalabel.items()} snake_case__ = idalabel snake_case__ = {v: k for k, v in idalabel.items()} snake_case__ = int(vit_name[-6:-4] ) snake_case__ = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith('''tiny''' ): snake_case__ = 192 snake_case__ = 768 snake_case__ = 12 snake_case__ = 3 elif vit_name[9:].startswith('''small''' ): snake_case__ = 384 snake_case__ = 1536 snake_case__ = 12 snake_case__ = 6 else: pass else: if vit_name[4:].startswith('''small''' ): snake_case__ = 768 snake_case__ = 2304 snake_case__ = 8 snake_case__ = 8 elif vit_name[4:].startswith('''base''' ): pass elif vit_name[4:].startswith('''large''' ): snake_case__ = 1024 snake_case__ = 4096 snake_case__ = 24 snake_case__ = 16 elif vit_name[4:].startswith('''huge''' ): snake_case__ = 1280 snake_case__ = 5120 snake_case__ = 32 snake_case__ = 16 # load original model from timm snake_case__ = timm.create_model(__lowerCAmelCase , pretrained=__lowerCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys snake_case__ = timm_model.state_dict() if base_model: remove_classification_head_(__lowerCAmelCase ) snake_case__ = create_rename_keys(__lowerCAmelCase , __lowerCAmelCase ) for src, dest in rename_keys: rename_key(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) read_in_q_k_v(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # load HuggingFace model if vit_name[-5:] == "in21k": snake_case__ = ViTModel(__lowerCAmelCase ).eval() else: snake_case__ = ViTForImageClassification(__lowerCAmelCase ).eval() model.load_state_dict(__lowerCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: snake_case__ = DeiTImageProcessor(size=config.image_size ) else: snake_case__ = ViTImageProcessor(size=config.image_size ) snake_case__ = image_processor(images=prepare_img() , return_tensors='''pt''' ) snake_case__ = encoding['''pixel_values'''] snake_case__ = model(__lowerCAmelCase ) if base_model: snake_case__ = timm_model.forward_features(__lowerCAmelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__lowerCAmelCase , outputs.pooler_output , atol=1e-3 ) else: snake_case__ = timm_model(__lowerCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__lowerCAmelCase , outputs.logits , atol=1e-3 ) Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) print(F"""Saving model {vit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCAmelCase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": lowerCamelCase__ : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--vit_name""", default="""vit_base_patch16_224""", type=str, help="""Name of the ViT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) lowerCamelCase__ : str = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
33
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase__ = logging.get_logger(__name__) class lowercase_ ( snake_case_ ): '''simple docstring''' __snake_case = 'encoder-decoder' __snake_case = True def __init__( self : Dict , **__UpperCAmelCase : Optional[Any] ) ->List[str]: """simple docstring""" super().__init__(**_a ) assert ( "encoder" in kwargs and "decoder" in kwargs ), "Config has to be initialized with encoder and decoder config" a = kwargs.pop('''encoder''' ) a = encoder_config.pop('''model_type''' ) a = kwargs.pop('''decoder''' ) a = decoder_config.pop('''model_type''' ) from ..auto.configuration_auto import AutoConfig a = AutoConfig.for_model(_a , **_a ) a = AutoConfig.for_model(_a , **_a ) a = True @classmethod def __lowerCAmelCase ( cls : List[Any] , __UpperCAmelCase : PretrainedConfig , __UpperCAmelCase : PretrainedConfig , **__UpperCAmelCase : int ) ->Union[str, Any]: """simple docstring""" logger.info('''Set `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config''' ) a = True a = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **_a ) def __lowerCAmelCase ( self : Tuple ) ->Dict: """simple docstring""" a = copy.deepcopy(self.__dict__ ) a = self.encoder.to_dict() a = self.decoder.to_dict() a = self.__class__.model_type return output
117
import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : List[str] = ['image_processor', 'tokenizer'] __lowercase : str = 'AutoImageProcessor' __lowercase : Dict = 'AutoTokenizer' def __init__( self:int , _a:List[str]=None , _a:Optional[Any]=None , **_a:List[str] ): snake_case__ = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _a , ) snake_case__ = kwargs.pop('''feature_extractor''' ) snake_case__ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_a , _a ) snake_case__ = self.image_processor snake_case__ = False def __call__( self:Optional[int] , *_a:str , **_a:int ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*_a , **_a ) snake_case__ = kwargs.pop('''images''' , _a ) snake_case__ = kwargs.pop('''text''' , _a ) if len(_a ) > 0: snake_case__ = args[0] snake_case__ = args[1:] if images is None and text is None: raise ValueError('''You need to specify either an `images` or `text` input to process.''' ) if images is not None: snake_case__ = self.image_processor(_a , *_a , **_a ) if text is not None: snake_case__ = self.tokenizer(_a , **_a ) if text is None: return inputs elif images is None: return encodings else: snake_case__ = encodings['''input_ids'''] return inputs def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] , *_a:Union[str, Any] , **_a:Any ): return self.tokenizer.batch_decode(*_a , **_a ) def SCREAMING_SNAKE_CASE__ ( self:Tuple , *_a:Union[str, Any] , **_a:Optional[int] ): return self.tokenizer.decode(*_a , **_a ) @contextmanager def SCREAMING_SNAKE_CASE__ ( self:Tuple ): warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your images inputs, or in a separate call.''' ) snake_case__ = True snake_case__ = self.tokenizer yield snake_case__ = self.image_processor snake_case__ = False def SCREAMING_SNAKE_CASE__ ( self:List[str] , _a:Dict , _a:Dict=False , _a:Optional[int]=None ): if added_vocab is None: snake_case__ = self.tokenizer.get_added_vocab() snake_case__ = {} while tokens: snake_case__ = re.search(r'''<s_(.*?)>''' , _a , re.IGNORECASE ) if start_token is None: break snake_case__ = start_token.group(1 ) snake_case__ = re.search(rF"""</s_{key}>""" , _a , re.IGNORECASE ) snake_case__ = start_token.group() if end_token is None: snake_case__ = tokens.replace(_a , '''''' ) else: snake_case__ = end_token.group() snake_case__ = re.escape(_a ) snake_case__ = re.escape(_a ) snake_case__ = re.search(F"""{start_token_escaped}(.*?){end_token_escaped}""" , _a , re.IGNORECASE ) if content is not None: snake_case__ = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node snake_case__ = self.tokenajson(_a , is_inner_value=_a , added_vocab=_a ) if value: if len(_a ) == 1: snake_case__ = value[0] snake_case__ = value else: # leaf nodes snake_case__ = [] for leaf in content.split(r'''<sep/>''' ): snake_case__ = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": snake_case__ = leaf[1:-2] # for categorical special tokens output[key].append(_a ) if len(output[key] ) == 1: snake_case__ = output[key][0] snake_case__ = tokens[tokens.find(_a ) + len(_a ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:] , is_inner_value=_a , added_vocab=_a ) if len(_a ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , _a , ) return self.image_processor_class @property def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , _a , ) return self.image_processor
33
0
import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): _lowerCamelCase : str = True from torch.cuda.amp import autocast _lowerCamelCase : List[Any] = logging.getLogger(__name__) @dataclass class lowerCAmelCase__ : '''simple docstring''' lowercase_ = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) lowercase_ = field( default=snake_case_ , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) lowercase_ = field( default=snake_case_ , metadata={"""help""": """Whether to freeze the feature extractor layers of the model."""} ) lowercase_ = field( default=snake_case_ , metadata={"""help""": """Whether to log verbose messages or not."""} , ) lowercase_ = field( default=2.0 , metadata={"""help""": """Maximum temperature for gumbel softmax."""} ) lowercase_ = field( default=0.5 , metadata={"""help""": """Minimum temperature for gumbel softmax."""} ) lowercase_ = field( default=0.99_99_95 , metadata={"""help""": """Decay of gumbel temperature during training."""} ) def A__ ( __A : Tuple , __A : Tuple ) ->Dict: logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''' , datefmt='''%m/%d/%Y %H:%M:%S''' , handlers=[logging.StreamHandler(sys.stdout )] , ) __A =logging.WARNING if model_args.verbose_logging: __A =logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank ): __A =logging.INFO logger.setLevel(__lowerCAmelCase ) @dataclass class lowerCAmelCase__ : '''simple docstring''' lowercase_ = field( default=snake_case_ , metadata={"""help""": """The name of the dataset to use (via the datasets library)."""} ) lowercase_ = field( default=snake_case_ , metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) lowercase_ = field( default="""train""" , metadata={ """help""": """The name of the training data set split to use (via the datasets library). Defaults to \'train\'""" } , ) lowercase_ = field( default="""validation""" , metadata={ """help""": ( """The name of the validation data set split to use (via the datasets library). Defaults to \'validation\'""" ) } , ) lowercase_ = field( default="""file""" , metadata={"""help""": """Column in the dataset that contains speech file path. Defaults to \'file\'"""} , ) lowercase_ = field( default=snake_case_ , metadata={"""help""": """Overwrite the cached preprocessed datasets or not."""} ) lowercase_ = field( default=1 , metadata={ """help""": """The percentage of the train set used as validation set in case there\'s no validation split""" } , ) lowercase_ = field( default=snake_case_ , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) lowercase_ = field( default=20.0 , metadata={"""help""": """Filter audio files that are longer than `max_duration_in_seconds` seconds"""} ) @dataclass class lowerCAmelCase__ : '''simple docstring''' lowercase_ = 42 lowercase_ = 42 lowercase_ = "longest" lowercase_ = None lowercase_ = None def __call__( self , lowercase__ ): '''simple docstring''' __A =self.feature_extractor.pad( _a , max_length=self.max_length , padding=self.padding , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='''pt''' , ) __A =self.model._get_feat_extract_output_lengths(batch['''input_values'''].shape[-1] ) __A =batch['''input_values'''].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula __A =self.model._get_feat_extract_output_lengths(batch['''attention_mask'''].sum(-1 ) ).to( torch.long ) __A =torch.zeros( (batch_size, mask_indices_seq_length) , dtype=torch.long , device=batch['''input_values'''].device ) # these two operations makes sure that all values # before the output lengths indices are attended to __A =1 __A =attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices __A =_compute_mask_indices( (batch_size, mask_indices_seq_length) , self.model.config.mask_time_prob , self.model.config.mask_time_length , attention_mask=_a , min_masks=2 , ) return batch class lowerCAmelCase__ ( snake_case_ ): '''simple docstring''' def __init__( self , *lowercase__ , lowercase__=1 , lowercase__=0 , lowercase__=1.0 , **lowercase__ ): '''simple docstring''' super().__init__(*_a , **_a ) __A =0 __A =max_gumbel_temp __A =min_gumbel_temp __A =gumbel_temp_decay def __UpperCamelCase ( self , lowercase__ , lowercase__ ): '''simple docstring''' model.train() __A =self._prepare_inputs(_a ) if self.use_amp: with autocast(): __A =self.compute_loss(_a , _a ) else: __A =self.compute_loss(_a , _a ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": __A =loss.mean() elif model.module.config.ctc_loss_reduction == "sum": __A =loss.sum() / (inputs['''mask_time_indices''']).sum() else: raise ValueError(f'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: __A =loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(_a ).backward() elif self.use_apex: with amp.scale_loss(_a , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(_a ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step , self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step , self.min_gumbel_temp ) ) return loss.detach() def A__ ( ) ->Union[str, Any]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. __A =HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) __A , __A , __A =parser.parse_args_into_dataclasses() configure_logger(__lowerCAmelCase , __lowerCAmelCase ) # Downloading and loading a dataset from the hub. __A =load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" __A =DatasetDict() __A =load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) __A =load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" __A =DatasetDict() __A =load_dataset( data_args.dataset_name , data_args.dataset_config_name , split='''validation''' , cache_dir=model_args.cache_dir , ) __A =load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported __A =WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=__lowerCAmelCase ) def prepare_dataset(__A : Dict ): # check that all files have the correct sampling rate __A , __A =librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate ) return batch # load audio files into numpy arrays __A =datasets.map( __lowerCAmelCase , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets['''train'''].column_names ) # filter audio files that are too long __A =vectorized_datasets.filter( lambda __A : len(data['''speech'''] ) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate ) ) def normalize(__A : Optional[int] ): return feature_extractor(batch['''speech'''] , sampling_rate=feature_extractor.sampling_rate ) # normalize and transform to `BatchFeatures` __A =vectorized_datasets.map( __lowerCAmelCase , batched=__lowerCAmelCase , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets['''train'''].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 __A =WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( '''PreTraining is only supported for ``config.do_stable_layer_norm=True`` and''' ''' ``config.feat_extract_norm=\'layer\'''' ) __A =WavaVecaForPreTraining(__lowerCAmelCase ) __A =DataCollatorForWavaVecaPretraining(model=__lowerCAmelCase , feature_extractor=__lowerCAmelCase ) __A =WavaVecaPreTrainer( model=__lowerCAmelCase , data_collator=__lowerCAmelCase , args=__lowerCAmelCase , train_dataset=vectorized_datasets['''train'''] , eval_dataset=vectorized_datasets['''validation'''] , tokenizer=__lowerCAmelCase , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
184
import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class __magic_name__ : '''simple docstring''' def __init__( self:Optional[Any] , _a:int , _a:str=3 , _a:Optional[int]=32 , _a:Optional[Any]=3 , _a:Tuple=10 , _a:List[Any]=[8, 16, 32, 64] , _a:str=[1, 1, 2, 1] , _a:Any=True , _a:List[Any]=True , _a:List[str]="relu" , _a:int=3 , _a:Tuple=None , _a:Tuple=["stage2", "stage3", "stage4"] , _a:List[Any]=[2, 3, 4] , _a:Union[str, Any]=1 , ): snake_case__ = parent snake_case__ = batch_size snake_case__ = image_size snake_case__ = num_channels snake_case__ = embeddings_size snake_case__ = hidden_sizes snake_case__ = depths snake_case__ = is_training snake_case__ = use_labels snake_case__ = hidden_act snake_case__ = num_labels snake_case__ = scope snake_case__ = len(_a ) snake_case__ = out_features snake_case__ = out_indices snake_case__ = num_groups def SCREAMING_SNAKE_CASE__ ( self:int ): snake_case__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case__ = None if self.use_labels: snake_case__ = ids_tensor([self.batch_size] , self.num_labels ) snake_case__ = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def SCREAMING_SNAKE_CASE__ ( self:Any , _a:Optional[int] , _a:Tuple , _a:int ): snake_case__ = BitModel(config=_a ) model.to(_a ) model.eval() snake_case__ = model(_a ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def SCREAMING_SNAKE_CASE__ ( self:int , _a:Tuple , _a:Any , _a:Union[str, Any] ): snake_case__ = self.num_labels snake_case__ = BitForImageClassification(_a ) model.to(_a ) model.eval() snake_case__ = model(_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self:str , _a:str , _a:List[str] , _a:Any ): snake_case__ = BitBackbone(config=_a ) model.to(_a ) model.eval() snake_case__ = model(_a ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None snake_case__ = None snake_case__ = BitBackbone(config=_a ) model.to(_a ) model.eval() snake_case__ = model(_a ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = self.prepare_config_and_inputs() snake_case__ , snake_case__ , snake_case__ = config_and_inputs snake_case__ = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __magic_name__ (snake_case_ ,snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : Any = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () __lowercase : int = ( {'feature-extraction': BitModel, 'image-classification': BitForImageClassification} if is_torch_available() else {} ) __lowercase : Tuple = False __lowercase : Optional[Any] = False __lowercase : str = False __lowercase : Tuple = False __lowercase : Tuple = False def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): snake_case__ = BitModelTester(self ) snake_case__ = ConfigTester(self , config_class=_a , has_text_modality=_a ) def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): return @unittest.skip(reason='''Bit does not output attentions''' ) def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): pass @unittest.skip(reason='''Bit does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): pass @unittest.skip(reason='''Bit does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): pass def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ , snake_case__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case__ = model_class(_a ) snake_case__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case__ = [*signature.parameters.keys()] snake_case__ = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _a ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): snake_case__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_a ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ , snake_case__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case__ = model_class(config=_a ) for name, module in model.named_modules(): if isinstance(_a , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): def check_hidden_states_output(_a:List[Any] , _a:int , _a:Union[str, Any] ): snake_case__ = model_class(_a ) model.to(_a ) model.eval() with torch.no_grad(): snake_case__ = model(**self._prepare_for_class(_a , _a ) ) snake_case__ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states snake_case__ = self.model_tester.num_stages self.assertEqual(len(_a ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) snake_case__ , snake_case__ = self.model_tester.prepare_config_and_inputs_for_common() snake_case__ = ['''preactivation''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: snake_case__ = layer_type snake_case__ = True check_hidden_states_output(_a , _a , _a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case__ = True check_hidden_states_output(_a , _a , _a ) @unittest.skip(reason='''Bit does not use feedforward chunking''' ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): pass def SCREAMING_SNAKE_CASE__ ( self:List[str] ): snake_case__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_a ) @slow def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case__ = BitModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def SCREAMING_SNAKE_CASE ( ) -> Any: snake_case__ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class __magic_name__ (unittest.TestCase ): '''simple docstring''' @cached_property def SCREAMING_SNAKE_CASE__ ( self:Tuple ): return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): snake_case__ = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(_a ) snake_case__ = self.default_image_processor snake_case__ = prepare_img() snake_case__ = image_processor(images=_a , return_tensors='''pt''' ).to(_a ) # forward pass with torch.no_grad(): snake_case__ = model(**_a ) # verify the logits snake_case__ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , _a ) snake_case__ = torch.tensor([[-0.6526, -0.5263, -1.4398]] ).to(_a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _a , atol=1e-4 ) ) @require_torch class __magic_name__ (snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : Optional[Any] = (BitBackbone,) if is_torch_available() else () __lowercase : int = BitConfig __lowercase : Any = False def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = BitModelTester(self )
33
0
import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class snake_case_ ( snake_case_ , unittest.TestCase ): '''simple docstring''' __UpperCamelCase = OpenAIGPTTokenizer __UpperCamelCase = OpenAIGPTTokenizerFast __UpperCamelCase = True __UpperCamelCase = False def UpperCAmelCase ( self : List[str] ) -> Union[str, Any]: '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __lowercase = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] __lowercase = dict(zip(_a , range(len(_a ) ) ) ) __lowercase = ['#version: 0.2', 'l o', 'lo w', 'e r</w>', ''] __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) __lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(_a ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(_a ) ) def UpperCAmelCase ( self : int , __lowerCamelCase : Dict ) -> Dict: '''simple docstring''' return "lower newer", "lower newer" def UpperCAmelCase ( self : Optional[Any] ) -> Union[str, Any]: '''simple docstring''' __lowercase = OpenAIGPTTokenizer(self.vocab_file , self.merges_file ) __lowercase = 'lower' __lowercase = ['low', 'er</w>'] __lowercase = tokenizer.tokenize(_a ) self.assertListEqual(_a , _a ) __lowercase = tokens + ['<unk>'] __lowercase = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , _a ) def UpperCAmelCase ( self : Optional[int] , __lowerCamelCase : Dict=15 ) -> List[Any]: '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): __lowercase = self.rust_tokenizer_class.from_pretrained(_a , **_a ) # Simple input __lowercase = 'This is a simple input' __lowercase = ['This is a simple input 1', 'This is a simple input 2'] __lowercase = ('This is a simple input', 'This is a pair') __lowercase = [ ('This is a simple input 1', 'This is a simple input 2'), ('This is a simple pair 1', 'This is a simple pair 2'), ] # Simple input tests self.assertRaises(_a , tokenizer_r.encode , _a , max_length=_a , padding='max_length' ) # Simple input self.assertRaises(_a , tokenizer_r.encode_plus , _a , max_length=_a , padding='max_length' ) # Simple input self.assertRaises( _a , tokenizer_r.batch_encode_plus , _a , max_length=_a , padding='max_length' , ) # Pair input self.assertRaises(_a , tokenizer_r.encode , _a , max_length=_a , padding='max_length' ) # Pair input self.assertRaises(_a , tokenizer_r.encode_plus , _a , max_length=_a , padding='max_length' ) # Pair input self.assertRaises( _a , tokenizer_r.batch_encode_plus , _a , max_length=_a , padding='max_length' , ) def UpperCAmelCase ( self : Any ) -> Union[str, Any]: '''simple docstring''' pass @require_ftfy @require_spacy @require_tokenizers class snake_case_ ( snake_case_ ): '''simple docstring''' pass
375
import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging lowerCamelCase__ : Any = """\ """ lowerCamelCase__ : List[str] = """ Perplexity (PPL) is one of the most common metrics for evaluating language models. It is defined as the exponentiated average negative log-likelihood of a sequence. For more information, see https://huggingface.co/docs/transformers/perplexity """ lowerCamelCase__ : Any = """ Args: model_id (str): model used for calculating Perplexity NOTE: Perplexity can only be calculated for causal language models. This includes models such as gpt2, causal variations of bert, causal versions of t5, and more (the full list can be found in the AutoModelForCausalLM documentation here: https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM ) input_texts (list of str): input text, each separate text snippet is one list entry. batch_size (int): the batch size to run texts through the model. Defaults to 16. add_start_token (bool): whether to add the start token to the texts, so the perplexity can include the probability of the first word. Defaults to True. device (str): device to run on, defaults to 'cuda' when available Returns: perplexity: dictionary containing the perplexity scores for the texts in the input list, as well as the mean perplexity. If one of the input texts is longer than the max input length of the model, then it is truncated to the max length for the perplexity computation. Examples: Example 1: >>> perplexity = datasets.load_metric(\"perplexity\") >>> input_texts = [\"lorem ipsum\", \"Happy Birthday!\", \"Bienvenue\"] >>> results = perplexity.compute(model_id='gpt2', ... add_start_token=False, ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) ['perplexities', 'mean_perplexity'] >>> print(round(results[\"mean_perplexity\"], 2)) 78.22 >>> print(round(results[\"perplexities\"][0], 2)) 11.11 Example 2: >>> perplexity = datasets.load_metric(\"perplexity\") >>> input_texts = datasets.load_dataset(\"wikitext\", ... \"wikitext-2-raw-v1\", ... split=\"test\")[\"text\"][:50] # doctest:+ELLIPSIS [...] >>> input_texts = [s for s in input_texts if s!=''] >>> results = perplexity.compute(model_id='gpt2', ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) ['perplexities', 'mean_perplexity'] >>> print(round(results[\"mean_perplexity\"], 2)) 60.35 >>> print(round(results[\"perplexities\"][0], 2)) 81.12 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class __magic_name__ (datasets.Metric ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self:int ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''input_texts''': datasets.Value('''string''' ), } ) , reference_urls=['''https://huggingface.co/docs/transformers/perplexity'''] , ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] , _a:int , _a:List[Any] , _a:int = 16 , _a:bool = True , _a:Any=None ): if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": snake_case__ = '''cuda''' else: snake_case__ = '''cuda''' if torch.cuda.is_available() else '''cpu''' snake_case__ = AutoModelForCausalLM.from_pretrained(_a ) snake_case__ = model.to(_a ) snake_case__ = AutoTokenizer.from_pretrained(_a ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: snake_case__ = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(_a ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({'''pad_token''': existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" snake_case__ = model.config.max_length - 1 else: snake_case__ = model.config.max_length snake_case__ = tokenizer( _a , add_special_tokens=_a , padding=_a , truncation=_a , max_length=_a , return_tensors='''pt''' , return_attention_mask=_a , ).to(_a ) snake_case__ = encodings['''input_ids'''] snake_case__ = encodings['''attention_mask'''] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." snake_case__ = [] snake_case__ = CrossEntropyLoss(reduction='''none''' ) for start_index in logging.tqdm(range(0 , len(_a ) , _a ) ): snake_case__ = min(start_index + batch_size , len(_a ) ) snake_case__ = encoded_texts[start_index:end_index] snake_case__ = attn_masks[start_index:end_index] if add_start_token: snake_case__ = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(_a ) snake_case__ = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) snake_case__ = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(_a ), attn_mask] , dim=1 ) snake_case__ = encoded_batch with torch.no_grad(): snake_case__ = model(_a , attention_mask=_a ).logits snake_case__ = out_logits[..., :-1, :].contiguous() snake_case__ = labels[..., 1:].contiguous() snake_case__ = attn_mask[..., 1:].contiguous() snake_case__ = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , _a ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(_a )}
33
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) lowercase : List[Any] = { """configuration_efficientformer""": [ """EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """EfficientFormerConfig""", ] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : str = ["""EfficientFormerImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Optional[int] = [ """EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """EfficientFormerForImageClassification""", """EfficientFormerForImageClassificationWithTeacher""", """EfficientFormerModel""", """EfficientFormerPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase : Tuple = [ """TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFEfficientFormerForImageClassification""", """TFEfficientFormerForImageClassificationWithTeacher""", """TFEfficientFormerModel""", """TFEfficientFormerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_efficientformer import EFFICIENTFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, EfficientFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_efficientformer import EfficientFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_efficientformer import ( EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, EfficientFormerForImageClassification, EfficientFormerForImageClassificationWithTeacher, EfficientFormerModel, EfficientFormerPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, TFEfficientFormerPreTrainedModel, ) else: import sys lowercase : int = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
568
import os from datetime import datetime as dt from github import Github lowerCamelCase__ : int = [ """good first issue""", """good second issue""", """good difficult issue""", """enhancement""", """new pipeline/model""", """new scheduler""", """wip""", ] def SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: snake_case__ = Github(os.environ['''GITHUB_TOKEN'''] ) snake_case__ = g.get_repo('''huggingface/diffusers''' ) snake_case__ = repo.get_issues(state='''open''' ) for issue in open_issues: snake_case__ = sorted(issue.get_comments() , key=lambda __lowerCAmelCase : i.created_at , reverse=__lowerCAmelCase ) snake_case__ = comments[0] if len(__lowerCAmelCase ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Closes the issue after 7 days of inactivity since the Stalebot notification. issue.edit(state='''closed''' ) elif ( "stale" in issue.get_labels() and last_comment is not None and last_comment.user.login != "github-actions[bot]" ): # Opens the issue if someone other than Stalebot commented. issue.edit(state='''open''' ) issue.remove_from_labels('''stale''' ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Post a Stalebot notification after 23 days of inactivity. issue.create_comment( '''This issue has been automatically marked as stale because it has not had ''' '''recent activity. If you think this still needs to be addressed ''' '''please comment on this thread.\n\nPlease note that issues that do not follow the ''' '''[contributing guidelines](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md) ''' '''are likely to be ignored.''' ) issue.add_to_labels('''stale''' ) if __name__ == "__main__": main()
33
0
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( UniSpeechConfig, UniSpeechForCTC, UniSpeechForPreTraining, WavaVecaFeatureExtractor, WavaVecaPhonemeCTCTokenizer, WavaVecaProcessor, logging, ) logging.set_verbosity_info() __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { """post_extract_proj""": """feature_projection.projection""", """encoder.pos_conv.0""": """encoder.pos_conv_embed.conv""", """self_attn.k_proj""": """encoder.layers.*.attention.k_proj""", """self_attn.v_proj""": """encoder.layers.*.attention.v_proj""", """self_attn.q_proj""": """encoder.layers.*.attention.q_proj""", """self_attn.out_proj""": """encoder.layers.*.attention.out_proj""", """self_attn_layer_norm""": """encoder.layers.*.layer_norm""", """fc1""": """encoder.layers.*.feed_forward.intermediate_dense""", """fc2""": """encoder.layers.*.feed_forward.output_dense""", """final_layer_norm""": """encoder.layers.*.final_layer_norm""", """encoder.layer_norm""": """encoder.layer_norm""", """w2v_model.layer_norm""": """feature_projection.layer_norm""", """quantizer.weight_proj""": """quantizer.weight_proj""", """quantizer.vars""": """quantizer.codevectors""", """project_q""": """project_q""", """final_proj""": """project_hid""", """w2v_encoder.proj""": """ctc_proj""", """mask_emb""": """masked_spec_embed""", } __lowerCAmelCase = [ """ctc_proj""", """quantizer.weight_proj""", """quantizer.codevectors""", """project_q""", """project_hid""", ] def _lowercase ( a__ : Tuple , a__ : Dict , a__ : Dict , a__ : List[str] , a__ : Union[str, Any] , a__ : Dict ) -> str: """simple docstring""" for attribute in key.split("." ): if is_finetuned: if attribute in ["quantizer", "project_q", "project_hid"]: # those layers are only relevant for pretraining and should be dropped return if attribute == "ctc_proj": # we should rename `ctc_proj` to `lm_head` for fine-tuned phoneme models _UpperCamelCase = "lm_head" _UpperCamelCase = getattr(__lowerCAmelCase , __lowerCAmelCase ) if weight_type is not None: _UpperCamelCase = getattr(__lowerCAmelCase , __lowerCAmelCase ).shape else: _UpperCamelCase = hf_pointer.shape assert hf_shape == value.shape, ( f'''Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be''' f''' {value.shape} for {full_name}''' ) if weight_type == "weight": _UpperCamelCase = value elif weight_type == "weight_g": _UpperCamelCase = value elif weight_type == "weight_v": _UpperCamelCase = value elif weight_type == "bias": _UpperCamelCase = value else: _UpperCamelCase = value logger.info(f'''{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.''' ) def _lowercase ( a__ : List[str] , a__ : Dict , a__ : Optional[Any] ) -> Tuple: """simple docstring""" _UpperCamelCase = [] _UpperCamelCase = fairseq_model.state_dict() _UpperCamelCase = hf_model.unispeech.feature_extractor for name, value in fairseq_dict.items(): _UpperCamelCase = False if "conv_layers" in name: load_conv_layer( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , hf_model.config.feat_extract_norm == "group" , ) _UpperCamelCase = True else: for key, mapped_key in MAPPING.items(): _UpperCamelCase = "unispeech." + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split("w2v_model." )[-1] == name.split("." )[0]: _UpperCamelCase = True if "*" in mapped_key: _UpperCamelCase = name.split(__lowerCAmelCase )[0].split("." )[-2] _UpperCamelCase = mapped_key.replace("*" , __lowerCAmelCase ) if "weight_g" in name: _UpperCamelCase = "weight_g" elif "weight_v" in name: _UpperCamelCase = "weight_v" elif "bias" in name: _UpperCamelCase = "bias" elif "weight" in name: # TODO: don't match quantizer.weight_proj _UpperCamelCase = "weight" else: _UpperCamelCase = None set_recursively(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) continue if not is_used: unused_weights.append(__lowerCAmelCase ) logger.warning(f'''Unused weights: {unused_weights}''' ) def _lowercase ( a__ : Optional[Any] , a__ : Optional[int] , a__ : Dict , a__ : List[Any] , a__ : str ) -> Optional[Any]: """simple docstring""" _UpperCamelCase = full_name.split("conv_layers." )[-1] _UpperCamelCase = name.split("." ) _UpperCamelCase = int(items[0] ) _UpperCamelCase = int(items[1] ) if type_id == 0: if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.bias.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.''' ) _UpperCamelCase = value logger.info(f'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].conv.weight.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.''' ) _UpperCamelCase = value logger.info(f'''Feat extract conv layer {layer_id} was initialized from {full_name}.''' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape, ( f'''{full_name} has size {value.shape}, but {feature_extractor[layer_id].layer_norm.bias.data.shape} was''' " found." ) _UpperCamelCase = value logger.info(f'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) elif "weight" in name: assert value.shape == feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape, ( f'''{full_name} has size {value.shape}, but''' f''' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.''' ) _UpperCamelCase = value logger.info(f'''Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.''' ) else: unused_weights.append(__lowerCAmelCase ) @torch.no_grad() def _lowercase ( a__ : str , a__ : Dict , a__ : Optional[int]=None , a__ : Tuple=None , a__ : Any=True ) -> Optional[Any]: """simple docstring""" if config_path is not None: _UpperCamelCase = UniSpeechConfig.from_pretrained(__lowerCAmelCase ) else: _UpperCamelCase = UniSpeechConfig() if is_finetuned: if dict_path: _UpperCamelCase = Dictionary.load_from_json(__lowerCAmelCase ) # important change bos & pad token id since CTC symbol is <pad> and # not <s> as in fairseq _UpperCamelCase = target_dict.pad_index _UpperCamelCase = target_dict.bos_index _UpperCamelCase = target_dict.eos_index _UpperCamelCase = len(target_dict.symbols ) _UpperCamelCase = os.path.join(__lowerCAmelCase , "vocab.json" ) if not os.path.isdir(__lowerCAmelCase ): logger.error("--pytorch_dump_folder_path ({}) should be a directory".format(__lowerCAmelCase ) ) return os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) _UpperCamelCase = target_dict.indices # fairseq has the <pad> and <s> switched _UpperCamelCase = 42 _UpperCamelCase = 43 with open(__lowerCAmelCase , "w" , encoding="utf-8" ) as vocab_handle: json.dump(__lowerCAmelCase , __lowerCAmelCase ) _UpperCamelCase = WavaVecaPhonemeCTCTokenizer( __lowerCAmelCase , unk_token=target_dict.unk_word , pad_token=target_dict.pad_word , bos_token=target_dict.bos_word , eos_token=target_dict.eos_word , word_delimiter_token="|" , do_lower_case=__lowerCAmelCase , ) _UpperCamelCase = True if config.feat_extract_norm == "layer" else False _UpperCamelCase = WavaVecaFeatureExtractor( feature_size=1 , sampling_rate=1_60_00 , padding_value=0 , do_normalize=__lowerCAmelCase , return_attention_mask=__lowerCAmelCase , ) _UpperCamelCase = WavaVecaProcessor(feature_extractor=__lowerCAmelCase , tokenizer=__lowerCAmelCase ) processor.save_pretrained(__lowerCAmelCase ) _UpperCamelCase = UniSpeechForCTC(__lowerCAmelCase ) else: _UpperCamelCase = UniSpeechForPreTraining(__lowerCAmelCase ) if is_finetuned: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path] , arg_overrides={"data": "/".join(dict_path.split("/" )[:-1] ), "w2v_path": checkpoint_path} ) else: _UpperCamelCase , _UpperCamelCase , _UpperCamelCase = fairseq.checkpoint_utils.load_model_ensemble_and_task([checkpoint_path] ) _UpperCamelCase = model[0].eval() recursively_load_weights(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) hf_unispeech.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": __lowerCAmelCase = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to fairseq checkpoint""") parser.add_argument("""--dict_path""", default=None, type=str, help="""Path to dict of fine-tuned model""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--not_finetuned""", action="""store_true""", help="""Whether the model to convert is a fine-tuned model or not""" ) __lowerCAmelCase = parser.parse_args() convert_unispeech_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
147
import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( '''kwargs, expected''' , [ ({'''num_shards''': 0, '''max_num_jobs''': 1}, []), ({'''num_shards''': 10, '''max_num_jobs''': 1}, [range(10 )]), ({'''num_shards''': 10, '''max_num_jobs''': 10}, [range(__lowerCAmelCase , i + 1 ) for i in range(10 )]), ({'''num_shards''': 1, '''max_num_jobs''': 10}, [range(1 )]), ({'''num_shards''': 10, '''max_num_jobs''': 3}, [range(0 , 4 ), range(4 , 7 ), range(7 , 10 )]), ({'''num_shards''': 3, '''max_num_jobs''': 10}, [range(0 , 1 ), range(1 , 2 ), range(2 , 3 )]), ] , ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> Optional[int]: snake_case__ = _distribute_shards(**__lowerCAmelCase ) assert out == expected @pytest.mark.parametrize( '''gen_kwargs, max_num_jobs, expected''' , [ ({'''foo''': 0}, 10, [{'''foo''': 0}]), ({'''shards''': [0, 1, 2, 3]}, 1, [{'''shards''': [0, 1, 2, 3]}]), ({'''shards''': [0, 1, 2, 3]}, 4, [{'''shards''': [0]}, {'''shards''': [1]}, {'''shards''': [2]}, {'''shards''': [3]}]), ({'''shards''': [0, 1]}, 4, [{'''shards''': [0]}, {'''shards''': [1]}]), ({'''shards''': [0, 1, 2, 3]}, 2, [{'''shards''': [0, 1]}, {'''shards''': [2, 3]}]), ] , ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Dict: snake_case__ = _split_gen_kwargs(__lowerCAmelCase , __lowerCAmelCase ) assert out == expected @pytest.mark.parametrize( '''gen_kwargs, expected''' , [ ({'''foo''': 0}, 1), ({'''shards''': [0]}, 1), ({'''shards''': [0, 1, 2, 3]}, 4), ({'''shards''': [0, 1, 2, 3], '''foo''': 0}, 4), ({'''shards''': [0, 1, 2, 3], '''other''': (0, 1)}, 4), ({'''shards''': [0, 1, 2, 3], '''shards2''': [0, 1]}, RuntimeError), ] , ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> List[Any]: if expected is RuntimeError: with pytest.raises(__lowerCAmelCase ): _number_of_shards_in_gen_kwargs(__lowerCAmelCase ) else: snake_case__ = _number_of_shards_in_gen_kwargs(__lowerCAmelCase ) assert out == expected
33
0
def __UpperCamelCase ( _A , _A ): lowerCAmelCase_ = len(__lowerCAmelCase ) + 1 lowerCAmelCase_ = len(__lowerCAmelCase ) + 1 # dp is a 2d matrix where dp[i][j] denotes whether prefix string of # length i of input_string matches with prefix string of length j of # given pattern. # "dp" stands for dynamic programming. lowerCAmelCase_ = [[0 for i in range(__lowerCAmelCase )] for j in range(__lowerCAmelCase )] # since string of zero length match pattern of zero length lowerCAmelCase_ = 1 # since pattern of zero length will never match with string of non-zero length for i in range(1 , __lowerCAmelCase ): lowerCAmelCase_ = 0 # since string of zero length will match with pattern where there # is at least one * alternatively for j in range(1 , __lowerCAmelCase ): lowerCAmelCase_ = dp[0][j - 2] if pattern[j - 1] == '''*''' else 0 # now using bottom-up approach to find for all remaining lengths for i in range(1 , __lowerCAmelCase ): for j in range(1 , __lowerCAmelCase ): if input_string[i - 1] == pattern[j - 1] or pattern[j - 1] == ".": lowerCAmelCase_ = dp[i - 1][j - 1] elif pattern[j - 1] == "*": if dp[i][j - 2] == 1: lowerCAmelCase_ = 1 elif pattern[j - 2] in (input_string[i - 1], "."): lowerCAmelCase_ = dp[i - 1][j] else: lowerCAmelCase_ = 0 else: lowerCAmelCase_ = 0 return bool(dp[-1][-1] ) if __name__ == "__main__": import doctest doctest.testmod() # inputing the strings # input_string = input("input a string :") # pattern = input("input a pattern :") _A = """aab""" _A = """c*a*b""" # using function to check whether given string matches the given pattern if match_pattern(input_string, pattern): print(f"{input_string} matches the given pattern {pattern}") else: print(f"{input_string} does not match with the given pattern {pattern}")
431
import random import unittest import torch from diffusers import IFImgaImgSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __magic_name__ (snake_case_ ,snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : str = IFImgaImgSuperResolutionPipeline __lowercase : Optional[Any] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'width', 'height'} __lowercase : int = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'original_image'} ) __lowercase : List[str] = PipelineTesterMixin.required_optional_params - {'latents'} def SCREAMING_SNAKE_CASE__ ( self:Dict ): return self._get_superresolution_dummy_components() def SCREAMING_SNAKE_CASE__ ( self:Dict , _a:int , _a:Optional[Any]=0 ): if str(_a ).startswith('''mps''' ): snake_case__ = torch.manual_seed(_a ) else: snake_case__ = torch.Generator(device=_a ).manual_seed(_a ) snake_case__ = floats_tensor((1, 3, 32, 32) , rng=random.Random(_a ) ).to(_a ) snake_case__ = floats_tensor((1, 3, 16, 16) , rng=random.Random(_a ) ).to(_a ) snake_case__ = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': image, '''original_image''': original_image, '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != '''cuda''' , reason='''float16 requires CUDA''' ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1e-1 ) def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def SCREAMING_SNAKE_CASE__ ( self:str ): self._test_save_load_local() def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
33
0
A__ : Optional[int] = """Input must be a string of 8 numbers plus letter""" A__ : List[str] = """TRWAGMYFPDXBNJZSQVHLCKE""" def UpperCamelCase( __UpperCamelCase : int ): if not isinstance(__lowerCAmelCase ,__lowerCAmelCase ): lowerCAmelCase_ : List[str] = f"""Expected string as input, found {type(__lowerCAmelCase ).__name__}""" raise TypeError(__lowerCAmelCase ) lowerCAmelCase_ : List[str] = spanish_id.replace('''-''' ,'''''' ).upper() if len(__lowerCAmelCase ) != 9: raise ValueError(__lowerCAmelCase ) try: lowerCAmelCase_ : int = int(spanish_id_clean[0:8] ) lowerCAmelCase_ : List[Any] = spanish_id_clean[8] except ValueError as ex: raise ValueError(__lowerCAmelCase ) from ex if letter.isdigit(): raise ValueError(__lowerCAmelCase ) return letter == LOOKUP_LETTERS[number % 23] if __name__ == "__main__": import doctest doctest.testmod()
171
import math class __magic_name__ : '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self:Optional[int] , _a:list[list[float]] , _a:list[int] ): snake_case__ = 0.0 snake_case__ = 0.0 for i in range(len(_a ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def SCREAMING_SNAKE_CASE__ ( self:Tuple , _a:list[list[int | float]] , _a:list[int] , _a:int , _a:float ): for i in range(len(_a ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def SCREAMING_SNAKE_CASE ( ) -> None: # Training Examples ( m, n ) snake_case__ = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) snake_case__ = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training snake_case__ = SelfOrganizingMap() snake_case__ = 3 snake_case__ = 0.5 for _ in range(__lowerCAmelCase ): for j in range(len(__lowerCAmelCase ) ): # training sample snake_case__ = training_samples[j] # Compute the winning vector snake_case__ = self_organizing_map.get_winner(__lowerCAmelCase , __lowerCAmelCase ) # Update the winning vector snake_case__ = self_organizing_map.update(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # classify test sample snake_case__ = [0, 0, 0, 1] snake_case__ = self_organizing_map.get_winner(__lowerCAmelCase , __lowerCAmelCase ) # results print(F"""Clusters that the test sample belongs to : {winner}""" ) print(F"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
33
0
def SCREAMING_SNAKE_CASE_ ( UpperCAmelCase_ : Any ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError('''The given input must be positive''' ) # get the generated string sequence SCREAMING_SNAKE_CASE_ : Optional[Any] =gray_code_sequence_string(__lowerCAmelCase ) # # convert them to integers for i in range(len(__lowerCAmelCase ) ): SCREAMING_SNAKE_CASE_ : Optional[int] =int(sequence[i] , 2 ) return sequence def SCREAMING_SNAKE_CASE_ ( UpperCAmelCase_ : Dict ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] SCREAMING_SNAKE_CASE_ : int =1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits SCREAMING_SNAKE_CASE_ : str =gray_code_sequence_string(bit_count - 1 ) SCREAMING_SNAKE_CASE_ : Optional[Any] =[] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): SCREAMING_SNAKE_CASE_ : Tuple ='''0''' + smaller_sequence[i] sequence.append(__lowerCAmelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): SCREAMING_SNAKE_CASE_ : int ='''1''' + smaller_sequence[i] sequence.append(__lowerCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
443
from __future__ import annotations from statistics import mean def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> list[int]: snake_case__ = [0] * no_of_processes snake_case__ = [0] * no_of_processes # Initialize remaining_time to waiting_time. for i in range(__lowerCAmelCase ): snake_case__ = burst_time[i] snake_case__ = [] snake_case__ = 0 snake_case__ = 0 # When processes are not completed, # A process whose arrival time has passed \ # and has remaining execution time is put into the ready_process. # The shortest process in the ready_process, target_process is executed. while completed != no_of_processes: snake_case__ = [] snake_case__ = -1 for i in range(__lowerCAmelCase ): if (arrival_time[i] <= total_time) and (remaining_time[i] > 0): ready_process.append(__lowerCAmelCase ) if len(__lowerCAmelCase ) > 0: snake_case__ = ready_process[0] for i in ready_process: if remaining_time[i] < remaining_time[target_process]: snake_case__ = i total_time += burst_time[target_process] completed += 1 snake_case__ = 0 snake_case__ = ( total_time - arrival_time[target_process] - burst_time[target_process] ) else: total_time += 1 return waiting_time def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> list[int]: snake_case__ = [0] * no_of_processes for i in range(__lowerCAmelCase ): snake_case__ = burst_time[i] + waiting_time[i] return turn_around_time if __name__ == "__main__": print("""[TEST CASE 01]""") lowerCamelCase__ : Tuple = 4 lowerCamelCase__ : Union[str, Any] = [2, 5, 3, 7] lowerCamelCase__ : Optional[Any] = [0, 0, 0, 0] lowerCamelCase__ : Dict = calculate_waitingtime(arrival_time, burst_time, no_of_processes) lowerCamelCase__ : Union[str, Any] = calculate_turnaroundtime( burst_time, no_of_processes, waiting_time ) # Printing the Result print("""PID\tBurst Time\tArrival Time\tWaiting Time\tTurnaround Time""") for i, process_id in enumerate(list(range(1, 5))): print( F"""{process_id}\t{burst_time[i]}\t\t\t{arrival_time[i]}\t\t\t\t""" F"""{waiting_time[i]}\t\t\t\t{turn_around_time[i]}""" ) print(F"""\nAverage waiting time = {mean(waiting_time):.5f}""") print(F"""Average turnaround time = {mean(turn_around_time):.5f}""")
33
0
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging from .tokenization_realm import RealmTokenizer lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} lowerCAmelCase__ = { """vocab_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/vocab.txt""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/vocab.txt""" ), """google/realm-orqa-nq-openqa""": """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-nq-reader""": """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/vocab.txt""", """google/realm-orqa-wq-openqa""": """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/vocab.txt""", """google/realm-orqa-wq-reader""": """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/vocab.txt""", }, """tokenizer_file""": { """google/realm-cc-news-pretrained-embedder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/tokenizer.jsont""" ), """google/realm-cc-news-pretrained-encoder""": ( """https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-scorer""": ( """https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/tokenizer.json""" ), """google/realm-cc-news-pretrained-openqa""": ( """https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/tokenizer.json""" ), """google/realm-orqa-nq-openqa""": ( """https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-nq-reader""": ( """https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-openqa""": ( """https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/tokenizer.json""" ), """google/realm-orqa-wq-reader""": ( """https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/tokenizer.json""" ), }, } lowerCAmelCase__ = { """google/realm-cc-news-pretrained-embedder""": 512, """google/realm-cc-news-pretrained-encoder""": 512, """google/realm-cc-news-pretrained-scorer""": 512, """google/realm-cc-news-pretrained-openqa""": 512, """google/realm-orqa-nq-openqa""": 512, """google/realm-orqa-nq-reader""": 512, """google/realm-orqa-wq-openqa""": 512, """google/realm-orqa-wq-reader""": 512, } lowerCAmelCase__ = { """google/realm-cc-news-pretrained-embedder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-encoder""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-scorer""": {"""do_lower_case""": True}, """google/realm-cc-news-pretrained-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-nq-reader""": {"""do_lower_case""": True}, """google/realm-orqa-wq-openqa""": {"""do_lower_case""": True}, """google/realm-orqa-wq-reader""": {"""do_lower_case""": True}, } class snake_case ( snake_case_ ): UpperCAmelCase__ = VOCAB_FILES_NAMES UpperCAmelCase__ = PRETRAINED_VOCAB_FILES_MAP UpperCAmelCase__ = PRETRAINED_INIT_CONFIGURATION UpperCAmelCase__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCAmelCase__ = RealmTokenizer def __init__(self , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_="[UNK]" , SCREAMING_SNAKE_CASE_="[SEP]" , SCREAMING_SNAKE_CASE_="[PAD]" , SCREAMING_SNAKE_CASE_="[CLS]" , SCREAMING_SNAKE_CASE_="[MASK]" , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=None , **SCREAMING_SNAKE_CASE_ , ): """simple docstring""" super().__init__( _a , tokenizer_file=_a , do_lower_case=_a , unk_token=_a , sep_token=_a , pad_token=_a , cls_token=_a , mask_token=_a , tokenize_chinese_chars=_a , strip_accents=_a , **_a , ) SCREAMING_SNAKE_CASE_ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('''lowercase''' , _a ) != do_lower_case or normalizer_state.get('''strip_accents''' , _a ) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , _a ) != tokenize_chinese_chars ): SCREAMING_SNAKE_CASE_ = getattr(_a , normalizer_state.pop('''type''' ) ) SCREAMING_SNAKE_CASE_ = do_lower_case SCREAMING_SNAKE_CASE_ = strip_accents SCREAMING_SNAKE_CASE_ = tokenize_chinese_chars SCREAMING_SNAKE_CASE_ = normalizer_class(**_a ) SCREAMING_SNAKE_CASE_ = do_lower_case def _lowercase (self , SCREAMING_SNAKE_CASE_ , **SCREAMING_SNAKE_CASE_ ): """simple docstring""" SCREAMING_SNAKE_CASE_ = PaddingStrategy.MAX_LENGTH SCREAMING_SNAKE_CASE_ = text SCREAMING_SNAKE_CASE_ = kwargs.pop('''text_pair''' , _a ) SCREAMING_SNAKE_CASE_ = kwargs.pop('''return_tensors''' , _a ) SCREAMING_SNAKE_CASE_ = { '''input_ids''': [], '''attention_mask''': [], '''token_type_ids''': [], } for idx, candidate_text in enumerate(_a ): if batch_text_pair is not None: SCREAMING_SNAKE_CASE_ = batch_text_pair[idx] else: SCREAMING_SNAKE_CASE_ = None SCREAMING_SNAKE_CASE_ = super().__call__(_a , _a , return_tensors=_a , **_a ) SCREAMING_SNAKE_CASE_ = encoded_candidates.get('''input_ids''' ) SCREAMING_SNAKE_CASE_ = encoded_candidates.get('''attention_mask''' ) SCREAMING_SNAKE_CASE_ = encoded_candidates.get('''token_type_ids''' ) if encoded_input_ids is not None: output_data["input_ids"].append(_a ) if encoded_attention_mask is not None: output_data["attention_mask"].append(_a ) if encoded_token_type_ids is not None: output_data["token_type_ids"].append(_a ) SCREAMING_SNAKE_CASE_ = {key: item for key, item in output_data.items() if len(_a ) != 0} return BatchEncoding(_a , tensor_type=_a ) def _lowercase (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None ): """simple docstring""" SCREAMING_SNAKE_CASE_ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _lowercase (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None ): """simple docstring""" SCREAMING_SNAKE_CASE_ = [self.sep_token_id] SCREAMING_SNAKE_CASE_ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _lowercase (self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = None ): """simple docstring""" SCREAMING_SNAKE_CASE_ = self._tokenizer.model.save(_a , name=_a ) return tuple(_a )
626
lowerCamelCase__ : List[str] = """Alexander Joslin""" import operator as op from .stack import Stack def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> int: snake_case__ = {'''*''': op.mul, '''/''': op.truediv, '''+''': op.add, '''-''': op.sub} snake_case__ = Stack() snake_case__ = Stack() for i in equation: if i.isdigit(): # RULE 1 operand_stack.push(int(__lowerCAmelCase ) ) elif i in operators: # RULE 2 operator_stack.push(__lowerCAmelCase ) elif i == ")": # RULE 4 snake_case__ = operator_stack.peek() operator_stack.pop() snake_case__ = operand_stack.peek() operand_stack.pop() snake_case__ = operand_stack.peek() operand_stack.pop() snake_case__ = operators[opr](__lowerCAmelCase , __lowerCAmelCase ) operand_stack.push(__lowerCAmelCase ) # RULE 5 return operand_stack.peek() if __name__ == "__main__": lowerCamelCase__ : Optional[Any] = """(5 + ((4 * 2) * (2 + 3)))""" # answer = 45 print(F"""{equation} = {dijkstras_two_stack_algorithm(equation)}""")
33
0
"""simple docstring""" import math class __lowerCamelCase : def UpperCAmelCase__ ( self , UpperCAmelCase , UpperCAmelCase ): lowerCamelCase_ = 0.0 lowerCamelCase_ = 0.0 for i in range(len(_a ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def UpperCAmelCase__ ( self , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ): for i in range(len(_a ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def lowercase ( ): # Training Examples ( m, n ) lowerCamelCase_ = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) lowerCamelCase_ = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training lowerCamelCase_ = SelfOrganizingMap() lowerCamelCase_ = 3 lowerCamelCase_ = 0.5 for _ in range(__lowerCAmelCase ): for j in range(len(__lowerCAmelCase ) ): # training sample lowerCamelCase_ = training_samples[j] # Compute the winning vector lowerCamelCase_ = self_organizing_map.get_winner(__lowerCAmelCase ,__lowerCAmelCase ) # Update the winning vector lowerCamelCase_ = self_organizing_map.update(__lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase ,__lowerCAmelCase ) # classify test sample lowerCamelCase_ = [0, 0, 0, 1] lowerCamelCase_ = self_organizing_map.get_winner(__lowerCAmelCase ,__lowerCAmelCase ) # results print(f"Clusters that the test sample belongs to : {winner}" ) print(f"Weights that have been trained : {weights}" ) # running the main() function if __name__ == "__main__": main()
29
import warnings from ...utils import logging from .image_processing_perceiver import PerceiverImageProcessor lowerCamelCase__ : int = logging.get_logger(__name__) class __magic_name__ (snake_case_ ): '''simple docstring''' def __init__( self:List[Any] , *_a:Dict , **_a:Tuple ): warnings.warn( '''The class PerceiverFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use PerceiverImageProcessor instead.''' , _a , ) super().__init__(*_a , **_a )
33
0
"""simple docstring""" import argparse import torch from transformers import OpenAIGPTConfig, OpenAIGPTModel, load_tf_weights_in_openai_gpt from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowerCamelCase__ ( __snake_case, __snake_case, __snake_case ) -> Union[str, Any]: """simple docstring""" if openai_config_file == "": _UpperCamelCase = OpenAIGPTConfig() else: _UpperCamelCase = OpenAIGPTConfig.from_json_file(__lowerCAmelCase ) _UpperCamelCase = OpenAIGPTModel(__lowerCAmelCase ) # Load weights from numpy load_tf_weights_in_openai_gpt(__lowerCAmelCase, __lowerCAmelCase, __lowerCAmelCase ) # Save pytorch-model _UpperCamelCase = pytorch_dump_folder_path + '''/''' + WEIGHTS_NAME _UpperCamelCase = pytorch_dump_folder_path + '''/''' + CONFIG_NAME print(F'''Save PyTorch model to {pytorch_weights_dump_path}''' ) torch.save(model.state_dict(), __lowerCAmelCase ) print(F'''Save configuration file to {pytorch_config_dump_path}''' ) with open(__lowerCAmelCase, '''w''', encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": _a = argparse.ArgumentParser() # Required parameters parser.add_argument( """--openai_checkpoint_folder_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--openai_config_file""", default="""""", type=str, help=( """An optional config json file corresponding to the pre-trained OpenAI model. \n""" """This specifies the model architecture.""" ), ) _a = parser.parse_args() convert_openai_checkpoint_to_pytorch( args.openai_checkpoint_folder_path, args.openai_config_file, args.pytorch_dump_folder_path )
19
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCamelCase__ : Tuple = { """configuration_roberta""": ["""ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """RobertaConfig""", """RobertaOnnxConfig"""], """tokenization_roberta""": ["""RobertaTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : Tuple = ["""RobertaTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : Optional[int] = [ """ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """RobertaForCausalLM""", """RobertaForMaskedLM""", """RobertaForMultipleChoice""", """RobertaForQuestionAnswering""", """RobertaForSequenceClassification""", """RobertaForTokenClassification""", """RobertaModel""", """RobertaPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : List[str] = [ """TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFRobertaForCausalLM""", """TFRobertaForMaskedLM""", """TFRobertaForMultipleChoice""", """TFRobertaForQuestionAnswering""", """TFRobertaForSequenceClassification""", """TFRobertaForTokenClassification""", """TFRobertaMainLayer""", """TFRobertaModel""", """TFRobertaPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : str = [ """FlaxRobertaForCausalLM""", """FlaxRobertaForMaskedLM""", """FlaxRobertaForMultipleChoice""", """FlaxRobertaForQuestionAnswering""", """FlaxRobertaForSequenceClassification""", """FlaxRobertaForTokenClassification""", """FlaxRobertaModel""", """FlaxRobertaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_roberta import ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaConfig, RobertaOnnxConfig from .tokenization_roberta import RobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roberta_fast import RobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta import ( ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaForCausalLM, RobertaForMaskedLM, RobertaForMultipleChoice, RobertaForQuestionAnswering, RobertaForSequenceClassification, RobertaForTokenClassification, RobertaModel, RobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta import ( TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForMultipleChoice, TFRobertaForQuestionAnswering, TFRobertaForSequenceClassification, TFRobertaForTokenClassification, TFRobertaMainLayer, TFRobertaModel, TFRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, FlaxRobertaPreTrainedModel, ) else: import sys lowerCamelCase__ : List[Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
33
0
import json import os import unittest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class lowercase_ ( snake_case_ , unittest.TestCase ): '''simple docstring''' __snake_case = CLIPTokenizer __snake_case = CLIPTokenizerFast __snake_case = True __snake_case = {} __snake_case = False def __lowerCAmelCase ( self : str ) ->str: """simple docstring""" super().setUp() # fmt: off a = ['''l''', '''o''', '''w''', '''e''', '''r''', '''s''', '''t''', '''i''', '''d''', '''n''', '''lo''', '''l</w>''', '''w</w>''', '''r</w>''', '''t</w>''', '''low</w>''', '''er</w>''', '''lowest</w>''', '''newer</w>''', '''wider''', '''<unk>''', '''<|startoftext|>''', '''<|endoftext|>'''] # fmt: on a = dict(zip(_a , range(len(_a ) ) ) ) a = ['''#version: 0.2''', '''l o''', '''lo w</w>''', '''e r</w>'''] a = {'''unk_token''': '''<unk>'''} a = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) a = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write(json.dumps(_a ) + '''\n''' ) with open(self.merges_file , '''w''' , encoding='''utf-8''' ) as fp: fp.write('''\n'''.join(_a ) ) def __lowerCAmelCase ( self : List[Any] , **__UpperCAmelCase : Optional[int] ) ->Optional[int]: """simple docstring""" kwargs.update(self.special_tokens_map ) return CLIPTokenizer.from_pretrained(self.tmpdirname , **_a ) def __lowerCAmelCase ( self : List[Any] , **__UpperCAmelCase : Union[str, Any] ) ->Any: """simple docstring""" kwargs.update(self.special_tokens_map ) return CLIPTokenizerFast.from_pretrained(self.tmpdirname , **_a ) def __lowerCAmelCase ( self : Tuple , __UpperCAmelCase : List[str] ) ->int: """simple docstring""" a = '''lower newer''' a = '''lower newer''' return input_text, output_text def __lowerCAmelCase ( self : Optional[Any] ) ->Any: """simple docstring""" a = CLIPTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map ) a = '''lower newer''' a = ['''lo''', '''w''', '''er</w>''', '''n''', '''e''', '''w''', '''er</w>'''] a = tokenizer.tokenize(_a ) self.assertListEqual(_a , _a ) a = tokens + [tokenizer.unk_token] a = [10, 2, 16, 9, 3, 2, 16, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , _a ) @require_ftfy def __lowerCAmelCase ( self : Optional[int] ) ->Any: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): a = self.tokenizer_class.from_pretrained(_a , **_a ) a = self.rust_tokenizer_class.from_pretrained(_a , **_a ) a = '''A\n\'ll 11p223RF☆ho!!to?\'d\'d\'\'d of a cat to-$\'\'d.''' a = tokenizer_s.tokenize(_a ) a = tokenizer_r.tokenize(_a ) self.assertListEqual(_a , _a ) # Test that the tokenization is identical on an example containing a character (Latin Small Letter A # with Tilde) encoded in 2 different ways a = '''xa\u0303y''' + ''' ''' + '''x\xe3y''' a = tokenizer_s.tokenize(_a ) a = tokenizer_r.tokenize(_a ) self.assertListEqual(_a , _a ) # Test that the tokenization is identical on unicode of space type a = [ '''\u0009''', # (horizontal tab, '\t') '''\u000B''', # (vertical tab) '''\u000C''', # (form feed) '''\u0020''', # (space, ' ') '''\u200E''', # (left-to-right mark):w '''\u200F''', # (right-to-left mark) ] for unicode_seq in spaces_unicodes: a = tokenizer_s.tokenize(_a ) a = tokenizer_r.tokenize(_a ) self.assertListEqual(_a , _a ) # Test that the tokenization is identical on unicode of line break type a = [ '''\u000A''', # (line feed, '\n') '''\r\n''', # (carriage return and line feed, '\r\n') '''\u000D''', # (carriage return, '\r') '''\r''', # (carriage return, '\r') '''\u000D''', # (carriage return, '\r') '''\u2028''', # (line separator) '''\u2029''', # (paragraph separator) # "\u0085", # (next line) ] # The tokenization is not identical for the character "\u0085" (next line). The slow version using ftfy transforms # it into the Horizontal Ellipsis character "…" ("\u2026") while the fast version transforms it into a # space (and thus into an empty list). for unicode_seq in line_break_unicodes: a = tokenizer_s.tokenize(_a ) a = tokenizer_r.tokenize(_a ) self.assertListEqual(_a , _a ) def __lowerCAmelCase ( self : Any ) ->Optional[Any]: """simple docstring""" for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"""{tokenizer.__class__.__name__} ({pretrained_name})""" ): a = '''hello''' # `hello` is a token in the vocabulary of `pretrained_name` a = F"""{text_of_1_token} {text_of_1_token}""" a = self.rust_tokenizer_class.from_pretrained( _a , use_fast=_a , ) a = tokenizer_r(_a , return_offsets_mapping=_a , add_special_tokens=_a ) self.assertEqual(encoding.offset_mapping[0] , (0, len(_a )) ) self.assertEqual( encoding.offset_mapping[1] , (len(_a ) + 1, len(_a ) + 1 + len(_a )) , ) a = F""" {text}""" a = self.rust_tokenizer_class.from_pretrained( _a , use_fast=_a , ) a = tokenizer_r(_a , return_offsets_mapping=_a , add_special_tokens=_a ) self.assertEqual(encoding.offset_mapping[0] , (1, 1 + len(_a )) ) self.assertEqual( encoding.offset_mapping[1] , (1 + len(_a ) + 1, 1 + len(_a ) + 1 + len(_a )) , ) def __lowerCAmelCase ( self : str ) ->str: """simple docstring""" with self.assertRaises(_a ) as context: self.rust_tokenizer_class.from_pretrained('''robot-test/old-clip-tokenizer''' ) self.assertTrue( context.exception.args[0].startswith( '''The `backend_tokenizer` provided does not match the expected format.''' ) ) @require_ftfy def __lowerCAmelCase ( self : int ) ->str: """simple docstring""" super().test_tokenization_python_rust_equals() def __lowerCAmelCase ( self : Optional[Any] ) ->int: """simple docstring""" pass
117
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer import diffusers from diffusers import ( AutoencoderKL, EulerDiscreteScheduler, StableDiffusionLatentUpscalePipeline, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.schedulers import KarrasDiffusionSchedulers from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> List[Any]: snake_case__ = [tensor.shape for tensor in tensor_list] return all(shape == shapes[0] for shape in shapes[1:] ) class __magic_name__ (snake_case_ ,snake_case_ ,snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : Dict = StableDiffusionLatentUpscalePipeline __lowercase : List[str] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { 'height', 'width', 'cross_attention_kwargs', 'negative_prompt_embeds', 'prompt_embeds', } __lowercase : List[Any] = PipelineTesterMixin.required_optional_params - {'num_images_per_prompt'} __lowercase : Any = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS __lowercase : int = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess __lowercase : List[Any] = frozenset([] ) __lowercase : Any = True @property def SCREAMING_SNAKE_CASE__ ( self:List[str] ): snake_case__ = 1 snake_case__ = 4 snake_case__ = (16, 16) snake_case__ = floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(_a ) return image def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): torch.manual_seed(0 ) snake_case__ = UNetaDConditionModel( act_fn='''gelu''' , attention_head_dim=8 , norm_num_groups=_a , block_out_channels=[32, 32, 64, 64] , time_cond_proj_dim=1_60 , conv_in_kernel=1 , conv_out_kernel=1 , cross_attention_dim=32 , down_block_types=( '''KDownBlock2D''', '''KCrossAttnDownBlock2D''', '''KCrossAttnDownBlock2D''', '''KCrossAttnDownBlock2D''', ) , in_channels=8 , mid_block_type=_a , only_cross_attention=_a , out_channels=5 , resnet_time_scale_shift='''scale_shift''' , time_embedding_type='''fourier''' , timestep_post_act='''gelu''' , up_block_types=('''KCrossAttnUpBlock2D''', '''KCrossAttnUpBlock2D''', '''KCrossAttnUpBlock2D''', '''KUpBlock2D''') , ) snake_case__ = AutoencoderKL( block_out_channels=[32, 32, 64, 64] , in_channels=3 , out_channels=3 , down_block_types=[ '''DownEncoderBlock2D''', '''DownEncoderBlock2D''', '''DownEncoderBlock2D''', '''DownEncoderBlock2D''', ] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D''', '''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) snake_case__ = EulerDiscreteScheduler(prediction_type='''sample''' ) snake_case__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act='''quick_gelu''' , projection_dim=5_12 , ) snake_case__ = CLIPTextModel(_a ) snake_case__ = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''' ) snake_case__ = { '''unet''': model.eval(), '''vae''': vae.eval(), '''scheduler''': scheduler, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def SCREAMING_SNAKE_CASE__ ( self:List[Any] , _a:Optional[Any] , _a:List[str]=0 ): if str(_a ).startswith('''mps''' ): snake_case__ = torch.manual_seed(_a ) else: snake_case__ = torch.Generator(device=_a ).manual_seed(_a ) snake_case__ = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': self.dummy_image.cpu(), '''generator''': generator, '''num_inference_steps''': 2, '''output_type''': '''numpy''', } return inputs def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = '''cpu''' snake_case__ = self.get_dummy_components() snake_case__ = self.pipeline_class(**_a ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) snake_case__ = self.get_dummy_inputs(_a ) snake_case__ = pipe(**_a ).images snake_case__ = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 2_56, 2_56, 3) ) snake_case__ = np.array( [0.47222412, 0.41921633, 0.44717434, 0.46874192, 0.42588258, 0.46150726, 0.4677534, 0.45583832, 0.48579055] ) snake_case__ = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(_a , 1e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): super().test_attention_slicing_forward_pass(expected_max_diff=7e-3 ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): super().test_cpu_offload_forward_pass(expected_max_diff=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:str ): super().test_dict_tuple_outputs_equivalent(expected_max_difference=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Any ): super().test_inference_batch_single_identical(expected_max_diff=7e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): super().test_pt_np_pil_outputs_equivalent(expected_max_diff=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): super().test_save_load_local(expected_max_difference=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:str ): super().test_save_load_optional_components(expected_max_difference=3e-3 ) def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = [ '''DDIMScheduler''', '''DDPMScheduler''', '''PNDMScheduler''', '''HeunDiscreteScheduler''', '''EulerAncestralDiscreteScheduler''', '''KDPM2DiscreteScheduler''', '''KDPM2AncestralDiscreteScheduler''', '''DPMSolverSDEScheduler''', ] snake_case__ = self.get_dummy_components() snake_case__ = self.pipeline_class(**_a ) # make sure that PNDM does not need warm-up pipe.scheduler.register_to_config(skip_prk_steps=_a ) pipe.to(_a ) pipe.set_progress_bar_config(disable=_a ) snake_case__ = self.get_dummy_inputs(_a ) snake_case__ = 2 snake_case__ = [] for scheduler_enum in KarrasDiffusionSchedulers: if scheduler_enum.name in skip_schedulers: # no sigma schedulers are not supported # no schedulers continue snake_case__ = getattr(_a , scheduler_enum.name ) snake_case__ = scheduler_cls.from_config(pipe.scheduler.config ) snake_case__ = pipe(**_a )[0] outputs.append(_a ) assert check_same_shape(_a ) @require_torch_gpu @slow class __magic_name__ (unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = torch.manual_seed(33 ) snake_case__ = StableDiffusionPipeline.from_pretrained('''CompVis/stable-diffusion-v1-4''' , torch_dtype=torch.floataa ) pipe.to('''cuda''' ) snake_case__ = StableDiffusionLatentUpscalePipeline.from_pretrained( '''stabilityai/sd-x2-latent-upscaler''' , torch_dtype=torch.floataa ) upscaler.to('''cuda''' ) snake_case__ = '''a photo of an astronaut high resolution, unreal engine, ultra realistic''' snake_case__ = pipe(_a , generator=_a , output_type='''latent''' ).images snake_case__ = upscaler( prompt=_a , image=_a , num_inference_steps=20 , guidance_scale=0 , generator=_a , output_type='''np''' , ).images[0] snake_case__ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/astronaut_1024.npy''' ) assert np.abs((expected_image - image).mean() ) < 5e-2 def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): snake_case__ = torch.manual_seed(33 ) snake_case__ = StableDiffusionLatentUpscalePipeline.from_pretrained( '''stabilityai/sd-x2-latent-upscaler''' , torch_dtype=torch.floataa ) upscaler.to('''cuda''' ) snake_case__ = '''the temple of fire by Ross Tran and Gerardo Dottori, oil on canvas''' snake_case__ = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_512.png''' ) snake_case__ = upscaler( prompt=_a , image=_a , num_inference_steps=20 , guidance_scale=0 , generator=_a , output_type='''np''' , ).images[0] snake_case__ = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/latent-upscaler/fire_temple_1024.npy''' ) assert np.abs((expected_image - image).max() ) < 5e-2
33
0
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv("""TEST_SAGEMAKER""" , """False""" ) ) is not True , reason="""Skipping test because should only be run when releasing minor transformers version""" , ) @pytest.mark.usefixtures("""sm_env""" ) @parameterized_class( [ { """framework""": """pytorch""", """script""": """run_glue_model_parallelism.py""", """model_name_or_path""": """roberta-large""", """instance_type""": """ml.p3dn.24xlarge""", """results""": {"""train_runtime""": 1_6_0_0, """eval_accuracy""": 0.3, """eval_loss""": 1.2}, }, { """framework""": """pytorch""", """script""": """run_glue.py""", """model_name_or_path""": """roberta-large""", """instance_type""": """ml.p3dn.24xlarge""", """results""": {"""train_runtime""": 1_6_0_0, """eval_accuracy""": 0.3, """eval_loss""": 1.2}, }, ] ) class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def __UpperCamelCase ( self ): '''simple docstring''' if self.framework == "pytorch": subprocess.run( f'''cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'''.split() , encoding='''utf-8''' , check=_a , ) assert hasattr(self , '''env''' ) def __UpperCamelCase ( self , lowercase__ ): '''simple docstring''' __A ={ '''enabled''': True, '''processes_per_host''': 8, } __A ={ '''enabled''': True, '''parameters''': { '''microbatches''': 4, '''placement_strategy''': '''spread''', '''pipeline''': '''interleaved''', '''optimize''': '''speed''', '''partitions''': 4, '''ddp''': True, }, } __A ={'''smdistributed''': {'''modelparallel''': smp_options}, '''mpi''': mpi_options} __A ='''trainer''' if self.script == '''run_glue.py''' else '''smtrainer''' # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=f'''{self.env.base_job_name}-{instance_count}-smp-{name_extension}''' , instance_count=_a , instance_type=self.instance_type , debugger_hook_config=_a , hyperparameters={ **self.env.hyperparameters, '''model_name_or_path''': self.model_name_or_path, '''max_steps''': 5_0_0, } , metric_definitions=self.env.metric_definitions , distribution=_a , py_version='''py36''' , ) def __UpperCamelCase ( self , lowercase__ ): '''simple docstring''' TrainingJobAnalytics(_a ).export_csv(f'''{self.env.test_path}/{job_name}_metrics.csv''' ) @parameterized.expand([(1,)] ) def __UpperCamelCase ( self , lowercase__ ): '''simple docstring''' __A =self.create_estimator(_a ) # run training estimator.fit() # result dataframe __A =TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis __A =list(result_metrics_df[result_metrics_df.metric_name == '''eval_accuracy''']['''value'''] ) __A =list(result_metrics_df[result_metrics_df.metric_name == '''eval_loss''']['''value'''] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping __A =( Session().describe_training_job(estimator.latest_training_job.name ).get('''TrainingTimeInSeconds''' , 9_9_9_9_9_9 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results['''eval_accuracy'''] for t in eval_accuracy ) assert all(t <= self.results['''eval_loss'''] for t in eval_loss ) # dump tests result into json file to share in PR with open(f'''{estimator.latest_training_job.name}.json''' , '''w''' ) as outfile: json.dump({'''train_time''': train_runtime, '''eval_accuracy''': eval_accuracy, '''eval_loss''': eval_loss} , _a )
184
import shutil import tempfile import unittest import numpy as np import pytest from transformers import is_speech_available, is_vision_available from transformers.testing_utils import require_torch if is_vision_available(): from transformers import TvltImageProcessor if is_speech_available(): from transformers import TvltFeatureExtractor from transformers import TvltProcessor @require_torch class __magic_name__ (unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = '''ZinengTang/tvlt-base''' snake_case__ = tempfile.mkdtemp() def SCREAMING_SNAKE_CASE__ ( self:Dict , **_a:List[Any] ): return TvltImageProcessor.from_pretrained(self.checkpoint , **_a ) def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] , **_a:Tuple ): return TvltFeatureExtractor.from_pretrained(self.checkpoint , **_a ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): shutil.rmtree(self.tmpdirname ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) processor.save_pretrained(self.tmpdirname ) snake_case__ = TvltProcessor.from_pretrained(self.tmpdirname ) self.assertIsInstance(processor.feature_extractor , _a ) self.assertIsInstance(processor.image_processor , _a ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) snake_case__ = np.ones([1_20_00] ) snake_case__ = feature_extractor(_a , return_tensors='''np''' ) snake_case__ = processor(audio=_a , return_tensors='''np''' ) for key in audio_dict.keys(): self.assertAlmostEqual(audio_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) snake_case__ = np.ones([3, 2_24, 2_24] ) snake_case__ = image_processor(_a , return_tensors='''np''' ) snake_case__ = processor(images=_a , return_tensors='''np''' ) for key in image_dict.keys(): self.assertAlmostEqual(image_dict[key].sum() , input_processor[key].sum() , delta=1e-2 ) def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) snake_case__ = np.ones([1_20_00] ) snake_case__ = np.ones([3, 2_24, 2_24] ) snake_case__ = processor(audio=_a , images=_a ) self.assertListEqual(list(inputs.keys() ) , ['''audio_values''', '''audio_mask''', '''pixel_values''', '''pixel_mask'''] ) # test if it raises when no input is passed with pytest.raises(_a ): processor() def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = self.get_image_processor() snake_case__ = self.get_feature_extractor() snake_case__ = TvltProcessor(image_processor=_a , feature_extractor=_a ) self.assertListEqual( processor.model_input_names , image_processor.model_input_names + feature_extractor.model_input_names , msg='''`processor` and `image_processor`+`feature_extractor` model input names do not match''' , )
33
0
from __future__ import annotations def SCREAMING_SNAKE_CASE ( snake_case , snake_case , snake_case , snake_case , snake_case , ) -> None: __lowercase = len(__lowerCAmelCase ) # If row is equal to the size of the board it means there are a queen in each row in # the current board (possible_board) if row == n: # We convert the variable possible_board that looks like this: [1, 3, 0, 2] to # this: ['. Q . . ', '. . . Q ', 'Q . . . ', '. . Q . '] boards.append(['. ' * i + 'Q ' + '. ' * (n - 1 - i) for i in possible_board] ) return # We iterate each column in the row to find all possible results in each row for col in range(__lowerCAmelCase ): # We apply that we learned previously. First we check that in the current board # (possible_board) there are not other same value because if there is it means # that there are a collision in vertical. Then we apply the two formulas we # learned before: # # 45º: y - x = b or 45: row - col = b # 135º: y + x = b or row + col = b. # # And we verify if the results of this two formulas not exist in their variables # respectively. (diagonal_right_collisions, diagonal_left_collisions) # # If any or these are True it means there is a collision so we continue to the # next value in the for loop. if ( col in possible_board or row - col in diagonal_right_collisions or row + col in diagonal_left_collisions ): continue # If it is False we call dfs function again and we update the inputs depth_first_search( [*possible_board, col] , [*diagonal_right_collisions, row - col] , [*diagonal_left_collisions, row + col] , __lowerCAmelCase , __lowerCAmelCase , ) def SCREAMING_SNAKE_CASE ( snake_case ) -> None: __lowercase = [] depth_first_search([] , [] , [] , __lowerCAmelCase , __lowerCAmelCase ) # Print all the boards for board in boards: for column in board: print(__lowerCAmelCase ) print('' ) print(len(__lowerCAmelCase ) , 'solutions were found.' ) if __name__ == "__main__": import doctest doctest.testmod() n_queens_solution(4)
375
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCamelCase__ : List[Any] = logging.get_logger(__name__) lowerCamelCase__ : Optional[int] = { """facebook/data2vec-vision-base-ft""": ( """https://huggingface.co/facebook/data2vec-vision-base-ft/resolve/main/config.json""" ), } class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : Optional[int] = 'data2vec-vision' def __init__( self:int , _a:Tuple=7_68 , _a:int=12 , _a:Any=12 , _a:Optional[int]=30_72 , _a:Optional[int]="gelu" , _a:Any=0.0 , _a:Any=0.0 , _a:List[str]=0.02 , _a:Dict=1e-12 , _a:Tuple=2_24 , _a:Any=16 , _a:str=3 , _a:str=False , _a:Union[str, Any]=False , _a:Optional[int]=False , _a:Any=False , _a:Dict=0.1 , _a:Dict=0.1 , _a:str=True , _a:str=[3, 5, 7, 11] , _a:List[str]=[1, 2, 3, 6] , _a:List[str]=True , _a:Any=0.4 , _a:str=2_56 , _a:Union[str, Any]=1 , _a:int=False , _a:Optional[int]=2_55 , **_a:Dict , ): super().__init__(**_a ) snake_case__ = hidden_size snake_case__ = num_hidden_layers snake_case__ = num_attention_heads snake_case__ = intermediate_size snake_case__ = hidden_act snake_case__ = hidden_dropout_prob snake_case__ = attention_probs_dropout_prob snake_case__ = initializer_range snake_case__ = layer_norm_eps snake_case__ = image_size snake_case__ = patch_size snake_case__ = num_channels snake_case__ = use_mask_token snake_case__ = use_absolute_position_embeddings snake_case__ = use_relative_position_bias snake_case__ = use_shared_relative_position_bias snake_case__ = layer_scale_init_value snake_case__ = drop_path_rate snake_case__ = use_mean_pooling # decode head attributes (semantic segmentation) snake_case__ = out_indices snake_case__ = pool_scales # auxiliary head attributes (semantic segmentation) snake_case__ = use_auxiliary_head snake_case__ = auxiliary_loss_weight snake_case__ = auxiliary_channels snake_case__ = auxiliary_num_convs snake_case__ = auxiliary_concat_input snake_case__ = semantic_loss_ignore_index class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : Any = version.parse('1.11' ) @property def SCREAMING_SNAKE_CASE__ ( self:List[str] ): return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ] ) @property def SCREAMING_SNAKE_CASE__ ( self:Tuple ): return 1e-4
33
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional from packaging import version if TYPE_CHECKING: from ... import PreTrainedTokenizer, TensorType from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import is_torch_available, logging lowercase : Optional[Any] = logging.get_logger(__name__) lowercase : List[str] = { """bigscience/bloom""": """https://huggingface.co/bigscience/bloom/resolve/main/config.json""", """bigscience/bloom-560m""": """https://huggingface.co/bigscience/bloom-560m/blob/main/config.json""", """bigscience/bloom-1b1""": """https://huggingface.co/bigscience/bloom-1b1/blob/main/config.json""", """bigscience/bloom-1b7""": """https://huggingface.co/bigscience/bloom-1b7/blob/main/config.json""", """bigscience/bloom-3b""": """https://huggingface.co/bigscience/bloom-3b/blob/main/config.json""", """bigscience/bloom-7b1""": """https://huggingface.co/bigscience/bloom-7b1/blob/main/config.json""", } class UpperCAmelCase_ ( snake_case_ ): '''simple docstring''' A : Any = 'bloom' A : str = ['past_key_values'] A : Optional[Any] = { 'num_hidden_layers': 'n_layer', 'num_attention_heads': 'n_head', } def __init__( self , _SCREAMING_SNAKE_CASE=25_0880 , _SCREAMING_SNAKE_CASE=64 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=8 , _SCREAMING_SNAKE_CASE=1e-5 , _SCREAMING_SNAKE_CASE=0.02 , _SCREAMING_SNAKE_CASE=True , _SCREAMING_SNAKE_CASE=1 , _SCREAMING_SNAKE_CASE=2 , _SCREAMING_SNAKE_CASE=False , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=0.0 , _SCREAMING_SNAKE_CASE=1 , _SCREAMING_SNAKE_CASE=False , **_SCREAMING_SNAKE_CASE , ) -> Union[str, Any]: snake_case_ : Union[str, Any] = vocab_size # Backward compatibility with n_embed kwarg snake_case_ : List[str] = kwargs.pop("n_embed" , _a ) snake_case_ : str = hidden_size if n_embed is None else n_embed snake_case_ : Any = n_layer snake_case_ : Dict = n_head snake_case_ : Any = layer_norm_epsilon snake_case_ : List[str] = initializer_range snake_case_ : Tuple = use_cache snake_case_ : Optional[Any] = pretraining_tp snake_case_ : Tuple = apply_residual_connection_post_layernorm snake_case_ : Any = hidden_dropout snake_case_ : Optional[int] = attention_dropout snake_case_ : Optional[int] = bos_token_id snake_case_ : Optional[Any] = eos_token_id snake_case_ : Optional[int] = slow_but_exact super().__init__(bos_token_id=_a , eos_token_id=_a , **_a ) class UpperCAmelCase_ ( snake_case_ ): '''simple docstring''' A : Optional[int] = version.parse('1.12' ) def __init__( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = "default" , _SCREAMING_SNAKE_CASE = None , _SCREAMING_SNAKE_CASE = False , ) -> List[Any]: super().__init__(_a , task=_a , patching_specs=_a , use_past=_a ) if not getattr(self._config , "pad_token_id" , _a ): # TODO: how to do that better? snake_case_ : Optional[int] = 0 @property def _lowerCAmelCase ( self ) -> Dict: snake_case_ : Any = OrderedDict({"input_ids": {0: "batch", 1: "sequence"}} ) if self.use_past: # BLOOM stores values on dynamic axis 2. For more details see: https://github.com/huggingface/transformers/pull/18344 self.fill_with_past_key_values_(_a , direction="inputs" , inverted_values_shape=_a ) snake_case_ : Tuple = {0: "batch", 1: "past_sequence + sequence"} else: snake_case_ : List[str] = {0: "batch", 1: "sequence"} return common_inputs @property def _lowerCAmelCase ( self ) -> List[str]: return self._config.n_layer @property def _lowerCAmelCase ( self ) -> Optional[Any]: return self._config.n_head @property def _lowerCAmelCase ( self ) -> Tuple: return 1e-3 def _lowerCAmelCase ( self , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE = -1 , _SCREAMING_SNAKE_CASE = -1 , _SCREAMING_SNAKE_CASE = False , _SCREAMING_SNAKE_CASE = None , ) -> Optional[int]: snake_case_ : Any = super(_a , self ).generate_dummy_inputs( _a , batch_size=_a , seq_length=_a , is_pair=_a , framework=_a ) # We need to order the input in the way they appears in the forward() snake_case_ : Any = OrderedDict({"input_ids": common_inputs["input_ids"]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch snake_case_ , snake_case_ : Optional[Any] = common_inputs["input_ids"].shape # Not using the same length for past_key_values snake_case_ : int = seqlen + 2 snake_case_ : Optional[int] = self._config.hidden_size // self.num_attention_heads snake_case_ : List[str] = ( batch * self.num_attention_heads, head_dim, past_key_values_length, ) snake_case_ : Optional[Any] = ( batch * self.num_attention_heads, past_key_values_length, head_dim, ) snake_case_ : Dict = [ (torch.zeros(_a ), torch.zeros(_a )) for _ in range(self.num_layers ) ] snake_case_ : str = common_inputs["attention_mask"] if self.use_past: snake_case_ : Dict = ordered_inputs["attention_mask"].dtype snake_case_ : Dict = torch.cat( [ordered_inputs["attention_mask"], torch.ones(_a , _a , dtype=_a )] , dim=1 ) return ordered_inputs @property def _lowerCAmelCase ( self ) -> Any: return 13
568
import os import sys lowerCamelCase__ : Optional[int] = os.path.join(os.path.dirname(__file__), """src""") sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) lowerCamelCase__ : Optional[int] = [ """torch""", """numpy""", """tokenizers""", """filelock""", """requests""", """tqdm""", """regex""", """sentencepiece""", """sacremoses""", """importlib_metadata""", """huggingface_hub""", ] @add_start_docstrings(AutoConfig.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Any: return AutoConfig.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoTokenizer.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> List[str]: return AutoTokenizer.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModel.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Tuple: return AutoModel.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Union[str, Any]: return AutoModelForCausalLM.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> List[Any]: return AutoModelForMaskedLM.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> List[str]: return AutoModelForSequenceClassification.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def SCREAMING_SNAKE_CASE ( *__lowerCAmelCase , **__lowerCAmelCase ) -> Union[str, Any]: return AutoModelForQuestionAnswering.from_pretrained(*__lowerCAmelCase , **__lowerCAmelCase )
33
0
from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowerCamelCase_ ( snake_case_ ): __lowercase : Optional[int] = ['image_processor', 'tokenizer'] __lowercase : List[str] = 'BlipImageProcessor' __lowercase : int = 'AutoTokenizer' def __init__( self , lowerCamelCase_ , lowerCamelCase_ ) -> int: """simple docstring""" _UpperCamelCase = False super().__init__(_a , _a ) _UpperCamelCase = self.image_processor def __call__( self , lowerCamelCase_ = None , lowerCamelCase_ = None , lowerCamelCase_ = True , lowerCamelCase_ = False , lowerCamelCase_ = None , lowerCamelCase_ = None , lowerCamelCase_ = 0 , lowerCamelCase_ = None , lowerCamelCase_ = None , lowerCamelCase_ = False , lowerCamelCase_ = False , lowerCamelCase_ = False , lowerCamelCase_ = False , lowerCamelCase_ = False , lowerCamelCase_ = True , lowerCamelCase_ = None , **lowerCamelCase_ , ) -> Optional[int]: """simple docstring""" if images is None and text is None: raise ValueError("You have to specify either images or text." ) # Get only text if images is None: _UpperCamelCase = self.tokenizer _UpperCamelCase = self.tokenizer( text=_a , add_special_tokens=_a , padding=_a , truncation=_a , max_length=_a , stride=_a , pad_to_multiple_of=_a , return_attention_mask=_a , return_overflowing_tokens=_a , return_special_tokens_mask=_a , return_offsets_mapping=_a , return_token_type_ids=_a , return_length=_a , verbose=_a , return_tensors=_a , **_a , ) return text_encoding # add pixel_values _UpperCamelCase = self.image_processor(_a , return_tensors=_a ) if text is not None: _UpperCamelCase = self.tokenizer( text=_a , add_special_tokens=_a , padding=_a , truncation=_a , max_length=_a , stride=_a , pad_to_multiple_of=_a , return_attention_mask=_a , return_overflowing_tokens=_a , return_special_tokens_mask=_a , return_offsets_mapping=_a , return_token_type_ids=_a , return_length=_a , verbose=_a , return_tensors=_a , **_a , ) else: _UpperCamelCase = None if text_encoding is not None: encoding_image_processor.update(_a ) return encoding_image_processor def lowercase ( self , *lowerCamelCase_ , **lowerCamelCase_ ) -> int: """simple docstring""" return self.tokenizer.batch_decode(*_a , **_a ) def lowercase ( self , *lowerCamelCase_ , **lowerCamelCase_ ) -> int: """simple docstring""" return self.tokenizer.decode(*_a , **_a ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def lowercase ( self ) -> int: """simple docstring""" _UpperCamelCase = self.tokenizer.model_input_names _UpperCamelCase = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) )
147
import torch from diffusers import CMStochasticIterativeScheduler from .test_schedulers import SchedulerCommonTest class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : str = (CMStochasticIterativeScheduler,) __lowercase : List[str] = 10 def SCREAMING_SNAKE_CASE__ ( self:int , **_a:Optional[int] ): snake_case__ = { '''num_train_timesteps''': 2_01, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } config.update(**_a ) return config def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): snake_case__ = 10 snake_case__ = self.get_scheduler_config() snake_case__ = self.scheduler_classes[0](**_a ) scheduler.set_timesteps(_a ) snake_case__ = scheduler.timesteps[0] snake_case__ = scheduler.timesteps[1] snake_case__ = self.dummy_sample snake_case__ = 0.1 * sample snake_case__ = scheduler.step(_a , _a , _a ).prev_sample snake_case__ = scheduler.step(_a , _a , _a ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def SCREAMING_SNAKE_CASE__ ( self:Any ): for timesteps in [10, 50, 1_00, 10_00]: self.check_over_configs(num_train_timesteps=_a ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): for clip_denoised in [True, False]: self.check_over_configs(clip_denoised=_a ) def SCREAMING_SNAKE_CASE__ ( self:int ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = 1 scheduler.set_timesteps(_a ) snake_case__ = scheduler.timesteps snake_case__ = torch.manual_seed(0 ) snake_case__ = self.dummy_model() snake_case__ = self.dummy_sample_deter * scheduler.init_noise_sigma for i, t in enumerate(_a ): # 1. scale model input snake_case__ = scheduler.scale_model_input(_a , _a ) # 2. predict noise residual snake_case__ = model(_a , _a ) # 3. predict previous sample x_t-1 snake_case__ = scheduler.step(_a , _a , _a , generator=_a ).prev_sample snake_case__ = pred_prev_sample snake_case__ = torch.sum(torch.abs(_a ) ) snake_case__ = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 192.7614 ) < 1e-2 assert abs(result_mean.item() - 0.2510 ) < 1e-3 def SCREAMING_SNAKE_CASE__ ( self:Tuple ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = [1_06, 0] scheduler.set_timesteps(timesteps=_a ) snake_case__ = scheduler.timesteps snake_case__ = torch.manual_seed(0 ) snake_case__ = self.dummy_model() snake_case__ = self.dummy_sample_deter * scheduler.init_noise_sigma for t in timesteps: # 1. scale model input snake_case__ = scheduler.scale_model_input(_a , _a ) # 2. predict noise residual snake_case__ = model(_a , _a ) # 3. predict previous sample x_t-1 snake_case__ = scheduler.step(_a , _a , _a , generator=_a ).prev_sample snake_case__ = pred_prev_sample snake_case__ = torch.sum(torch.abs(_a ) ) snake_case__ = torch.mean(torch.abs(_a ) ) assert abs(result_sum.item() - 347.6357 ) < 1e-2 assert abs(result_mean.item() - 0.4527 ) < 1e-3 def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = [39, 30, 12, 15, 0] with self.assertRaises(_a , msg='''`timesteps` must be in descending order.''' ): scheduler.set_timesteps(timesteps=_a ) def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = [39, 30, 12, 1, 0] snake_case__ = len(_a ) with self.assertRaises(_a , msg='''Can only pass one of `num_inference_steps` or `timesteps`.''' ): scheduler.set_timesteps(num_inference_steps=_a , timesteps=_a ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): snake_case__ = self.scheduler_classes[0] snake_case__ = self.get_scheduler_config() snake_case__ = scheduler_class(**_a ) snake_case__ = [scheduler.config.num_train_timesteps] with self.assertRaises( _a , msg='''`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}''' , ): scheduler.set_timesteps(timesteps=_a )
33
0
import qiskit def __UpperCamelCase ( _A = 2 ): lowerCAmelCase_ = qubits # Using Aer's simulator lowerCAmelCase_ = qiskit.Aer.get_backend('''aer_simulator''' ) # Creating a Quantum Circuit acting on the q register lowerCAmelCase_ = qiskit.QuantumCircuit(__lowerCAmelCase , __lowerCAmelCase ) # Adding a H gate on qubit 0 (now q0 in superposition) circuit.h(0 ) for i in range(1 , __lowerCAmelCase ): # Adding CX (CNOT) gate circuit.cx(i - 1 , __lowerCAmelCase ) # Mapping the quantum measurement to the classical bits circuit.measure(list(range(__lowerCAmelCase ) ) , list(range(__lowerCAmelCase ) ) ) # Now measuring any one qubit would affect other qubits to collapse # their super position and have same state as the measured one. # Executing the circuit on the simulator lowerCAmelCase_ = qiskit.execute(__lowerCAmelCase , __lowerCAmelCase , shots=1000 ) return job.result().get_counts(__lowerCAmelCase ) if __name__ == "__main__": print(f"Total count for various states are: {quantum_entanglement(3)}")
431
import numpy as np def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> np.ndarray: return 1 / (1 + np.exp(-vector )) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> np.ndarray: return vector * sigmoid(__lowerCAmelCase ) if __name__ == "__main__": import doctest doctest.testmod()
33
0
import functools def UpperCamelCase( __UpperCamelCase : List[str] ,__UpperCamelCase : List[Any] ): # Validation if not isinstance(__lowerCAmelCase ,__lowerCAmelCase ) or not all(isinstance(__lowerCAmelCase ,__lowerCAmelCase ) for day in days ): raise ValueError('''The parameter days should be a list of integers''' ) if len(__lowerCAmelCase ) != 3 or not all(isinstance(__lowerCAmelCase ,__lowerCAmelCase ) for cost in costs ): raise ValueError('''The parameter costs should be a list of three integers''' ) if len(__lowerCAmelCase ) == 0: return 0 if min(__lowerCAmelCase ) <= 0: raise ValueError('''All days elements should be greater than 0''' ) if max(__lowerCAmelCase ) >= 366: raise ValueError('''All days elements should be less than 366''' ) lowerCAmelCase_ : Any = set(__lowerCAmelCase ) @functools.cache def dynamic_programming(__UpperCamelCase : Dict ) -> int: if index > 365: return 0 if index not in days_set: return dynamic_programming(index + 1 ) return min( costs[0] + dynamic_programming(index + 1 ) ,costs[1] + dynamic_programming(index + 7 ) ,costs[2] + dynamic_programming(index + 30 ) ,) return dynamic_programming(1 ) if __name__ == "__main__": import doctest doctest.testmod()
171
def SCREAMING_SNAKE_CASE ( __lowerCAmelCase = 100 ) -> int: snake_case__ = set() snake_case__ = 0 snake_case__ = n + 1 # maximum limit for a in range(2 , __lowerCAmelCase ): for b in range(2 , __lowerCAmelCase ): snake_case__ = a**b # calculates the current power collect_powers.add(__lowerCAmelCase ) # adds the result to the set return len(__lowerCAmelCase ) if __name__ == "__main__": print("""Number of terms """, solution(int(str(input()).strip())))
33
0
def SCREAMING_SNAKE_CASE_ ( UpperCAmelCase_ : List[Any] ) -> int: if not grid or not grid[0]: raise TypeError('''The grid does not contain the appropriate information''' ) for cell_n in range(1 , len(grid[0] ) ): grid[0][cell_n] += grid[0][cell_n - 1] SCREAMING_SNAKE_CASE_ : List[str] =grid[0] for row_n in range(1 , len(__lowerCAmelCase ) ): SCREAMING_SNAKE_CASE_ : int =grid[row_n] SCREAMING_SNAKE_CASE_ : Tuple =fill_row(__lowerCAmelCase , __lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : str =grid[row_n] return grid[-1][-1] def SCREAMING_SNAKE_CASE_ ( UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : Dict ) -> list: current_row[0] += row_above[0] for cell_n in range(1 , len(__lowerCAmelCase ) ): current_row[cell_n] += min(current_row[cell_n - 1] , row_above[cell_n] ) return current_row if __name__ == "__main__": import doctest doctest.testmod()
443
from copy import deepcopy class __magic_name__ : '''simple docstring''' def __init__( self:int , _a:list[int] | None = None , _a:int | None = None ): if arr is None and size is not None: snake_case__ = size snake_case__ = [0] * size elif arr is not None: self.init(_a ) else: raise ValueError('''Either arr or size must be specified''' ) def SCREAMING_SNAKE_CASE__ ( self:Any , _a:list[int] ): snake_case__ = len(_a ) snake_case__ = deepcopy(_a ) for i in range(1 , self.size ): snake_case__ = self.next_(_a ) if j < self.size: self.tree[j] += self.tree[i] def SCREAMING_SNAKE_CASE__ ( self:Any ): snake_case__ = self.tree[:] for i in range(self.size - 1 , 0 , -1 ): snake_case__ = self.next_(_a ) if j < self.size: arr[j] -= arr[i] return arr @staticmethod def SCREAMING_SNAKE_CASE__ ( _a:int ): return index + (index & (-index)) @staticmethod def SCREAMING_SNAKE_CASE__ ( _a:int ): return index - (index & (-index)) def SCREAMING_SNAKE_CASE__ ( self:List[Any] , _a:int , _a:int ): if index == 0: self.tree[0] += value return while index < self.size: self.tree[index] += value snake_case__ = self.next_(_a ) def SCREAMING_SNAKE_CASE__ ( self:Dict , _a:int , _a:int ): self.add(_a , value - self.get(_a ) ) def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] , _a:int ): if right == 0: return 0 snake_case__ = self.tree[0] right -= 1 # make right inclusive while right > 0: result += self.tree[right] snake_case__ = self.prev(_a ) return result def SCREAMING_SNAKE_CASE__ ( self:Dict , _a:int , _a:int ): return self.prefix(_a ) - self.prefix(_a ) def SCREAMING_SNAKE_CASE__ ( self:str , _a:int ): return self.query(_a , index + 1 ) def SCREAMING_SNAKE_CASE__ ( self:str , _a:int ): value -= self.tree[0] if value < 0: return -1 snake_case__ = 1 # Largest power of 2 <= size while j * 2 < self.size: j *= 2 snake_case__ = 0 while j > 0: if i + j < self.size and self.tree[i + j] <= value: value -= self.tree[i + j] i += j j //= 2 return i if __name__ == "__main__": import doctest doctest.testmod()
33
0
"""simple docstring""" from __future__ import annotations def _lowerCamelCase ( __a, __a, __a, ): if (stress, tangential_force, area).count(0 ) != 1: raise ValueError('''You cannot supply more or less than 2 values''' ) elif stress < 0: raise ValueError('''Stress cannot be negative''' ) elif tangential_force < 0: raise ValueError('''Tangential Force cannot be negative''' ) elif area < 0: raise ValueError('''Area cannot be negative''' ) elif stress == 0: return ( "stress", tangential_force / area, ) elif tangential_force == 0: return ( "tangential_force", stress * area, ) else: return ( "area", tangential_force / stress, ) if __name__ == "__main__": import doctest doctest.testmod()
626
from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFAutoModelForSeqaSeqLM, TFBlenderbotForConditionalGeneration, TFBlenderbotModel @require_tf class __magic_name__ : '''simple docstring''' __lowercase : int = BlenderbotConfig __lowercase : Any = {} __lowercase : Optional[Any] = 'gelu' def __init__( self:Tuple , _a:Optional[Any] , _a:Optional[Any]=13 , _a:Tuple=7 , _a:Union[str, Any]=True , _a:int=False , _a:int=99 , _a:Optional[int]=32 , _a:List[str]=2 , _a:List[str]=4 , _a:List[Any]=37 , _a:Any=0.1 , _a:int=0.1 , _a:List[Any]=20 , _a:List[str]=2 , _a:int=1 , _a:Dict=0 , ): snake_case__ = parent snake_case__ = batch_size snake_case__ = seq_length snake_case__ = is_training snake_case__ = use_labels snake_case__ = vocab_size snake_case__ = hidden_size snake_case__ = num_hidden_layers snake_case__ = num_attention_heads snake_case__ = intermediate_size snake_case__ = hidden_dropout_prob snake_case__ = attention_probs_dropout_prob snake_case__ = max_position_embeddings snake_case__ = eos_token_id snake_case__ = pad_token_id snake_case__ = bos_token_id def SCREAMING_SNAKE_CASE__ ( self:int ): snake_case__ = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) snake_case__ = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) snake_case__ = tf.concat([input_ids, eos_tensor] , axis=1 ) snake_case__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) snake_case__ = self.config_cls( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_ids=[2] , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , decoder_start_token_id=self.pad_token_id , **self.config_updates , ) snake_case__ = prepare_blenderbot_inputs_dict(_a , _a , _a ) return config, inputs_dict def SCREAMING_SNAKE_CASE__ ( self:int , _a:Optional[Any] , _a:int ): snake_case__ = TFBlenderbotModel(config=_a ).get_decoder() snake_case__ = inputs_dict['''input_ids'''] snake_case__ = input_ids[:1, :] snake_case__ = inputs_dict['''attention_mask'''][:1, :] snake_case__ = inputs_dict['''head_mask'''] snake_case__ = 1 # first forward pass snake_case__ = model(_a , attention_mask=_a , head_mask=_a , use_cache=_a ) snake_case__ , snake_case__ = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids snake_case__ = ids_tensor((self.batch_size, 3) , config.vocab_size ) snake_case__ = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and snake_case__ = tf.concat([input_ids, next_tokens] , axis=-1 ) snake_case__ = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) snake_case__ = model(_a , attention_mask=_a )[0] snake_case__ = model(_a , attention_mask=_a , past_key_values=_a )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice snake_case__ = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) snake_case__ = output_from_no_past[:, -3:, random_slice_idx] snake_case__ = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(_a , _a , rtol=1e-3 ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=None , ) -> Tuple: if attention_mask is None: snake_case__ = tf.cast(tf.math.not_equal(__lowerCAmelCase , config.pad_token_id ) , tf.inta ) if decoder_attention_mask is None: snake_case__ = tf.concat( [ tf.ones(decoder_input_ids[:, :1].shape , dtype=tf.inta ), tf.cast(tf.math.not_equal(decoder_input_ids[:, 1:] , config.pad_token_id ) , tf.inta ), ] , axis=-1 , ) if head_mask is None: snake_case__ = tf.ones((config.encoder_layers, config.encoder_attention_heads) ) if decoder_head_mask is None: snake_case__ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) if cross_attn_head_mask is None: snake_case__ = tf.ones((config.decoder_layers, config.decoder_attention_heads) ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": decoder_attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } @require_tf class __magic_name__ (snake_case_ ,snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : List[str] = (TFBlenderbotForConditionalGeneration, TFBlenderbotModel) if is_tf_available() else () __lowercase : Any = (TFBlenderbotForConditionalGeneration,) if is_tf_available() else () __lowercase : Tuple = ( { 'conversational': TFBlenderbotForConditionalGeneration, 'feature-extraction': TFBlenderbotModel, 'summarization': TFBlenderbotForConditionalGeneration, 'text2text-generation': TFBlenderbotForConditionalGeneration, 'translation': TFBlenderbotForConditionalGeneration, } if is_tf_available() else {} ) __lowercase : Any = True __lowercase : int = False __lowercase : int = False def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): snake_case__ = TFBlenderbotModelTester(self ) snake_case__ = ConfigTester(self , config_class=_a ) def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): snake_case__ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*_a ) @require_tokenizers @require_tf class __magic_name__ (unittest.TestCase ): '''simple docstring''' __lowercase : Optional[int] = ['My friends are cool but they eat too many carbs.'] __lowercase : Optional[int] = 'facebook/blenderbot-400M-distill' @cached_property def SCREAMING_SNAKE_CASE__ ( self:Tuple ): return BlenderbotTokenizer.from_pretrained(self.model_name ) @cached_property def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): snake_case__ = TFAutoModelForSeqaSeqLM.from_pretrained(self.model_name ) return model @slow def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = self.tokenizer(self.src_text , return_tensors='''tf''' ) snake_case__ = self.model.generate( model_inputs.input_ids , ) snake_case__ = self.tokenizer.batch_decode(generated_ids.numpy() , skip_special_tokens=_a )[0] assert ( generated_words == " That's unfortunate. Are they trying to lose weight or are they just trying to be healthier?" )
33
0
"""simple docstring""" from __future__ import annotations from statistics import mean def lowercase ( lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ): lowerCamelCase_ = [0] * no_of_processes lowerCamelCase_ = [0] * no_of_processes # Initialize remaining_time to waiting_time. for i in range(__lowerCAmelCase ): lowerCamelCase_ = burst_time[i] lowerCamelCase_ = [] lowerCamelCase_ = 0 lowerCamelCase_ = 0 # When processes are not completed, # A process whose arrival time has passed \ # and has remaining execution time is put into the ready_process. # The shortest process in the ready_process, target_process is executed. while completed != no_of_processes: lowerCamelCase_ = [] lowerCamelCase_ = -1 for i in range(__lowerCAmelCase ): if (arrival_time[i] <= total_time) and (remaining_time[i] > 0): ready_process.append(__lowerCAmelCase ) if len(__lowerCAmelCase ) > 0: lowerCamelCase_ = ready_process[0] for i in ready_process: if remaining_time[i] < remaining_time[target_process]: lowerCamelCase_ = i total_time += burst_time[target_process] completed += 1 lowerCamelCase_ = 0 lowerCamelCase_ = ( total_time - arrival_time[target_process] - burst_time[target_process] ) else: total_time += 1 return waiting_time def lowercase ( lowerCAmelCase__ ,lowerCAmelCase__ ,lowerCAmelCase__ ): lowerCamelCase_ = [0] * no_of_processes for i in range(__lowerCAmelCase ): lowerCamelCase_ = burst_time[i] + waiting_time[i] return turn_around_time if __name__ == "__main__": print("""[TEST CASE 01]""") A_ = 4 A_ = [2, 5, 3, 7] A_ = [0, 0, 0, 0] A_ = calculate_waitingtime(arrival_time, burst_time, no_of_processes) A_ = calculate_turnaroundtime( burst_time, no_of_processes, waiting_time ) # Printing the Result print("""PID\tBurst Time\tArrival Time\tWaiting Time\tTurnaround Time""") for i, process_id in enumerate(list(range(1, 5))): print( f"{process_id}\t{burst_time[i]}\t\t\t{arrival_time[i]}\t\t\t\t" f"{waiting_time[i]}\t\t\t\t{turn_around_time[i]}" ) print(f"\nAverage waiting time = {mean(waiting_time):.5f}") print(f"Average turnaround time = {mean(turn_around_time):.5f}")
29
import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class __magic_name__ (unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = 0 def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): snake_case__ = AutoImageProcessor.from_pretrained('''openai/clip-vit-base-patch32''' ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:str ): with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = Path(_a ) / '''preprocessor_config.json''' snake_case__ = Path(_a ) / '''config.json''' json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) json.dump({'''model_type''': '''clip'''} , open(_a , '''w''' ) ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = Path(_a ) / '''preprocessor_config.json''' snake_case__ = Path(_a ) / '''config.json''' json.dump( {'''feature_extractor_type''': '''CLIPFeatureExtractor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) json.dump({'''model_type''': '''clip'''} , open(_a , '''w''' ) ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = CLIPConfig() # Create a dummy config file with image_proceesor_type snake_case__ = Path(_a ) / '''preprocessor_config.json''' snake_case__ = Path(_a ) / '''config.json''' json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) json.dump({'''model_type''': '''clip'''} , open(_a , '''w''' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally snake_case__ = AutoImageProcessor.from_pretrained(_a ).to_dict() config_dict.pop('''image_processor_type''' ) snake_case__ = CLIPImageProcessor(**_a ) # save in new folder model_config.save_pretrained(_a ) config.save_pretrained(_a ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) # make sure private variable is not incorrectly saved snake_case__ = json.loads(config.to_json_string() ) self.assertTrue('''_processor_class''' not in dict_as_saved ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:List[str] ): with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = Path(_a ) / '''preprocessor_config.json''' json.dump( {'''image_processor_type''': '''CLIPImageProcessor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): with self.assertRaisesRegex( _a , '''clip-base is not a local folder and is not a valid model identifier''' ): snake_case__ = AutoImageProcessor.from_pretrained('''clip-base''' ) def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): with self.assertRaisesRegex( _a , r'''aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)''' ): snake_case__ = AutoImageProcessor.from_pretrained(_a , revision='''aaaaaa''' ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): with self.assertRaisesRegex( _a , '''hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.''' , ): snake_case__ = AutoImageProcessor.from_pretrained('''hf-internal-testing/config-no-model''' ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(_a ): snake_case__ = AutoImageProcessor.from_pretrained('''hf-internal-testing/test_dynamic_image_processor''' ) # If remote code is disabled, we can't load this config. with self.assertRaises(_a ): snake_case__ = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=_a ) snake_case__ = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=_a ) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_a ) snake_case__ = AutoImageProcessor.from_pretrained(_a , trust_remote_code=_a ) self.assertEqual(reloaded_image_processor.__class__.__name__ , '''NewImageProcessor''' ) def SCREAMING_SNAKE_CASE__ ( self:Dict ): try: AutoConfig.register('''custom''' , _a ) AutoImageProcessor.register(_a , _a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_a ): AutoImageProcessor.register(_a , _a ) with tempfile.TemporaryDirectory() as tmpdirname: snake_case__ = Path(_a ) / '''preprocessor_config.json''' snake_case__ = Path(_a ) / '''config.json''' json.dump( {'''feature_extractor_type''': '''CLIPFeatureExtractor''', '''processor_class''': '''CLIPProcessor'''} , open(_a , '''w''' ) , ) json.dump({'''model_type''': '''clip'''} , open(_a , '''w''' ) ) snake_case__ = CustomImageProcessor.from_pretrained(_a ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(_a ) snake_case__ = AutoImageProcessor.from_pretrained(_a ) self.assertIsInstance(_a , _a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : List[str] = True try: AutoConfig.register('''custom''' , _a ) AutoImageProcessor.register(_a , _a ) # If remote code is not set, the default is to use local snake_case__ = AutoImageProcessor.from_pretrained('''hf-internal-testing/test_dynamic_image_processor''' ) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. snake_case__ = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=_a ) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub snake_case__ = AutoImageProcessor.from_pretrained( '''hf-internal-testing/test_dynamic_image_processor''' , trust_remote_code=_a ) self.assertEqual(image_processor.__class__.__name__ , '''NewImageProcessor''' ) self.assertTrue(not hasattr(_a , '''is_local''' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
33
0
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Image from .base import TaskTemplate @dataclass(frozen=snake_case_ ) class _UpperCAmelCase( snake_case_ ): lowercase__ = field(default='image-classification' , metadata={'include_in_asdict_even_if_is_default': True} ) lowercase__ = Features({'image': Image()} ) lowercase__ = Features({'labels': ClassLabel} ) lowercase__ = "image" lowercase__ = "labels" def UpperCAmelCase ( self , __a) -> Optional[Any]: '''simple docstring''' if self.label_column not in features: raise ValueError(F'''Column {self.label_column} is not present in features.''') if not isinstance(features[self.label_column] , _a): raise ValueError(F'''Column {self.label_column} is not a ClassLabel.''') _UpperCamelCase = copy.deepcopy(self) _UpperCamelCase = self.label_schema.copy() _UpperCamelCase = features[self.label_column] _UpperCamelCase = label_schema return task_template @property def UpperCAmelCase ( self) -> List[Any]: '''simple docstring''' return { self.image_column: "image", self.label_column: "labels", }
19
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTImageProcessor, ViTConfig, ViTForImageClassification, ViTImageProcessor, ViTModel from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase__ : int = logging.get_logger(__name__) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase=False ) -> int: snake_case__ = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((F"""blocks.{i}.norm1.weight""", F"""vit.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""blocks.{i}.norm1.bias""", F"""vit.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append((F"""blocks.{i}.attn.proj.weight""", F"""vit.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.attn.proj.bias""", F"""vit.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append((F"""blocks.{i}.norm2.weight""", F"""vit.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""blocks.{i}.norm2.bias""", F"""vit.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.weight""", F"""vit.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc1.bias""", F"""vit.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.weight""", F"""vit.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""blocks.{i}.mlp.fc2.bias""", F"""vit.encoder.layer.{i}.output.dense.bias""") ) # projection layer + position embeddings rename_keys.extend( [ ('''cls_token''', '''vit.embeddings.cls_token'''), ('''patch_embed.proj.weight''', '''vit.embeddings.patch_embeddings.projection.weight'''), ('''patch_embed.proj.bias''', '''vit.embeddings.patch_embeddings.projection.bias'''), ('''pos_embed''', '''vit.embeddings.position_embeddings'''), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ('''norm.weight''', '''layernorm.weight'''), ('''norm.bias''', '''layernorm.bias'''), ('''pre_logits.fc.weight''', '''pooler.dense.weight'''), ('''pre_logits.fc.bias''', '''pooler.dense.bias'''), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" snake_case__ = [(pair[0], pair[1][4:]) if pair[1].startswith('''vit''' ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ('''norm.weight''', '''vit.layernorm.weight'''), ('''norm.bias''', '''vit.layernorm.bias'''), ('''head.weight''', '''classifier.weight'''), ('''head.bias''', '''classifier.bias'''), ] ) return rename_keys def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False ) -> Dict: for i in range(config.num_hidden_layers ): if base_model: snake_case__ = '''''' else: snake_case__ = '''vit.''' # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) snake_case__ = state_dict.pop(F"""blocks.{i}.attn.qkv.weight""" ) snake_case__ = state_dict.pop(F"""blocks.{i}.attn.qkv.bias""" ) # next, add query, keys and values (in that order) to the state dict snake_case__ = in_proj_weight[ : config.hidden_size, : ] snake_case__ = in_proj_bias[: config.hidden_size] snake_case__ = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] snake_case__ = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] snake_case__ = in_proj_weight[ -config.hidden_size :, : ] snake_case__ = in_proj_bias[-config.hidden_size :] def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> Optional[Any]: snake_case__ = ['''head.weight''', '''head.bias'''] for k in ignore_keys: state_dict.pop(__lowerCAmelCase , __lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Union[str, Any]: snake_case__ = dct.pop(__lowerCAmelCase ) snake_case__ = val def SCREAMING_SNAKE_CASE ( ) -> str: snake_case__ = '''http://images.cocodataset.org/val2017/000000039769.jpg''' snake_case__ = Image.open(requests.get(__lowerCAmelCase , stream=__lowerCAmelCase ).raw ) return im @torch.no_grad() def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> Dict: snake_case__ = ViTConfig() snake_case__ = False # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size if vit_name[-5:] == "in21k": snake_case__ = True snake_case__ = int(vit_name[-12:-10] ) snake_case__ = int(vit_name[-9:-6] ) else: snake_case__ = 1000 snake_case__ = '''huggingface/label-files''' snake_case__ = '''imagenet-1k-id2label.json''' snake_case__ = json.load(open(hf_hub_download(__lowerCAmelCase , __lowerCAmelCase , repo_type='''dataset''' ) , '''r''' ) ) snake_case__ = {int(__lowerCAmelCase ): v for k, v in idalabel.items()} snake_case__ = idalabel snake_case__ = {v: k for k, v in idalabel.items()} snake_case__ = int(vit_name[-6:-4] ) snake_case__ = int(vit_name[-3:] ) # size of the architecture if "deit" in vit_name: if vit_name[9:].startswith('''tiny''' ): snake_case__ = 192 snake_case__ = 768 snake_case__ = 12 snake_case__ = 3 elif vit_name[9:].startswith('''small''' ): snake_case__ = 384 snake_case__ = 1536 snake_case__ = 12 snake_case__ = 6 else: pass else: if vit_name[4:].startswith('''small''' ): snake_case__ = 768 snake_case__ = 2304 snake_case__ = 8 snake_case__ = 8 elif vit_name[4:].startswith('''base''' ): pass elif vit_name[4:].startswith('''large''' ): snake_case__ = 1024 snake_case__ = 4096 snake_case__ = 24 snake_case__ = 16 elif vit_name[4:].startswith('''huge''' ): snake_case__ = 1280 snake_case__ = 5120 snake_case__ = 32 snake_case__ = 16 # load original model from timm snake_case__ = timm.create_model(__lowerCAmelCase , pretrained=__lowerCAmelCase ) timm_model.eval() # load state_dict of original model, remove and rename some keys snake_case__ = timm_model.state_dict() if base_model: remove_classification_head_(__lowerCAmelCase ) snake_case__ = create_rename_keys(__lowerCAmelCase , __lowerCAmelCase ) for src, dest in rename_keys: rename_key(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) read_in_q_k_v(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # load HuggingFace model if vit_name[-5:] == "in21k": snake_case__ = ViTModel(__lowerCAmelCase ).eval() else: snake_case__ = ViTForImageClassification(__lowerCAmelCase ).eval() model.load_state_dict(__lowerCAmelCase ) # Check outputs on an image, prepared by ViTImageProcessor/DeiTImageProcessor if "deit" in vit_name: snake_case__ = DeiTImageProcessor(size=config.image_size ) else: snake_case__ = ViTImageProcessor(size=config.image_size ) snake_case__ = image_processor(images=prepare_img() , return_tensors='''pt''' ) snake_case__ = encoding['''pixel_values'''] snake_case__ = model(__lowerCAmelCase ) if base_model: snake_case__ = timm_model.forward_features(__lowerCAmelCase ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(__lowerCAmelCase , outputs.pooler_output , atol=1e-3 ) else: snake_case__ = timm_model(__lowerCAmelCase ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(__lowerCAmelCase , outputs.logits , atol=1e-3 ) Path(__lowerCAmelCase ).mkdir(exist_ok=__lowerCAmelCase ) print(F"""Saving model {vit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__lowerCAmelCase ) print(F"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__lowerCAmelCase ) if __name__ == "__main__": lowerCamelCase__ : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--vit_name""", default="""vit_base_patch16_224""", type=str, help="""Name of the ViT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) lowerCamelCase__ : str = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path)
33
0
import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py UpperCAmelCase__ = """src/diffusers""" UpperCAmelCase__ = """.""" # This is to make sure the diffusers module imported is the one in the repo. UpperCAmelCase__ = importlib.util.spec_from_file_location( "diffusers", os.path.join(DIFFUSERS_PATH, "__init__.py"), submodule_search_locations=[DIFFUSERS_PATH], ) UpperCAmelCase__ = spec.loader.load_module() def _a ( a :str , a :int ) -> Optional[Any]: return line.startswith(__lowerCAmelCase ) or len(__lowerCAmelCase ) <= 1 or re.search(r'''^\s*\)(\s*->.*:|:)\s*$''' , __lowerCAmelCase ) is not None def _a ( a :Any ) -> Optional[int]: a = object_name.split('''.''' ) a = 0 # First let's find the module where our object lives. a = parts[i] while i < len(__lowerCAmelCase ) and not os.path.isfile(os.path.join(__lowerCAmelCase , F"""{module}.py""" ) ): i += 1 if i < len(__lowerCAmelCase ): a = os.path.join(__lowerCAmelCase , parts[i] ) if i >= len(__lowerCAmelCase ): raise ValueError(F"""`object_name` should begin with the name of a module of diffusers but got {object_name}.""" ) with open(os.path.join(__lowerCAmelCase , F"""{module}.py""" ) , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: a = f.readlines() # Now let's find the class / func in the code! a = '''''' a = 0 for name in parts[i + 1 :]: while ( line_index < len(__lowerCAmelCase ) and re.search(rF"""^{indent}(class|def)\s+{name}(\(|\:)""" , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(__lowerCAmelCase ): raise ValueError(F""" {object_name} does not match any function or class in {module}.""" ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). a = line_index while line_index < len(__lowerCAmelCase ) and _should_continue(lines[line_index] , __lowerCAmelCase ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 a = lines[start_index:line_index] return "".join(__lowerCAmelCase ) UpperCAmelCase__ = re.compile(R"^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)") UpperCAmelCase__ = re.compile(R"^\s*(\S+)->(\S+)(\s+.*|$)") UpperCAmelCase__ = re.compile(R"<FILL\s+[^>]*>") def _a ( a :Dict ) -> Optional[Any]: a = code.split('''\n''' ) a = 0 while idx < len(__lowerCAmelCase ) and len(lines[idx] ) == 0: idx += 1 if idx < len(__lowerCAmelCase ): return re.search(r'''^(\s*)\S''' , lines[idx] ).groups()[0] return "" def _a ( a :Dict ) -> Optional[Any]: a = len(get_indent(__lowerCAmelCase ) ) > 0 if has_indent: a = F"""class Bla:\n{code}""" a = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=__lowerCAmelCase ) a = black.format_str(__lowerCAmelCase , mode=__lowerCAmelCase ) a , a = style_docstrings_in_code(__lowerCAmelCase ) return result[len('''class Bla:\n''' ) :] if has_indent else result def _a ( a :Union[str, Any] , a :List[Any]=False ) -> Any: with open(__lowerCAmelCase , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: a = f.readlines() a = [] a = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(__lowerCAmelCase ): a = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. a , a , a = search.groups() a = find_code_in_diffusers(__lowerCAmelCase ) a = get_indent(__lowerCAmelCase ) a = line_index + 1 if indent == theoretical_indent else line_index + 2 a = theoretical_indent a = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. a = True while line_index < len(__lowerCAmelCase ) and should_continue: line_index += 1 if line_index >= len(__lowerCAmelCase ): break a = lines[line_index] a = _should_continue(__lowerCAmelCase , __lowerCAmelCase ) and re.search(F"""^{indent}# End copy""" , __lowerCAmelCase ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 a = lines[start_index:line_index] a = ''''''.join(__lowerCAmelCase ) # Remove any nested `Copied from` comments to avoid circular copies a = [line for line in theoretical_code.split('''\n''' ) if _re_copy_warning.search(__lowerCAmelCase ) is None] a = '''\n'''.join(__lowerCAmelCase ) # Before comparing, use the `replace_pattern` on the original code. if len(__lowerCAmelCase ) > 0: a = replace_pattern.replace('''with''' , '''''' ).split(''',''' ) a = [_re_replace_pattern.search(__lowerCAmelCase ) for p in patterns] for pattern in patterns: if pattern is None: continue a , a , a = pattern.groups() a = re.sub(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) if option.strip() == "all-casing": a = re.sub(obja.lower() , obja.lower() , __lowerCAmelCase ) a = re.sub(obja.upper() , obja.upper() , __lowerCAmelCase ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line a = blackify(lines[start_index - 1] + theoretical_code ) a = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: a = lines[:start_index] + [theoretical_code] + lines[line_index:] a = start_index + 1 if overwrite and len(__lowerCAmelCase ) > 0: # Warn the user a file has been modified. print(F"""Detected changes, rewriting {filename}.""" ) with open(__lowerCAmelCase , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(__lowerCAmelCase ) return diffs def _a ( a :Dict = False ) -> str: a = glob.glob(os.path.join(__lowerCAmelCase , '''**/*.py''' ) , recursive=__lowerCAmelCase ) a = [] for filename in all_files: a = is_copy_consistent(__lowerCAmelCase , __lowerCAmelCase ) diffs += [F"""- {filename}: copy does not match {d[0]} at line {d[1]}""" for d in new_diffs] if not overwrite and len(__lowerCAmelCase ) > 0: a = '''\n'''.join(__lowerCAmelCase ) raise Exception( '''Found the following copy inconsistencies:\n''' + diff + '''\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.''' ) if __name__ == "__main__": UpperCAmelCase__ = argparse.ArgumentParser() parser.add_argument("--fix_and_overwrite", action="store_true", help="Whether to fix inconsistencies.") UpperCAmelCase__ = parser.parse_args() check_copies(args.fix_and_overwrite)
117
import re import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class __magic_name__ (snake_case_ ): '''simple docstring''' __lowercase : List[str] = ['image_processor', 'tokenizer'] __lowercase : str = 'AutoImageProcessor' __lowercase : Dict = 'AutoTokenizer' def __init__( self:int , _a:List[str]=None , _a:Optional[Any]=None , **_a:List[str] ): snake_case__ = None if "feature_extractor" in kwargs: warnings.warn( '''The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`''' ''' instead.''' , _a , ) snake_case__ = kwargs.pop('''feature_extractor''' ) snake_case__ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('''You need to specify an `image_processor`.''' ) if tokenizer is None: raise ValueError('''You need to specify a `tokenizer`.''' ) super().__init__(_a , _a ) snake_case__ = self.image_processor snake_case__ = False def __call__( self:Optional[int] , *_a:str , **_a:int ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*_a , **_a ) snake_case__ = kwargs.pop('''images''' , _a ) snake_case__ = kwargs.pop('''text''' , _a ) if len(_a ) > 0: snake_case__ = args[0] snake_case__ = args[1:] if images is None and text is None: raise ValueError('''You need to specify either an `images` or `text` input to process.''' ) if images is not None: snake_case__ = self.image_processor(_a , *_a , **_a ) if text is not None: snake_case__ = self.tokenizer(_a , **_a ) if text is None: return inputs elif images is None: return encodings else: snake_case__ = encodings['''input_ids'''] return inputs def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] , *_a:Union[str, Any] , **_a:Any ): return self.tokenizer.batch_decode(*_a , **_a ) def SCREAMING_SNAKE_CASE__ ( self:Tuple , *_a:Union[str, Any] , **_a:Optional[int] ): return self.tokenizer.decode(*_a , **_a ) @contextmanager def SCREAMING_SNAKE_CASE__ ( self:Tuple ): warnings.warn( '''`as_target_processor` is deprecated and will be removed in v5 of Transformers. You can process your ''' '''labels by using the argument `text` of the regular `__call__` method (either in the same call as ''' '''your images inputs, or in a separate call.''' ) snake_case__ = True snake_case__ = self.tokenizer yield snake_case__ = self.image_processor snake_case__ = False def SCREAMING_SNAKE_CASE__ ( self:List[str] , _a:Dict , _a:Dict=False , _a:Optional[int]=None ): if added_vocab is None: snake_case__ = self.tokenizer.get_added_vocab() snake_case__ = {} while tokens: snake_case__ = re.search(r'''<s_(.*?)>''' , _a , re.IGNORECASE ) if start_token is None: break snake_case__ = start_token.group(1 ) snake_case__ = re.search(rF"""</s_{key}>""" , _a , re.IGNORECASE ) snake_case__ = start_token.group() if end_token is None: snake_case__ = tokens.replace(_a , '''''' ) else: snake_case__ = end_token.group() snake_case__ = re.escape(_a ) snake_case__ = re.escape(_a ) snake_case__ = re.search(F"""{start_token_escaped}(.*?){end_token_escaped}""" , _a , re.IGNORECASE ) if content is not None: snake_case__ = content.group(1 ).strip() if r"<s_" in content and r"</s_" in content: # non-leaf node snake_case__ = self.tokenajson(_a , is_inner_value=_a , added_vocab=_a ) if value: if len(_a ) == 1: snake_case__ = value[0] snake_case__ = value else: # leaf nodes snake_case__ = [] for leaf in content.split(r'''<sep/>''' ): snake_case__ = leaf.strip() if leaf in added_vocab and leaf[0] == "<" and leaf[-2:] == "/>": snake_case__ = leaf[1:-2] # for categorical special tokens output[key].append(_a ) if len(output[key] ) == 1: snake_case__ = output[key][0] snake_case__ = tokens[tokens.find(_a ) + len(_a ) :].strip() if tokens[:6] == r"<sep/>": # non-leaf nodes return [output] + self.tokenajson(tokens[6:] , is_inner_value=_a , added_vocab=_a ) if len(_a ): return [output] if is_inner_value else output else: return [] if is_inner_value else {"text_sequence": tokens} @property def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): warnings.warn( '''`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.''' , _a , ) return self.image_processor_class @property def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): warnings.warn( '''`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.''' , _a , ) return self.image_processor
33
0
import os import pickle import unittest from transformers import AutoTokenizer from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.models.bert_japanese.tokenization_bert_japanese import ( VOCAB_FILES_NAMES, BertJapaneseTokenizer, CharacterTokenizer, JumanppTokenizer, MecabTokenizer, SudachiTokenizer, WordpieceTokenizer, ) from transformers.testing_utils import custom_tokenizers, require_jumanpp, require_sudachi from ...test_tokenization_common import TokenizerTesterMixin @custom_tokenizers class lowerCAmelCase__ ( snake_case_ , unittest.TestCase ): '''simple docstring''' lowercase_ = BertJapaneseTokenizer lowercase_ = False lowercase_ = True def __UpperCamelCase ( self ): '''simple docstring''' super().setUp() __A =[ '''[UNK]''', '''[CLS]''', '''[SEP]''', '''こんにちは''', '''こん''', '''にちは''', '''ばんは''', '''##こん''', '''##にちは''', '''##ばんは''', '''世界''', '''##世界''', '''、''', '''##、''', '''。''', '''##。''', ] __A =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __UpperCamelCase ( self , lowercase__ ): '''simple docstring''' __A ='''こんにちは、世界。 \nこんばんは、世界。''' __A ='''こんにちは 、 世界 。 こんばんは 、 世界 。''' return input_text, output_text def __UpperCamelCase ( self , lowercase__ ): '''simple docstring''' __A , __A =self.get_input_output_texts(_a ) __A =tokenizer.encode(_a , add_special_tokens=_a ) __A =tokenizer.decode(_a , clean_up_tokenization_spaces=_a ) return text, ids def __UpperCamelCase ( self ): '''simple docstring''' pass # TODO add if relevant def __UpperCamelCase ( self ): '''simple docstring''' pass # TODO add if relevant def __UpperCamelCase ( self ): '''simple docstring''' pass # TODO add if relevant def __UpperCamelCase ( self ): '''simple docstring''' __A =self.tokenizer_class(self.vocab_file ) __A =tokenizer.tokenize('''こんにちは、世界。\nこんばんは、世界。''' ) self.assertListEqual(_a , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) def __UpperCamelCase ( self ): '''simple docstring''' __A =self.tokenizer_class(self.vocab_file , word_tokenizer_type='''mecab''' ) self.assertIsNotNone(_a ) __A ='''こんにちは、世界。\nこんばんは、世界。''' __A =tokenizer.tokenize(_a ) self.assertListEqual(_a , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __A =os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(_a , '''wb''' ) as handle: pickle.dump(_a , _a ) with open(_a , '''rb''' ) as handle: __A =pickle.load(_a ) __A =tokenizer_new.tokenize(_a ) self.assertListEqual(_a , _a ) def __UpperCamelCase ( self ): '''simple docstring''' __A =MecabTokenizer(mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __UpperCamelCase ( self ): '''simple docstring''' try: __A =MecabTokenizer(mecab_dic='''unidic_lite''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __UpperCamelCase ( self ): '''simple docstring''' try: __A =MecabTokenizer(mecab_dic='''unidic''' ) except ModuleNotFoundError: return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __UpperCamelCase ( self ): '''simple docstring''' __A =MecabTokenizer(do_lower_case=_a , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iphone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) def __UpperCamelCase ( self ): '''simple docstring''' try: __A =MecabTokenizer( do_lower_case=_a , normalize_text=_a , mecab_option='''-d /usr/local/lib/mecab/dic/jumandic''' ) except RuntimeError: # if dict doesn't exist in the system, previous code raises this error. return self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) def __UpperCamelCase ( self ): '''simple docstring''' __A =MecabTokenizer(normalize_text=_a , mecab_dic='''ipadic''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップルストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。'''] , ) @require_sudachi def __UpperCamelCase ( self ): '''simple docstring''' __A =self.tokenizer_class(self.vocab_file , word_tokenizer_type='''sudachi''' ) self.assertIsNotNone(_a ) __A ='''こんにちは、世界。\nこんばんは、世界。''' __A =tokenizer.tokenize(_a ) self.assertListEqual(_a , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __A =os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(_a , '''wb''' ) as handle: pickle.dump(_a , _a ) with open(_a , '''rb''' ) as handle: __A =pickle.load(_a ) __A =tokenizer_new.tokenize(_a ) self.assertListEqual(_a , _a ) @require_sudachi def __UpperCamelCase ( self ): '''simple docstring''' __A =SudachiTokenizer(sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __UpperCamelCase ( self ): '''simple docstring''' __A =SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''A''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国''', '''人''', '''参政''', '''権'''] ) @require_sudachi def __UpperCamelCase ( self ): '''simple docstring''' __A =SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''B''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人''', '''参政権'''] ) @require_sudachi def __UpperCamelCase ( self ): '''simple docstring''' __A =SudachiTokenizer(sudachi_dict_type='''core''' , sudachi_split_mode='''C''' ) self.assertListEqual(tokenizer.tokenize('''外国人参政権''' ) , ['''外国人参政権'''] ) @require_sudachi def __UpperCamelCase ( self ): '''simple docstring''' __A =SudachiTokenizer(do_lower_case=_a , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', ''' ''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __UpperCamelCase ( self ): '''simple docstring''' __A =SudachiTokenizer(normalize_text=_a , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , [''' ''', '''\t''', '''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', ''' ''', '''が''', ''' ''', ''' ''', '''\n ''', '''発売''', '''さ''', '''れ''', '''た''', '''\u3000''', '''。''', ''' ''', ''' '''] , ) @require_sudachi def __UpperCamelCase ( self ): '''simple docstring''' __A =SudachiTokenizer(trim_whitespace=_a , sudachi_dict_type='''core''' ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れ''', '''た''', '''。'''] , ) @require_jumanpp def __UpperCamelCase ( self ): '''simple docstring''' __A =self.tokenizer_class(self.vocab_file , word_tokenizer_type='''jumanpp''' ) self.assertIsNotNone(_a ) __A ='''こんにちは、世界。\nこんばんは、世界。''' __A =tokenizer.tokenize(_a ) self.assertListEqual(_a , ['''こんにちは''', '''、''', '''世界''', '''。''', '''こん''', '''##ばんは''', '''、''', '''世界''', '''。'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_a ) , [3, 1_2, 1_0, 1_4, 4, 9, 1_2, 1_0, 1_4] ) __A =os.path.join(self.tmpdirname , '''tokenizer.bin''' ) with open(_a , '''wb''' ) as handle: pickle.dump(_a , _a ) with open(_a , '''rb''' ) as handle: __A =pickle.load(_a ) __A =tokenizer_new.tokenize(_a ) self.assertListEqual(_a , _a ) @require_jumanpp def __UpperCamelCase ( self ): '''simple docstring''' __A =JumanppTokenizer() self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __UpperCamelCase ( self ): '''simple docstring''' __A =JumanppTokenizer(do_lower_case=_a ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iphone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __UpperCamelCase ( self ): '''simple docstring''' __A =JumanppTokenizer(normalize_text=_a ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''ア''', '''ッ''', '''フ''', '''゚''', '''ル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''\u3000''', '''が''', '''\u3000''', '''\u3000''', '''\u3000''', '''発売''', '''さ''', '''れた''', '''\u3000''', '''。'''] , ) @require_jumanpp def __UpperCamelCase ( self ): '''simple docstring''' __A =JumanppTokenizer(trim_whitespace=_a ) self.assertListEqual( tokenizer.tokenize(''' \tアップルストアでiPhone8 が \n 発売された 。 ''' ) , ['''アップル''', '''ストア''', '''で''', '''iPhone''', '''8''', '''が''', '''発売''', '''さ''', '''れた''', '''。'''] , ) @require_jumanpp def __UpperCamelCase ( self ): '''simple docstring''' __A =JumanppTokenizer() self.assertListEqual( tokenizer.tokenize('''ありがとうございますm(_ _)m見つけるのが大変です。''' ) , ['''ありがとう''', '''ございます''', '''m(_ _)m''', '''見つける''', '''の''', '''が''', '''大変です''', '''。'''] , ) def __UpperCamelCase ( self ): '''simple docstring''' __A =['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こんにちは''', '''こん''', '''にちは''', '''ばんは''', '''##こん''', '''##にちは''', '''##ばんは'''] __A ={} for i, token in enumerate(_a ): __A =i __A =WordpieceTokenizer(vocab=_a , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こんにちは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは''' ) , ['''こん''', '''##ばんは'''] ) self.assertListEqual(tokenizer.tokenize('''こんばんは こんばんにちは こんにちは''' ) , ['''こん''', '''##ばんは''', '''[UNK]''', '''こんにちは'''] ) def __UpperCamelCase ( self ): '''simple docstring''' __A =BertJapaneseTokenizer.from_pretrained('''nlp-waseda/roberta-base-japanese-with-auto-jumanpp''' ) __A =tokenizer.subword_tokenizer __A =subword_tokenizer.tokenize('''国境 の 長い トンネル を 抜ける と 雪国 であった 。''' ) self.assertListEqual(_a , ['''▁国境''', '''▁の''', '''▁長い''', '''▁トンネル''', '''▁を''', '''▁抜ける''', '''▁と''', '''▁雪''', '''国''', '''▁であった''', '''▁。'''] ) __A =subword_tokenizer.tokenize('''こんばんは こんばん にち は こんにちは''' ) self.assertListEqual(_a , ['''▁こん''', '''ばん''', '''は''', '''▁こん''', '''ばん''', '''▁に''', '''ち''', '''▁は''', '''▁こんにちは'''] ) def __UpperCamelCase ( self ): '''simple docstring''' __A =self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese''' ) __A =tokenizer.encode('''ありがとう。''' , add_special_tokens=_a ) __A =tokenizer.encode('''どういたしまして。''' , add_special_tokens=_a ) __A =tokenizer.build_inputs_with_special_tokens(_a ) __A =tokenizer.build_inputs_with_special_tokens(_a , _a ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class lowerCAmelCase__ ( snake_case_ , unittest.TestCase ): '''simple docstring''' lowercase_ = BertJapaneseTokenizer lowercase_ = False def __UpperCamelCase ( self ): '''simple docstring''' super().setUp() __A =['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こ''', '''ん''', '''に''', '''ち''', '''は''', '''ば''', '''世''', '''界''', '''、''', '''。'''] __A =os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file'''] ) with open(self.vocab_file , '''w''' , encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) def __UpperCamelCase ( self , **lowercase__ ): '''simple docstring''' return BertJapaneseTokenizer.from_pretrained(self.tmpdirname , subword_tokenizer_type='''character''' , **_a ) def __UpperCamelCase ( self , lowercase__ ): '''simple docstring''' __A ='''こんにちは、世界。 \nこんばんは、世界。''' __A ='''こ ん に ち は 、 世 界 。 こ ん ば ん は 、 世 界 。''' return input_text, output_text def __UpperCamelCase ( self ): '''simple docstring''' pass # TODO add if relevant def __UpperCamelCase ( self ): '''simple docstring''' pass # TODO add if relevant def __UpperCamelCase ( self ): '''simple docstring''' pass # TODO add if relevant def __UpperCamelCase ( self ): '''simple docstring''' __A =self.tokenizer_class(self.vocab_file , subword_tokenizer_type='''character''' ) __A =tokenizer.tokenize('''こんにちは、世界。 \nこんばんは、世界。''' ) self.assertListEqual( _a , ['''こ''', '''ん''', '''に''', '''ち''', '''は''', '''、''', '''世''', '''界''', '''。''', '''こ''', '''ん''', '''ば''', '''ん''', '''は''', '''、''', '''世''', '''界''', '''。'''] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(_a ) , [3, 4, 5, 6, 7, 1_1, 9, 1_0, 1_2, 3, 4, 8, 4, 7, 1_1, 9, 1_0, 1_2] ) def __UpperCamelCase ( self ): '''simple docstring''' __A =['''[UNK]''', '''[CLS]''', '''[SEP]''', '''こ''', '''ん''', '''に''', '''ち''', '''は''', '''ば''', '''世''', '''界''', '''、''', '''。'''] __A ={} for i, token in enumerate(_a ): __A =i __A =CharacterTokenizer(vocab=_a , unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) , [] ) self.assertListEqual(tokenizer.tokenize('''こんにちは''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''は'''] ) self.assertListEqual(tokenizer.tokenize('''こんにちほ''' ) , ['''こ''', '''ん''', '''に''', '''ち''', '''[UNK]'''] ) def __UpperCamelCase ( self ): '''simple docstring''' __A =self.tokenizer_class.from_pretrained('''cl-tohoku/bert-base-japanese-char''' ) __A =tokenizer.encode('''ありがとう。''' , add_special_tokens=_a ) __A =tokenizer.encode('''どういたしまして。''' , add_special_tokens=_a ) __A =tokenizer.build_inputs_with_special_tokens(_a ) __A =tokenizer.build_inputs_with_special_tokens(_a , _a ) # 2 is for "[CLS]", 3 is for "[SEP]" assert encoded_sentence == [2] + text + [3] assert encoded_pair == [2] + text + [3] + text_a + [3] @custom_tokenizers class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def __UpperCamelCase ( self ): '''simple docstring''' __A ='''cl-tohoku/bert-base-japanese''' __A =AutoTokenizer.from_pretrained(_a ) self.assertIsInstance(_a , _a ) class lowerCAmelCase__ ( unittest.TestCase ): '''simple docstring''' def __UpperCamelCase ( self ): '''simple docstring''' __A ='''cl-tohoku/bert-base-japanese''' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertTokenizer.from_pretrained(_a ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) ) __A ='''bert-base-cased''' with self.assertLogs('''transformers''' , level='''WARNING''' ) as cm: BertJapaneseTokenizer.from_pretrained(_a ) self.assertTrue( cm.records[0].message.startswith( '''The tokenizer class you load from this checkpoint is not the same type as the class this function''' ''' is called from.''' ) )
184
import inspect import unittest from transformers import BitConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_backbone_common import BackboneTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import BitBackbone, BitForImageClassification, BitImageProcessor, BitModel from transformers.models.bit.modeling_bit import BIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image class __magic_name__ : '''simple docstring''' def __init__( self:Optional[Any] , _a:int , _a:str=3 , _a:Optional[int]=32 , _a:Optional[Any]=3 , _a:Tuple=10 , _a:List[Any]=[8, 16, 32, 64] , _a:str=[1, 1, 2, 1] , _a:Any=True , _a:List[Any]=True , _a:List[str]="relu" , _a:int=3 , _a:Tuple=None , _a:Tuple=["stage2", "stage3", "stage4"] , _a:List[Any]=[2, 3, 4] , _a:Union[str, Any]=1 , ): snake_case__ = parent snake_case__ = batch_size snake_case__ = image_size snake_case__ = num_channels snake_case__ = embeddings_size snake_case__ = hidden_sizes snake_case__ = depths snake_case__ = is_training snake_case__ = use_labels snake_case__ = hidden_act snake_case__ = num_labels snake_case__ = scope snake_case__ = len(_a ) snake_case__ = out_features snake_case__ = out_indices snake_case__ = num_groups def SCREAMING_SNAKE_CASE__ ( self:int ): snake_case__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) snake_case__ = None if self.use_labels: snake_case__ = ids_tensor([self.batch_size] , self.num_labels ) snake_case__ = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): return BitConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , out_features=self.out_features , out_indices=self.out_indices , num_groups=self.num_groups , ) def SCREAMING_SNAKE_CASE__ ( self:Any , _a:Optional[int] , _a:Tuple , _a:int ): snake_case__ = BitModel(config=_a ) model.to(_a ) model.eval() snake_case__ = model(_a ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def SCREAMING_SNAKE_CASE__ ( self:int , _a:Tuple , _a:Any , _a:Union[str, Any] ): snake_case__ = self.num_labels snake_case__ = BitForImageClassification(_a ) model.to(_a ) model.eval() snake_case__ = model(_a , labels=_a ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self:str , _a:str , _a:List[str] , _a:Any ): snake_case__ = BitBackbone(config=_a ) model.to(_a ) model.eval() snake_case__ = model(_a ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None snake_case__ = None snake_case__ = BitBackbone(config=_a ) model.to(_a ) model.eval() snake_case__ = model(_a ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = self.prepare_config_and_inputs() snake_case__ , snake_case__ , snake_case__ = config_and_inputs snake_case__ = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class __magic_name__ (snake_case_ ,snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : Any = (BitModel, BitForImageClassification, BitBackbone) if is_torch_available() else () __lowercase : int = ( {'feature-extraction': BitModel, 'image-classification': BitForImageClassification} if is_torch_available() else {} ) __lowercase : Tuple = False __lowercase : Optional[Any] = False __lowercase : str = False __lowercase : Tuple = False __lowercase : Tuple = False def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): snake_case__ = BitModelTester(self ) snake_case__ = ConfigTester(self , config_class=_a , has_text_modality=_a ) def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): return @unittest.skip(reason='''Bit does not output attentions''' ) def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): pass @unittest.skip(reason='''Bit does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): pass @unittest.skip(reason='''Bit does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): pass def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ , snake_case__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case__ = model_class(_a ) snake_case__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic snake_case__ = [*signature.parameters.keys()] snake_case__ = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , _a ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def SCREAMING_SNAKE_CASE__ ( self:Union[str, Any] ): snake_case__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_backbone(*_a ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): snake_case__ , snake_case__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: snake_case__ = model_class(config=_a ) for name, module in model.named_modules(): if isinstance(_a , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F"""Parameter {name} of model {model_class} seems not properly initialized""" , ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] ): def check_hidden_states_output(_a:List[Any] , _a:int , _a:Union[str, Any] ): snake_case__ = model_class(_a ) model.to(_a ) model.eval() with torch.no_grad(): snake_case__ = model(**self._prepare_for_class(_a , _a ) ) snake_case__ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states snake_case__ = self.model_tester.num_stages self.assertEqual(len(_a ) , expected_num_stages + 1 ) # Bit's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) snake_case__ , snake_case__ = self.model_tester.prepare_config_and_inputs_for_common() snake_case__ = ['''preactivation''', '''bottleneck'''] for model_class in self.all_model_classes: for layer_type in layers_type: snake_case__ = layer_type snake_case__ = True check_hidden_states_output(_a , _a , _a ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] snake_case__ = True check_hidden_states_output(_a , _a , _a ) @unittest.skip(reason='''Bit does not use feedforward chunking''' ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): pass def SCREAMING_SNAKE_CASE__ ( self:List[str] ): snake_case__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_a ) @slow def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] ): for model_name in BIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: snake_case__ = BitModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def SCREAMING_SNAKE_CASE ( ) -> Any: snake_case__ = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class __magic_name__ (unittest.TestCase ): '''simple docstring''' @cached_property def SCREAMING_SNAKE_CASE__ ( self:Tuple ): return ( BitImageProcessor.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def SCREAMING_SNAKE_CASE__ ( self:Optional[int] ): snake_case__ = BitForImageClassification.from_pretrained(BIT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(_a ) snake_case__ = self.default_image_processor snake_case__ = prepare_img() snake_case__ = image_processor(images=_a , return_tensors='''pt''' ).to(_a ) # forward pass with torch.no_grad(): snake_case__ = model(**_a ) # verify the logits snake_case__ = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , _a ) snake_case__ = torch.tensor([[-0.6526, -0.5263, -1.4398]] ).to(_a ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , _a , atol=1e-4 ) ) @require_torch class __magic_name__ (snake_case_ ,unittest.TestCase ): '''simple docstring''' __lowercase : Optional[Any] = (BitBackbone,) if is_torch_available() else () __lowercase : int = BitConfig __lowercase : Any = False def SCREAMING_SNAKE_CASE__ ( self:str ): snake_case__ = BitModelTester(self )
33
0
from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging SCREAMING_SNAKE_CASE_ : str = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ : Optional[int] = { """t5-small""": """https://huggingface.co/t5-small/resolve/main/config.json""", """t5-base""": """https://huggingface.co/t5-base/resolve/main/config.json""", """t5-large""": """https://huggingface.co/t5-large/resolve/main/config.json""", """t5-3b""": """https://huggingface.co/t5-3b/resolve/main/config.json""", """t5-11b""": """https://huggingface.co/t5-11b/resolve/main/config.json""", } class snake_case_ ( snake_case_ ): '''simple docstring''' __UpperCamelCase = 't5' __UpperCamelCase = ['past_key_values'] __UpperCamelCase = {'hidden_size': 'd_model', 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers'} def __init__( self : Tuple , __lowerCamelCase : Any=32_128 , __lowerCamelCase : Union[str, Any]=512 , __lowerCamelCase : str=64 , __lowerCamelCase : List[Any]=2_048 , __lowerCamelCase : Dict=6 , __lowerCamelCase : Tuple=None , __lowerCamelCase : str=8 , __lowerCamelCase : List[str]=32 , __lowerCamelCase : Dict=128 , __lowerCamelCase : Any=0.1 , __lowerCamelCase : List[str]=1E-6 , __lowerCamelCase : List[Any]=1.0 , __lowerCamelCase : Tuple="relu" , __lowerCamelCase : Union[str, Any]=True , __lowerCamelCase : int=True , __lowerCamelCase : Dict=0 , __lowerCamelCase : Optional[Any]=1 , **__lowerCamelCase : str , ) -> Optional[int]: '''simple docstring''' __lowercase = vocab_size __lowercase = d_model __lowercase = d_kv __lowercase = d_ff __lowercase = num_layers __lowercase = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry __lowercase = num_heads __lowercase = relative_attention_num_buckets __lowercase = relative_attention_max_distance __lowercase = dropout_rate __lowercase = layer_norm_epsilon __lowercase = initializer_factor __lowercase = feed_forward_proj __lowercase = use_cache __lowercase = self.feed_forward_proj.split('-' ) __lowercase = act_info[-1] __lowercase = act_info[0] == 'gated' if len(_a ) > 1 and act_info[0] != "gated" or len(_a ) > 2: raise ValueError( F"`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer." 'Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ' '\'gated-gelu\' or \'relu\'' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": __lowercase = 'gelu_new' super().__init__( pad_token_id=_a , eos_token_id=_a , is_encoder_decoder=_a , **_a , ) class snake_case_ ( snake_case_ ): '''simple docstring''' @property def UpperCAmelCase ( self : int ) -> List[str]: '''simple docstring''' __lowercase = { 'input_ids': {0: 'batch', 1: 'encoder_sequence'}, 'attention_mask': {0: 'batch', 1: 'encoder_sequence'}, } if self.use_past: __lowercase = 'past_encoder_sequence + sequence' __lowercase = {0: 'batch'} __lowercase = {0: 'batch', 1: 'past_decoder_sequence + sequence'} else: __lowercase = {0: 'batch', 1: 'decoder_sequence'} __lowercase = {0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(_a , direction='inputs' ) return common_inputs @property def UpperCAmelCase ( self : Dict ) -> List[str]: '''simple docstring''' return 13
375
import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging lowerCamelCase__ : Any = """\ """ lowerCamelCase__ : List[str] = """ Perplexity (PPL) is one of the most common metrics for evaluating language models. It is defined as the exponentiated average negative log-likelihood of a sequence. For more information, see https://huggingface.co/docs/transformers/perplexity """ lowerCamelCase__ : Any = """ Args: model_id (str): model used for calculating Perplexity NOTE: Perplexity can only be calculated for causal language models. This includes models such as gpt2, causal variations of bert, causal versions of t5, and more (the full list can be found in the AutoModelForCausalLM documentation here: https://huggingface.co/docs/transformers/master/en/model_doc/auto#transformers.AutoModelForCausalLM ) input_texts (list of str): input text, each separate text snippet is one list entry. batch_size (int): the batch size to run texts through the model. Defaults to 16. add_start_token (bool): whether to add the start token to the texts, so the perplexity can include the probability of the first word. Defaults to True. device (str): device to run on, defaults to 'cuda' when available Returns: perplexity: dictionary containing the perplexity scores for the texts in the input list, as well as the mean perplexity. If one of the input texts is longer than the max input length of the model, then it is truncated to the max length for the perplexity computation. Examples: Example 1: >>> perplexity = datasets.load_metric(\"perplexity\") >>> input_texts = [\"lorem ipsum\", \"Happy Birthday!\", \"Bienvenue\"] >>> results = perplexity.compute(model_id='gpt2', ... add_start_token=False, ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) ['perplexities', 'mean_perplexity'] >>> print(round(results[\"mean_perplexity\"], 2)) 78.22 >>> print(round(results[\"perplexities\"][0], 2)) 11.11 Example 2: >>> perplexity = datasets.load_metric(\"perplexity\") >>> input_texts = datasets.load_dataset(\"wikitext\", ... \"wikitext-2-raw-v1\", ... split=\"test\")[\"text\"][:50] # doctest:+ELLIPSIS [...] >>> input_texts = [s for s in input_texts if s!=''] >>> results = perplexity.compute(model_id='gpt2', ... input_texts=input_texts) # doctest:+ELLIPSIS >>> print(list(results.keys())) ['perplexities', 'mean_perplexity'] >>> print(round(results[\"mean_perplexity\"], 2)) 60.35 >>> print(round(results[\"perplexities\"][0], 2)) 81.12 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class __magic_name__ (datasets.Metric ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self:int ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''input_texts''': datasets.Value('''string''' ), } ) , reference_urls=['''https://huggingface.co/docs/transformers/perplexity'''] , ) def SCREAMING_SNAKE_CASE__ ( self:List[Any] , _a:int , _a:List[Any] , _a:int = 16 , _a:bool = True , _a:Any=None ): if device is not None: assert device in ["gpu", "cpu", "cuda"], "device should be either gpu or cpu." if device == "gpu": snake_case__ = '''cuda''' else: snake_case__ = '''cuda''' if torch.cuda.is_available() else '''cpu''' snake_case__ = AutoModelForCausalLM.from_pretrained(_a ) snake_case__ = model.to(_a ) snake_case__ = AutoTokenizer.from_pretrained(_a ) # if batch_size > 1 (which generally leads to padding being required), and # if there is not an already assigned pad_token, assign an existing # special token to also be the padding token if tokenizer.pad_token is None and batch_size > 1: snake_case__ = list(tokenizer.special_tokens_map_extended.values() ) # check that the model already has at least one special token defined assert ( len(_a ) > 0 ), "If batch_size > 1, model must have at least one special token to use for padding. Please use a different model or set batch_size=1." # assign one of the special tokens to also be the pad token tokenizer.add_special_tokens({'''pad_token''': existing_special_tokens[0]} ) if add_start_token: # leave room for <BOS> token to be added: assert ( tokenizer.bos_token is not None ), "Input model must already have a BOS token if using add_start_token=True. Please use a different model, or set add_start_token=False" snake_case__ = model.config.max_length - 1 else: snake_case__ = model.config.max_length snake_case__ = tokenizer( _a , add_special_tokens=_a , padding=_a , truncation=_a , max_length=_a , return_tensors='''pt''' , return_attention_mask=_a , ).to(_a ) snake_case__ = encodings['''input_ids'''] snake_case__ = encodings['''attention_mask'''] # check that each input is long enough: if add_start_token: assert torch.all(torch.ge(attn_masks.sum(1 ) , 1 ) ), "Each input text must be at least one token long." else: assert torch.all( torch.ge(attn_masks.sum(1 ) , 2 ) ), "When add_start_token=False, each input text must be at least two tokens long. Run with add_start_token=True if inputting strings of only one token, and remove all empty input strings." snake_case__ = [] snake_case__ = CrossEntropyLoss(reduction='''none''' ) for start_index in logging.tqdm(range(0 , len(_a ) , _a ) ): snake_case__ = min(start_index + batch_size , len(_a ) ) snake_case__ = encoded_texts[start_index:end_index] snake_case__ = attn_masks[start_index:end_index] if add_start_token: snake_case__ = torch.tensor([[tokenizer.bos_token_id]] * encoded_batch.size(dim=0 ) ).to(_a ) snake_case__ = torch.cat([bos_tokens_tensor, encoded_batch] , dim=1 ) snake_case__ = torch.cat( [torch.ones(bos_tokens_tensor.size() , dtype=torch.intaa ).to(_a ), attn_mask] , dim=1 ) snake_case__ = encoded_batch with torch.no_grad(): snake_case__ = model(_a , attention_mask=_a ).logits snake_case__ = out_logits[..., :-1, :].contiguous() snake_case__ = labels[..., 1:].contiguous() snake_case__ = attn_mask[..., 1:].contiguous() snake_case__ = torch.expa( (loss_fct(shift_logits.transpose(1 , 2 ) , _a ) * shift_attention_mask_batch).sum(1 ) / shift_attention_mask_batch.sum(1 ) ) ppls += perplexity_batch.tolist() return {"perplexities": ppls, "mean_perplexity": np.mean(_a )}
33
0
from collections.abc import Sequence def lowerCAmelCase__ ( _a : int , _a : Optional[Any] = False ): if not arr: return 0 snake_case_ : Tuple = 0 if allow_empty_subarrays else float("-inf" ) snake_case_ : List[str] = 0.0 for num in arr: snake_case_ : List[str] = max(0 if allow_empty_subarrays else num , curr_sum + num ) snake_case_ : int = max(__lowerCAmelCase , __lowerCAmelCase ) return max_sum if __name__ == "__main__": from doctest import testmod testmod() lowercase : Any = [-2, 1, -3, 4, -1, 2, 1, -5, 4] print(F"""{max_subarray_sum(nums) = }""")
568
import os from datetime import datetime as dt from github import Github lowerCamelCase__ : int = [ """good first issue""", """good second issue""", """good difficult issue""", """enhancement""", """new pipeline/model""", """new scheduler""", """wip""", ] def SCREAMING_SNAKE_CASE ( ) -> Optional[Any]: snake_case__ = Github(os.environ['''GITHUB_TOKEN'''] ) snake_case__ = g.get_repo('''huggingface/diffusers''' ) snake_case__ = repo.get_issues(state='''open''' ) for issue in open_issues: snake_case__ = sorted(issue.get_comments() , key=lambda __lowerCAmelCase : i.created_at , reverse=__lowerCAmelCase ) snake_case__ = comments[0] if len(__lowerCAmelCase ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Closes the issue after 7 days of inactivity since the Stalebot notification. issue.edit(state='''closed''' ) elif ( "stale" in issue.get_labels() and last_comment is not None and last_comment.user.login != "github-actions[bot]" ): # Opens the issue if someone other than Stalebot commented. issue.edit(state='''open''' ) issue.remove_from_labels('''stale''' ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # Post a Stalebot notification after 23 days of inactivity. issue.create_comment( '''This issue has been automatically marked as stale because it has not had ''' '''recent activity. If you think this still needs to be addressed ''' '''please comment on this thread.\n\nPlease note that issues that do not follow the ''' '''[contributing guidelines](https://github.com/huggingface/diffusers/blob/main/CONTRIBUTING.md) ''' '''are likely to be ignored.''' ) issue.add_to_labels('''stale''' ) if __name__ == "__main__": main()
33
0
import importlib.util import os import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import ( is_accelerate_available, is_flax_available, is_safetensors_available, is_tf_available, is_torch_available, ) from . import BaseTransformersCLICommand def _lowercase ( a__ : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" return EnvironmentCommand() def _lowercase ( a__ : Optional[Any] ) -> List[str]: """simple docstring""" return EnvironmentCommand(args.accelerate_config_file ) class lowerCamelCase_ ( snake_case_ ): @staticmethod def lowercase ( lowerCamelCase_ ) -> int: """simple docstring""" _UpperCamelCase = parser.add_parser("env" ) download_parser.set_defaults(func=_a ) download_parser.add_argument( "--accelerate-config_file" , default=_a , help="The accelerate config file to use for the default values in the launching script." , ) download_parser.set_defaults(func=_a ) def __init__( self , lowerCamelCase_ , *lowerCamelCase_ ) -> Any: """simple docstring""" _UpperCamelCase = accelerate_config_file def lowercase ( self ) -> List[str]: """simple docstring""" _UpperCamelCase = "not installed" if is_safetensors_available(): import safetensors _UpperCamelCase = safetensors.__version__ elif importlib.util.find_spec("safetensors" ) is not None: import safetensors _UpperCamelCase = f'''{safetensors.__version__} but is ignored because of PyTorch version too old.''' _UpperCamelCase = "not installed" _UpperCamelCase = _UpperCamelCase = "not found" if is_accelerate_available(): import accelerate from accelerate.commands.config import default_config_file, load_config_from_file _UpperCamelCase = accelerate.__version__ # Get the default from the config file. if self._accelerate_config_file is not None or os.path.isfile(_a ): _UpperCamelCase = load_config_from_file(self._accelerate_config_file ).to_dict() _UpperCamelCase = ( "\n".join([f'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(_a , _a ) else f'''\t{accelerate_config}''' ) _UpperCamelCase = "not installed" _UpperCamelCase = "NA" if is_torch_available(): import torch _UpperCamelCase = torch.__version__ _UpperCamelCase = torch.cuda.is_available() _UpperCamelCase = "not installed" _UpperCamelCase = "NA" if is_tf_available(): import tensorflow as tf _UpperCamelCase = tf.__version__ try: # deprecated in v2.1 _UpperCamelCase = tf.test.is_gpu_available() except AttributeError: # returns list of devices, convert to bool _UpperCamelCase = bool(tf.config.list_physical_devices("GPU" ) ) _UpperCamelCase = "not installed" _UpperCamelCase = "not installed" _UpperCamelCase = "not installed" _UpperCamelCase = "NA" if is_flax_available(): import flax import jax import jaxlib _UpperCamelCase = flax.__version__ _UpperCamelCase = jax.__version__ _UpperCamelCase = jaxlib.__version__ _UpperCamelCase = jax.lib.xla_bridge.get_backend().platform _UpperCamelCase = { "`transformers` version": version, "Platform": platform.platform(), "Python version": platform.python_version(), "Huggingface_hub version": huggingface_hub.__version__, "Safetensors version": f'''{safetensors_version}''', "Accelerate version": f'''{accelerate_version}''', "Accelerate config": f'''{accelerate_config_str}''', "PyTorch version (GPU?)": f'''{pt_version} ({pt_cuda_available})''', "Tensorflow version (GPU?)": f'''{tf_version} ({tf_cuda_available})''', "Flax version (CPU?/GPU?/TPU?)": f'''{flax_version} ({jax_backend})''', "Jax version": f'''{jax_version}''', "JaxLib version": f'''{jaxlib_version}''', "Using GPU in script?": "<fill in>", "Using distributed or parallel set-up in script?": "<fill in>", } print("\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n" ) print(self.format_dict(_a ) ) return info @staticmethod def lowercase ( lowerCamelCase_ ) -> str: """simple docstring""" return "\n".join([f'''- {prop}: {val}''' for prop, val in d.items()] ) + "\n"
147
import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( '''kwargs, expected''' , [ ({'''num_shards''': 0, '''max_num_jobs''': 1}, []), ({'''num_shards''': 10, '''max_num_jobs''': 1}, [range(10 )]), ({'''num_shards''': 10, '''max_num_jobs''': 10}, [range(__lowerCAmelCase , i + 1 ) for i in range(10 )]), ({'''num_shards''': 1, '''max_num_jobs''': 10}, [range(1 )]), ({'''num_shards''': 10, '''max_num_jobs''': 3}, [range(0 , 4 ), range(4 , 7 ), range(7 , 10 )]), ({'''num_shards''': 3, '''max_num_jobs''': 10}, [range(0 , 1 ), range(1 , 2 ), range(2 , 3 )]), ] , ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> Optional[int]: snake_case__ = _distribute_shards(**__lowerCAmelCase ) assert out == expected @pytest.mark.parametrize( '''gen_kwargs, max_num_jobs, expected''' , [ ({'''foo''': 0}, 10, [{'''foo''': 0}]), ({'''shards''': [0, 1, 2, 3]}, 1, [{'''shards''': [0, 1, 2, 3]}]), ({'''shards''': [0, 1, 2, 3]}, 4, [{'''shards''': [0]}, {'''shards''': [1]}, {'''shards''': [2]}, {'''shards''': [3]}]), ({'''shards''': [0, 1]}, 4, [{'''shards''': [0]}, {'''shards''': [1]}]), ({'''shards''': [0, 1, 2, 3]}, 2, [{'''shards''': [0, 1]}, {'''shards''': [2, 3]}]), ] , ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> Dict: snake_case__ = _split_gen_kwargs(__lowerCAmelCase , __lowerCAmelCase ) assert out == expected @pytest.mark.parametrize( '''gen_kwargs, expected''' , [ ({'''foo''': 0}, 1), ({'''shards''': [0]}, 1), ({'''shards''': [0, 1, 2, 3]}, 4), ({'''shards''': [0, 1, 2, 3], '''foo''': 0}, 4), ({'''shards''': [0, 1, 2, 3], '''other''': (0, 1)}, 4), ({'''shards''': [0, 1, 2, 3], '''shards2''': [0, 1]}, RuntimeError), ] , ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> List[Any]: if expected is RuntimeError: with pytest.raises(__lowerCAmelCase ): _number_of_shards_in_gen_kwargs(__lowerCAmelCase ) else: snake_case__ = _number_of_shards_in_gen_kwargs(__lowerCAmelCase ) assert out == expected
33
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) SCREAMING_SNAKE_CASE_ = {'configuration_opt': ['OPT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'OPTConfig']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ = [ 'OPT_PRETRAINED_MODEL_ARCHIVE_LIST', 'OPTForCausalLM', 'OPTModel', 'OPTPreTrainedModel', 'OPTForSequenceClassification', 'OPTForQuestionAnswering', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ = ['TFOPTForCausalLM', 'TFOPTModel', 'TFOPTPreTrainedModel'] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ = [ 'FlaxOPTForCausalLM', 'FlaxOPTModel', 'FlaxOPTPreTrainedModel', ] if TYPE_CHECKING: from .configuration_opt import OPT_PRETRAINED_CONFIG_ARCHIVE_MAP, OPTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_opt import ( OPT_PRETRAINED_MODEL_ARCHIVE_LIST, OPTForCausalLM, OPTForQuestionAnswering, OPTForSequenceClassification, OPTModel, OPTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_opt import TFOPTForCausalLM, TFOPTModel, TFOPTPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_opt import FlaxOPTForCausalLM, FlaxOPTModel, FlaxOPTPreTrainedModel else: import sys SCREAMING_SNAKE_CASE_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
34
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class snake_case_ ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): """simple docstring""" A_ = StableDiffusionInpaintPipeline A_ = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS A_ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS A_ = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess A_ = frozenset([] ) def UpperCAmelCase__ ( self) -> List[Any]: torch.manual_seed(0) UpperCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=9 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=3_2 , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase_ , ) UpperCamelCase = PNDMScheduler(skip_prk_steps=lowerCamelCase_) torch.manual_seed(0) UpperCamelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0) UpperCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='''gelu''' , projection_dim=5_1_2 , ) UpperCamelCase = CLIPTextModel(lowerCamelCase_) UpperCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') UpperCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_=0) -> Dict: # TODO: use tensor inputs instead of PIL, this is here just to leave the old expected_slices untouched UpperCamelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCamelCase_)).to(lowerCamelCase_) UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1)[0] UpperCamelCase = Image.fromarray(np.uinta(lowerCamelCase_)).convert('''RGB''').resize((6_4, 6_4)) UpperCamelCase = Image.fromarray(np.uinta(image + 4)).convert('''RGB''').resize((6_4, 6_4)) if str(lowerCamelCase_).startswith('''mps'''): UpperCamelCase = torch.manual_seed(lowerCamelCase_) else: UpperCamelCase = torch.Generator(device=lowerCamelCase_).manual_seed(lowerCamelCase_) UpperCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': init_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def UpperCAmelCase__ ( self) -> Optional[Any]: UpperCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCamelCase = self.get_dummy_components() UpperCamelCase = StableDiffusionInpaintPipeline(**lowerCamelCase_) UpperCamelCase = sd_pipe.to(lowerCamelCase_) sd_pipe.set_progress_bar_config(disable=lowerCamelCase_) UpperCamelCase = self.get_dummy_inputs(lowerCamelCase_) UpperCamelCase = sd_pipe(**lowerCamelCase_).images UpperCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) UpperCamelCase = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 def UpperCAmelCase__ ( self) -> Union[str, Any]: super().test_inference_batch_single_identical(expected_max_diff=3e-3) @slow @require_torch_gpu class snake_case_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self) -> List[Any]: UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''') UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''') UpperCamelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench.npy''') UpperCamelCase = '''stabilityai/stable-diffusion-2-inpainting''' UpperCamelCase = StableDiffusionInpaintPipeline.from_pretrained(lowerCamelCase_ , safety_checker=lowerCamelCase_) pipe.to(lowerCamelCase_) pipe.set_progress_bar_config(disable=lowerCamelCase_) pipe.enable_attention_slicing() UpperCamelCase = '''Face of a yellow cat, high resolution, sitting on a park bench''' UpperCamelCase = torch.manual_seed(0) UpperCamelCase = pipe( prompt=lowerCamelCase_ , image=lowerCamelCase_ , mask_image=lowerCamelCase_ , generator=lowerCamelCase_ , output_type='''np''' , ) UpperCamelCase = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image).max() < 9e-3 def UpperCAmelCase__ ( self) -> Optional[Any]: UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''') UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''') UpperCamelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench_fp16.npy''') UpperCamelCase = '''stabilityai/stable-diffusion-2-inpainting''' UpperCamelCase = StableDiffusionInpaintPipeline.from_pretrained( lowerCamelCase_ , torch_dtype=torch.floataa , safety_checker=lowerCamelCase_ , ) pipe.to(lowerCamelCase_) pipe.set_progress_bar_config(disable=lowerCamelCase_) pipe.enable_attention_slicing() UpperCamelCase = '''Face of a yellow cat, high resolution, sitting on a park bench''' UpperCamelCase = torch.manual_seed(0) UpperCamelCase = pipe( prompt=lowerCamelCase_ , image=lowerCamelCase_ , mask_image=lowerCamelCase_ , generator=lowerCamelCase_ , output_type='''np''' , ) UpperCamelCase = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image).max() < 5e-1 def UpperCAmelCase__ ( self) -> List[str]: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''') UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''') UpperCamelCase = '''stabilityai/stable-diffusion-2-inpainting''' UpperCamelCase = PNDMScheduler.from_pretrained(lowerCamelCase_ , subfolder='''scheduler''') UpperCamelCase = StableDiffusionInpaintPipeline.from_pretrained( lowerCamelCase_ , safety_checker=lowerCamelCase_ , scheduler=lowerCamelCase_ , torch_dtype=torch.floataa , ) pipe.to(lowerCamelCase_) pipe.set_progress_bar_config(disable=lowerCamelCase_) pipe.enable_attention_slicing(1) pipe.enable_sequential_cpu_offload() UpperCamelCase = '''Face of a yellow cat, high resolution, sitting on a park bench''' UpperCamelCase = torch.manual_seed(0) UpperCamelCase = pipe( prompt=lowerCamelCase_ , image=lowerCamelCase_ , mask_image=lowerCamelCase_ , generator=lowerCamelCase_ , num_inference_steps=2 , output_type='''np''' , ) UpperCamelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 1_0**9
34
1
"""simple docstring""" import operator def __snake_case ( _lowercase ,_lowercase = False ,_lowercase = None ): """simple docstring""" UpperCamelCase = operator.lt if reverse else operator.gt UpperCamelCase = solution or [] if not arr: return solution UpperCamelCase = [arr.pop(0 )] for i, item in enumerate(_lowercase ): if _operator(_lowercase ,sublist[-1] ): sublist.append(_lowercase ) arr.pop(_lowercase ) # merging sublist into solution list if not solution: solution.extend(_lowercase ) else: while sublist: UpperCamelCase = sublist.pop(0 ) for i, xx in enumerate(_lowercase ): if not _operator(_lowercase ,_lowercase ): solution.insert(_lowercase ,_lowercase ) break else: solution.append(_lowercase ) strand_sort(_lowercase ,_lowercase ,_lowercase ) return solution if __name__ == "__main__": assert strand_sort([4, 3, 5, 1, 2]) == [1, 2, 3, 4, 5] assert strand_sort([4, 3, 5, 1, 2], reverse=True) == [5, 4, 3, 2, 1]
34
"""simple docstring""" import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('3.8'): import importlib_metadata else: import importlib.metadata as importlib_metadata def __snake_case ( _lowercase ,_lowercase=False ): """simple docstring""" try: UpperCamelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. UpperCamelCase = default else: # KEY is set, convert it to True or False. try: UpperCamelCase = strtobool(_lowercase ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f'If set, {key} must be yes or no.' ) return _value SCREAMING_SNAKE_CASE_ = parse_flag_from_env('RUN_SLOW', default=False) SCREAMING_SNAKE_CASE_ = parse_flag_from_env('RUN_REMOTE', default=False) SCREAMING_SNAKE_CASE_ = parse_flag_from_env('RUN_LOCAL', default=True) SCREAMING_SNAKE_CASE_ = parse_flag_from_env('RUN_PACKAGED', default=True) # Compression SCREAMING_SNAKE_CASE_ = pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='test requires lz4') SCREAMING_SNAKE_CASE_ = pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='test requires py7zr') SCREAMING_SNAKE_CASE_ = pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='test requires zstandard') # Audio SCREAMING_SNAKE_CASE_ = pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('soundfile') is None or version.parse(importlib_metadata.version('soundfile')) < version.parse('0.12.0'), reason='test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ', ) # Beam SCREAMING_SNAKE_CASE_ = pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('0.3.2'), reason='test requires apache-beam and a compatible dill version', ) # Dill-cloudpickle compatibility SCREAMING_SNAKE_CASE_ = pytest.mark.skipif( config.DILL_VERSION <= version.parse('0.3.2'), reason='test requires dill>0.3.2 for cloudpickle compatibility', ) # Windows SCREAMING_SNAKE_CASE_ = pytest.mark.skipif( sys.platform == 'win32', reason='test should not be run on Windows', ) def __snake_case ( _lowercase ): """simple docstring""" try: import faiss # noqa except ImportError: UpperCamelCase = unittest.skip('''test requires faiss''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" try: import regex # noqa except ImportError: UpperCamelCase = unittest.skip('''test requires regex''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" try: import elasticsearch # noqa except ImportError: UpperCamelCase = unittest.skip('''test requires elasticsearch''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" try: import sqlalchemy # noqa except ImportError: UpperCamelCase = unittest.skip('''test requires sqlalchemy''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" if not config.TORCH_AVAILABLE: UpperCamelCase = unittest.skip('''test requires PyTorch''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" if not config.TF_AVAILABLE: UpperCamelCase = unittest.skip('''test requires TensorFlow''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" if not config.JAX_AVAILABLE: UpperCamelCase = unittest.skip('''test requires JAX''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" if not config.PIL_AVAILABLE: UpperCamelCase = unittest.skip('''test requires Pillow''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" try: import transformers # noqa F401 except ImportError: return unittest.skip('''test requires transformers''' )(_lowercase ) else: return test_case def __snake_case ( _lowercase ): """simple docstring""" try: import tiktoken # noqa F401 except ImportError: return unittest.skip('''test requires tiktoken''' )(_lowercase ) else: return test_case def __snake_case ( _lowercase ): """simple docstring""" try: import spacy # noqa F401 except ImportError: return unittest.skip('''test requires spacy''' )(_lowercase ) else: return test_case def __snake_case ( _lowercase ): """simple docstring""" def _require_spacy_model(_lowercase ): try: import spacy # noqa F401 spacy.load(_lowercase ) except ImportError: return unittest.skip('''test requires spacy''' )(_lowercase ) except OSError: return unittest.skip('''test requires spacy model \'{}\''''.format(_lowercase ) )(_lowercase ) else: return test_case return _require_spacy_model def __snake_case ( _lowercase ): """simple docstring""" try: import pyspark # noqa F401 except ImportError: return unittest.skip('''test requires pyspark''' )(_lowercase ) else: return test_case def __snake_case ( _lowercase ): """simple docstring""" try: import joblibspark # noqa F401 except ImportError: return unittest.skip('''test requires joblibspark''' )(_lowercase ) else: return test_case def __snake_case ( _lowercase ): """simple docstring""" if not _run_slow_tests or _run_slow_tests == 0: UpperCamelCase = unittest.skip('''test is slow''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" if not _run_local_tests or _run_local_tests == 0: UpperCamelCase = unittest.skip('''test is local''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" if not _run_packaged_tests or _run_packaged_tests == 0: UpperCamelCase = unittest.skip('''test is packaged''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" if not _run_remote_tests or _run_remote_tests == 0: UpperCamelCase = unittest.skip('''test requires remote''' )(_lowercase ) return test_case def __snake_case ( *_lowercase ): """simple docstring""" def decorate(cls ): for name, fn in cls.__dict__.items(): if callable(_lowercase ) and name.startswith('''test''' ): for decorator in decorators: UpperCamelCase = decorator(_lowercase ) setattr(cls ,_lowercase ,_lowercase ) return cls return decorate class snake_case_ ( lowerCamelCase_ ): """simple docstring""" pass class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = 0 A_ = 1 A_ = 2 @contextmanager def __snake_case ( _lowercase=OfflineSimulationMode.CONNECTION_FAILS ,_lowercase=1e-16 ): """simple docstring""" UpperCamelCase = requests.Session().request def timeout_request(_lowercase ,_lowercase ,_lowercase ,**_lowercase ): # Change the url to an invalid url so that the connection hangs UpperCamelCase = '''https://10.255.255.1''' if kwargs.get('''timeout''' ) is None: raise RequestWouldHangIndefinitelyError( f'Tried a call to {url} in offline mode with no timeout set. Please set a timeout.' ) UpperCamelCase = timeout try: return online_request(_lowercase ,_lowercase ,**_lowercase ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier UpperCamelCase = url UpperCamelCase = e.args[0] UpperCamelCase = (max_retry_error.args[0].replace('''10.255.255.1''' ,f'OfflineMock[{url}]' ),) UpperCamelCase = (max_retry_error,) raise def raise_connection_error(_lowercase ,_lowercase ,**_lowercase ): raise requests.ConnectionError('''Offline mode is enabled.''' ,request=_lowercase ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch('''requests.Session.send''' ,_lowercase ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch('''requests.Session.request''' ,_lowercase ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch('''datasets.config.HF_DATASETS_OFFLINE''' ,_lowercase ): yield else: raise ValueError('''Please use a value from the OfflineSimulationMode enum.''' ) @contextmanager def __snake_case ( *_lowercase ,**_lowercase ): """simple docstring""" UpperCamelCase = str(Path().resolve() ) with tempfile.TemporaryDirectory(*_lowercase ,**_lowercase ) as tmp_dir: try: os.chdir(_lowercase ) yield finally: os.chdir(_lowercase ) @contextmanager def __snake_case ( ): """simple docstring""" import gc gc.collect() UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def __snake_case ( ): """simple docstring""" import gc gc.collect() UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" return deepcopy(_lowercase ).integers(0 ,100 ,10 ).tolist() == deepcopy(_lowercase ).integers(0 ,100 ,10 ).tolist() def __snake_case ( _lowercase ): """simple docstring""" import decorator from requests.exceptions import HTTPError def _wrapper(_lowercase ,*_lowercase ,**_lowercase ): try: return func(*_lowercase ,**_lowercase ) except HTTPError as err: if str(_lowercase ).startswith('''500''' ) or str(_lowercase ).startswith('''502''' ): pytest.xfail(str(_lowercase ) ) raise err return decorator.decorator(_wrapper ,_lowercase ) class snake_case_ : """simple docstring""" def __init__( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_) -> Dict: UpperCamelCase = returncode UpperCamelCase = stdout UpperCamelCase = stderr async def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" while True: UpperCamelCase = await stream.readline() if line: callback(_lowercase ) else: break async def __snake_case ( _lowercase ,_lowercase=None ,_lowercase=None ,_lowercase=None ,_lowercase=False ,_lowercase=False ): """simple docstring""" if echo: print('''\nRunning: ''' ,''' '''.join(_lowercase ) ) UpperCamelCase = await asyncio.create_subprocess_exec( cmd[0] ,*cmd[1:] ,stdin=_lowercase ,stdout=asyncio.subprocess.PIPE ,stderr=asyncio.subprocess.PIPE ,env=_lowercase ,) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) UpperCamelCase = [] UpperCamelCase = [] def tee(_lowercase ,_lowercase ,_lowercase ,_lowercase="" ): UpperCamelCase = line.decode('''utf-8''' ).rstrip() sink.append(_lowercase ) if not quiet: print(_lowercase ,_lowercase ,file=_lowercase ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout ,lambda _lowercase : tee(_lowercase ,_lowercase ,sys.stdout ,label='''stdout:''' ) ), _read_stream(p.stderr ,lambda _lowercase : tee(_lowercase ,_lowercase ,sys.stderr ,label='''stderr:''' ) ), ] ,timeout=_lowercase ,) return _RunOutput(await p.wait() ,_lowercase ,_lowercase ) def __snake_case ( _lowercase ,_lowercase=None ,_lowercase=None ,_lowercase=180 ,_lowercase=False ,_lowercase=True ): """simple docstring""" UpperCamelCase = asyncio.get_event_loop() UpperCamelCase = loop.run_until_complete( _stream_subprocess(_lowercase ,env=_lowercase ,stdin=_lowercase ,timeout=_lowercase ,quiet=_lowercase ,echo=_lowercase ) ) UpperCamelCase = ''' '''.join(_lowercase ) if result.returncode > 0: UpperCamelCase = '''\n'''.join(result.stderr ) raise RuntimeError( f'\'{cmd_str}\' failed with returncode {result.returncode}\n\n' f'The combined stderr from workers follows:\n{stderr}' ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(f'\'{cmd_str}\' produced no output.' ) return result def __snake_case ( ): """simple docstring""" UpperCamelCase = os.environ.get('''PYTEST_XDIST_WORKER''' ,'''gw0''' ) UpperCamelCase = re.sub(r'''^gw''' ,'''''' ,_lowercase ,0 ,re.M ) return int(_lowercase ) def __snake_case ( ): """simple docstring""" UpperCamelCase = 2_9500 UpperCamelCase = pytest_xdist_worker_id() return port + uniq_delta
34
1
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_squeezebert import SqueezeBertTokenizer SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} SCREAMING_SNAKE_CASE_ = { 'vocab_file': { 'squeezebert/squeezebert-uncased': ( 'https://huggingface.co/squeezebert/squeezebert-uncased/resolve/main/vocab.txt' ), 'squeezebert/squeezebert-mnli': 'https://huggingface.co/squeezebert/squeezebert-mnli/resolve/main/vocab.txt', 'squeezebert/squeezebert-mnli-headless': ( 'https://huggingface.co/squeezebert/squeezebert-mnli-headless/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'squeezebert/squeezebert-uncased': ( 'https://huggingface.co/squeezebert/squeezebert-uncased/resolve/main/tokenizer.json' ), 'squeezebert/squeezebert-mnli': ( 'https://huggingface.co/squeezebert/squeezebert-mnli/resolve/main/tokenizer.json' ), 'squeezebert/squeezebert-mnli-headless': ( 'https://huggingface.co/squeezebert/squeezebert-mnli-headless/resolve/main/tokenizer.json' ), }, } SCREAMING_SNAKE_CASE_ = { 'squeezebert/squeezebert-uncased': 512, 'squeezebert/squeezebert-mnli': 512, 'squeezebert/squeezebert-mnli-headless': 512, } SCREAMING_SNAKE_CASE_ = { 'squeezebert/squeezebert-uncased': {'do_lower_case': True}, 'squeezebert/squeezebert-mnli': {'do_lower_case': True}, 'squeezebert/squeezebert-mnli-headless': {'do_lower_case': True}, } class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = VOCAB_FILES_NAMES A_ = PRETRAINED_VOCAB_FILES_MAP A_ = PRETRAINED_INIT_CONFIGURATION A_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ = SqueezeBertTokenizer def __init__( self , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=True , lowerCamelCase_="[UNK]" , lowerCamelCase_="[SEP]" , lowerCamelCase_="[PAD]" , lowerCamelCase_="[CLS]" , lowerCamelCase_="[MASK]" , lowerCamelCase_=True , lowerCamelCase_=None , **lowerCamelCase_ , ) -> Union[str, Any]: super().__init__( lowerCamelCase_ , tokenizer_file=lowerCamelCase_ , do_lower_case=lowerCamelCase_ , unk_token=lowerCamelCase_ , sep_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , cls_token=lowerCamelCase_ , mask_token=lowerCamelCase_ , tokenize_chinese_chars=lowerCamelCase_ , strip_accents=lowerCamelCase_ , **lowerCamelCase_ , ) UpperCamelCase = json.loads(self.backend_tokenizer.normalizer.__getstate__()) if ( normalizer_state.get('''lowercase''' , lowerCamelCase_) != do_lower_case or normalizer_state.get('''strip_accents''' , lowerCamelCase_) != strip_accents or normalizer_state.get('''handle_chinese_chars''' , lowerCamelCase_) != tokenize_chinese_chars ): UpperCamelCase = getattr(lowerCamelCase_ , normalizer_state.pop('''type''')) UpperCamelCase = do_lower_case UpperCamelCase = strip_accents UpperCamelCase = tokenize_chinese_chars UpperCamelCase = normalizer_class(**lowerCamelCase_) UpperCamelCase = do_lower_case def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_=None) -> Dict: UpperCamelCase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> List[int]: UpperCamelCase = [self.sep_token_id] UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> Tuple[str]: UpperCamelCase = self._tokenizer.model.save(lowerCamelCase_ , name=lowerCamelCase_) return tuple(lowerCamelCase_)
34
"""simple docstring""" import operator def __snake_case ( _lowercase ,_lowercase = False ,_lowercase = None ): """simple docstring""" UpperCamelCase = operator.lt if reverse else operator.gt UpperCamelCase = solution or [] if not arr: return solution UpperCamelCase = [arr.pop(0 )] for i, item in enumerate(_lowercase ): if _operator(_lowercase ,sublist[-1] ): sublist.append(_lowercase ) arr.pop(_lowercase ) # merging sublist into solution list if not solution: solution.extend(_lowercase ) else: while sublist: UpperCamelCase = sublist.pop(0 ) for i, xx in enumerate(_lowercase ): if not _operator(_lowercase ,_lowercase ): solution.insert(_lowercase ,_lowercase ) break else: solution.append(_lowercase ) strand_sort(_lowercase ,_lowercase ,_lowercase ) return solution if __name__ == "__main__": assert strand_sort([4, 3, 5, 1, 2]) == [1, 2, 3, 4, 5] assert strand_sort([4, 3, 5, 1, 2], reverse=True) == [5, 4, 3, 2, 1]
34
1
"""simple docstring""" def __snake_case ( _lowercase ): """simple docstring""" UpperCamelCase = [0] * len(_lowercase ) for i in range(1 ,len(_lowercase ) ): # use last results for better performance - dynamic programming UpperCamelCase = prefix_result[i - 1] while j > 0 and input_string[i] != input_string[j]: UpperCamelCase = prefix_result[j - 1] if input_string[i] == input_string[j]: j += 1 UpperCamelCase = j return prefix_result def __snake_case ( _lowercase ): """simple docstring""" return max(prefix_function(_lowercase ) ) if __name__ == "__main__": import doctest doctest.testmod()
34
"""simple docstring""" from scipy.stats import pearsonr import datasets SCREAMING_SNAKE_CASE_ = '\nPearson correlation coefficient and p-value for testing non-correlation.\nThe Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases.\nThe p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets.\n' SCREAMING_SNAKE_CASE_ = '\nArgs:\n predictions (`list` of `int`): Predicted class labels, as returned by a model.\n references (`list` of `int`): Ground truth labels.\n return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`.\n\nReturns:\n pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation.\n p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities.\n\nExamples:\n\n Example 1-A simple example using only predictions and references.\n >>> pearsonr_metric = datasets.load_metric("pearsonr")\n >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5])\n >>> print(round(results[\'pearsonr\'], 2))\n -0.74\n\n Example 2-The same as Example 1, but that also returns the `p-value`.\n >>> pearsonr_metric = datasets.load_metric("pearsonr")\n >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True)\n >>> print(sorted(list(results.keys())))\n [\'p-value\', \'pearsonr\']\n >>> print(round(results[\'pearsonr\'], 2))\n -0.74\n >>> print(round(results[\'p-value\'], 2))\n 0.15\n' SCREAMING_SNAKE_CASE_ = '\n@article{2020SciPy-NMeth,\nauthor = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, Ilhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Antonio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\ntitle = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\njournal = {Nature Methods},\nyear = {2020},\nvolume = {17},\npages = {261--272},\nadsurl = {https://rdcu.be/b08Wh},\ndoi = {10.1038/s41592-019-0686-2},\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class snake_case_ ( datasets.Metric ): """simple docstring""" def UpperCAmelCase__ ( self) -> Tuple: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''float'''), '''references''': datasets.Value('''float'''), }) , reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html'''] , ) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=False) -> Any: if return_pvalue: UpperCamelCase = pearsonr(lowerCamelCase_ , lowerCamelCase_) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(lowerCamelCase_ , lowerCamelCase_)[0])}
34
1
"""simple docstring""" import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('''TEST_SAGEMAKER''' , '''False''' ) ) is not True , reason='''Skipping test because should only be run when releasing minor transformers version''' , ) @pytest.mark.usefixtures('''sm_env''' ) @parameterized_class( [ { '''framework''': '''pytorch''', '''script''': '''run_glue.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.p3.16xlarge''', '''results''': {'''train_runtime''': 650, '''eval_accuracy''': 0.7, '''eval_loss''': 0.6}, }, { '''framework''': '''pytorch''', '''script''': '''run_ddp.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.p3.16xlarge''', '''results''': {'''train_runtime''': 600, '''eval_accuracy''': 0.7, '''eval_loss''': 0.6}, }, { '''framework''': '''tensorflow''', '''script''': '''run_tf_dist.py''', '''model_name_or_path''': '''distilbert-base-cased''', '''instance_type''': '''ml.p3.16xlarge''', '''results''': {'''train_runtime''': 600, '''eval_accuracy''': 0.6, '''eval_loss''': 0.7}, }, ] ) class snake_case_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self) -> Optional[int]: if self.framework == "pytorch": subprocess.run( F'cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py'.split() , encoding='''utf-8''' , check=lowerCamelCase_ , ) assert hasattr(self , '''env''') def UpperCAmelCase__ ( self , lowerCamelCase_) -> List[str]: UpperCamelCase = F'{self.env.base_job_name}-{instance_count}-{"ddp" if "ddp" in self.script else "smd"}' # distributed data settings UpperCamelCase = {'''smdistributed''': {'''dataparallel''': {'''enabled''': True}}} if self.script != '''run_ddp.py''' else None # creates estimator return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=lowerCamelCase_ , instance_count=lowerCamelCase_ , instance_type=self.instance_type , debugger_hook_config=lowerCamelCase_ , hyperparameters={**self.env.distributed_hyperparameters, '''model_name_or_path''': self.model_name_or_path} , metric_definitions=self.env.metric_definitions , distribution=lowerCamelCase_ , py_version='''py36''' , ) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Optional[Any]: TrainingJobAnalytics(lowerCamelCase_).export_csv(F'{self.env.test_path}/{job_name}_metrics.csv') @parameterized.expand([(2,)]) def UpperCAmelCase__ ( self , lowerCamelCase_) -> List[str]: # create estimator UpperCamelCase = self.create_estimator(lowerCamelCase_) # run training estimator.fit() # result dataframe UpperCamelCase = TrainingJobAnalytics(estimator.latest_training_job.name).dataframe() # extract kpis UpperCamelCase = list(result_metrics_df[result_metrics_df.metric_name == '''eval_accuracy''']['''value''']) UpperCamelCase = list(result_metrics_df[result_metrics_df.metric_name == '''eval_loss''']['''value''']) # get train time from SageMaker job, this includes starting, preprocessing, stopping UpperCamelCase = ( Session().describe_training_job(estimator.latest_training_job.name).get('''TrainingTimeInSeconds''' , 9_9_9_9_9_9) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results['''eval_accuracy'''] for t in eval_accuracy) assert all(t <= self.results['''eval_loss'''] for t in eval_loss) # dump tests result into json file to share in PR with open(F'{estimator.latest_training_job.name}.json' , '''w''') as outfile: json.dump({'''train_time''': train_runtime, '''eval_accuracy''': eval_accuracy, '''eval_loss''': eval_loss} , lowerCamelCase_)
34
"""simple docstring""" import unittest from dataclasses import dataclass import pytest from accelerate.commands.config.config_args import SageMakerConfig from accelerate.utils import ComputeEnvironment from accelerate.utils.launch import _convert_nargs_to_dict @dataclass class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = ComputeEnvironment.AMAZON_SAGEMAKER A_ = True A_ = '''ml.p3.2xlarge''' A_ = '''accelerate_sagemaker_execution_role''' A_ = '''hf-sm''' A_ = '''us-east-1''' A_ = 1 A_ = '''accelerate-sagemaker-1''' A_ = '''1.6''' A_ = '''4.4''' A_ = '''train.py''' A_ = [ '''--model_name_or_path''', '''bert''', '''--do_train''', '''False''', '''--epochs''', '''3''', '''--learning_rate''', '''5e-5''', '''--max_steps''', '''50.5''', ] A_ = [ '''--model_name_or_path''', '''bert''', '''--do_train''', '''--do_test''', '''False''', '''--do_predict''', '''--epochs''', '''3''', '''--learning_rate''', '''5e-5''', '''--max_steps''', '''50.5''', ] class snake_case_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self) -> List[Any]: # If no defaults are changed, `to_kwargs` returns an empty dict. UpperCamelCase = _convert_nargs_to_dict(MockLaunchConfig.success_training_script_args) assert isinstance(converted_args['''model_name_or_path'''] , lowerCamelCase_) assert isinstance(converted_args['''do_train'''] , lowerCamelCase_) assert isinstance(converted_args['''epochs'''] , lowerCamelCase_) assert isinstance(converted_args['''learning_rate'''] , lowerCamelCase_) assert isinstance(converted_args['''max_steps'''] , lowerCamelCase_) with pytest.raises(lowerCamelCase_): _convert_nargs_to_dict(MockLaunchConfig.fail_training_script_args)
34
1
"""simple docstring""" from typing import Dict, List, Optional from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = { 'nielsr/canine-s': 2048, } # Unicode defines 1,114,112 total “codepoints” SCREAMING_SNAKE_CASE_ = 1114112 # Below: Constants defining canonical codepoints for special, pseudo-characters. # Copied from https://github.com/google-research/language/blob/master/language/canine/special_codepoints.py SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = 0Xe_000 SCREAMING_SNAKE_CASE_ = 0Xe_001 SCREAMING_SNAKE_CASE_ = 0Xe_002 SCREAMING_SNAKE_CASE_ = 0Xe_003 SCREAMING_SNAKE_CASE_ = 0Xe_004 # Maps special codepoints to human-readable names. SCREAMING_SNAKE_CASE_ = { # Special symbols are represented using codepoints values that are valid, # but designated as "Private Use", meaning that they will never be assigned # characters by the Unicode Consortium, and are thus safe for use here. # # NOTE: Do *NOT* add any sort of [UNK_CHAR] here. They are explicitly # excluded and should fail with a hard error. CLS: "[CLS]", SEP: "[SEP]", BOS: "[BOS]", MASK: "[MASK]", PAD: "[PAD]", RESERVED: "[RESERVED]", } # Maps special codepoint human-readable names to their codepoint values. SCREAMING_SNAKE_CASE_ = {name: codepoint for codepoint, name in SPECIAL_CODEPOINTS.items()} class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , lowerCamelCase_=chr(lowerCamelCase_) , lowerCamelCase_=chr(lowerCamelCase_) , lowerCamelCase_=chr(lowerCamelCase_) , lowerCamelCase_=chr(lowerCamelCase_) , lowerCamelCase_=chr(lowerCamelCase_) , lowerCamelCase_=chr(lowerCamelCase_) , lowerCamelCase_=False , lowerCamelCase_=2_0_4_8 , **lowerCamelCase_ , ) -> List[Any]: UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_) if isinstance(lowerCamelCase_ , lowerCamelCase_) else bos_token UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_) if isinstance(lowerCamelCase_ , lowerCamelCase_) else eos_token UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_) if isinstance(lowerCamelCase_ , lowerCamelCase_) else sep_token UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_) if isinstance(lowerCamelCase_ , lowerCamelCase_) else cls_token UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_) if isinstance(lowerCamelCase_ , lowerCamelCase_) else pad_token # Mask token behave like a normal word, i.e. include the space before it UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_) if isinstance(lowerCamelCase_ , lowerCamelCase_) else mask_token super().__init__( bos_token=lowerCamelCase_ , eos_token=lowerCamelCase_ , sep_token=lowerCamelCase_ , cls_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , mask_token=lowerCamelCase_ , add_prefix_space=lowerCamelCase_ , model_max_length=lowerCamelCase_ , **lowerCamelCase_ , ) # Creates a mapping for looking up the IDs of special symbols. UpperCamelCase = {} for codepoint, name in SPECIAL_CODEPOINTS.items(): UpperCamelCase = codepoint # Creates a mapping for looking up the string forms of special symbol IDs. UpperCamelCase = { codepoint: name for name, codepoint in self._special_codepoints.items() } UpperCamelCase = UNICODE_VOCAB_SIZE UpperCamelCase = len(self._special_codepoints) @property def UpperCAmelCase__ ( self) -> int: return self._unicode_vocab_size def UpperCAmelCase__ ( self , lowerCamelCase_) -> List[str]: return list(lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> int: try: return ord(lowerCamelCase_) except TypeError: raise ValueError(F'invalid token: \'{token}\'') def UpperCAmelCase__ ( self , lowerCamelCase_) -> str: try: if index in SPECIAL_CODEPOINTS: return SPECIAL_CODEPOINTS[index] return chr(lowerCamelCase_) except TypeError: raise ValueError(F'invalid id: {index}') def UpperCAmelCase__ ( self , lowerCamelCase_) -> Optional[int]: return "".join(lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> List[int]: UpperCamelCase = [self.sep_token_id] UpperCamelCase = [self.cls_token_id] UpperCamelCase = cls + token_ids_a + sep if token_ids_a is not None: result += token_ids_a + sep return result def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None , lowerCamelCase_ = False) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ , token_ids_a=lowerCamelCase_ , already_has_special_tokens=lowerCamelCase_) UpperCamelCase = [1] + ([0] * len(lowerCamelCase_)) + [1] if token_ids_a is not None: result += ([0] * len(lowerCamelCase_)) + [1] return result def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> List[int]: UpperCamelCase = [self.sep_token_id] UpperCamelCase = [self.cls_token_id] UpperCamelCase = len(cls + token_ids_a + sep) * [0] if token_ids_a is not None: result += len(token_ids_a + sep) * [1] return result def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> List[str]: return ()
34
"""simple docstring""" from typing import List import jiwer import jiwer.transforms as tr from packaging import version import datasets from datasets.config import PY_VERSION if PY_VERSION < version.parse('3.8'): import importlib_metadata else: import importlib.metadata as importlib_metadata SCREAMING_SNAKE_CASE_ = '' if version.parse(importlib_metadata.version('jiwer')) < version.parse('2.3.0'): class snake_case_ ( tr.AbstractTransform ): """simple docstring""" def __init__( self , lowerCamelCase_ = " ") -> List[str]: UpperCamelCase = sentence_delimiter def UpperCAmelCase__ ( self , lowerCamelCase_) -> Tuple: return list(lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Optional[Any]: UpperCamelCase = [] for sent_idx, sentence in enumerate(lowerCamelCase_): chars.extend(self.process_string(lowerCamelCase_)) if self.sentence_delimiter is not None and self.sentence_delimiter != "" and sent_idx < len(lowerCamelCase_) - 1: chars.append(self.sentence_delimiter) return chars SCREAMING_SNAKE_CASE_ = tr.Compose( [tr.RemoveMultipleSpaces(), tr.Strip(), SentencesToListOfCharacters(SENTENCE_DELIMITER)] ) else: SCREAMING_SNAKE_CASE_ = tr.Compose( [ tr.RemoveMultipleSpaces(), tr.Strip(), tr.ReduceToSingleSentence(SENTENCE_DELIMITER), tr.ReduceToListOfListOfChars(), ] ) SCREAMING_SNAKE_CASE_ = '\\n@inproceedings{inproceedings,\n author = {Morris, Andrew and Maier, Viktoria and Green, Phil},\n year = {2004},\n month = {01},\n pages = {},\n title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.}\n}\n' SCREAMING_SNAKE_CASE_ = '\\nCharacter error rate (CER) is a common metric of the performance of an automatic speech recognition system.\n\nCER is similar to Word Error Rate (WER), but operates on character instead of word. Please refer to docs of WER for further information.\n\nCharacter error rate can be computed as:\n\nCER = (S + D + I) / N = (S + D + I) / (S + D + C)\n\nwhere\n\nS is the number of substitutions,\nD is the number of deletions,\nI is the number of insertions,\nC is the number of correct characters,\nN is the number of characters in the reference (N=S+D+C).\n\nCER\'s output is not always a number between 0 and 1, in particular when there is a high number of insertions. This value is often associated to the percentage of characters that were incorrectly predicted. The lower the value, the better the\nperformance of the ASR system with a CER of 0 being a perfect score.\n' SCREAMING_SNAKE_CASE_ = '\nComputes CER score of transcribed segments against references.\nArgs:\n references: list of references for each speech input.\n predictions: list of transcribtions to score.\n concatenate_texts: Whether or not to concatenate sentences before evaluation, set to True for more accurate result.\nReturns:\n (float): the character error rate\n\nExamples:\n\n >>> predictions = ["this is the prediction", "there is an other sample"]\n >>> references = ["this is the reference", "there is another one"]\n >>> cer = datasets.load_metric("cer")\n >>> cer_score = cer.compute(predictions=predictions, references=references)\n >>> print(cer_score)\n 0.34146341463414637\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class snake_case_ ( datasets.Metric ): """simple docstring""" def UpperCAmelCase__ ( self) -> Dict: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence'''), '''references''': datasets.Value('''string''' , id='''sequence'''), }) , codebase_urls=['''https://github.com/jitsi/jiwer/'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/Word_error_rate''', '''https://sites.google.com/site/textdigitisation/qualitymeasures/computingerrorrates''', ] , ) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=False) -> List[Any]: if concatenate_texts: return jiwer.compute_measures( lowerCamelCase_ , lowerCamelCase_ , truth_transform=lowerCamelCase_ , hypothesis_transform=lowerCamelCase_ , )["wer"] UpperCamelCase = 0 UpperCamelCase = 0 for prediction, reference in zip(lowerCamelCase_ , lowerCamelCase_): UpperCamelCase = jiwer.compute_measures( lowerCamelCase_ , lowerCamelCase_ , truth_transform=lowerCamelCase_ , hypothesis_transform=lowerCamelCase_ , ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
34
1
"""simple docstring""" import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class snake_case_ ( unittest.TestCase ): """simple docstring""" def __init__( self , lowerCamelCase_ , lowerCamelCase_=1_3 , lowerCamelCase_=7 , lowerCamelCase_=True , lowerCamelCase_=True , lowerCamelCase_=True , lowerCamelCase_=True , lowerCamelCase_=9_9 , lowerCamelCase_=3_2 , lowerCamelCase_=5 , lowerCamelCase_=4 , lowerCamelCase_=3_7 , lowerCamelCase_="gelu" , lowerCamelCase_=0.1 , lowerCamelCase_=0.1 , lowerCamelCase_=5_1_2 , lowerCamelCase_=1_6 , lowerCamelCase_=2 , lowerCamelCase_=0.02 , lowerCamelCase_=4 , ) -> Any: UpperCamelCase = parent UpperCamelCase = batch_size UpperCamelCase = seq_length UpperCamelCase = is_training UpperCamelCase = use_attention_mask UpperCamelCase = use_token_type_ids UpperCamelCase = use_labels UpperCamelCase = vocab_size UpperCamelCase = hidden_size UpperCamelCase = num_hidden_layers UpperCamelCase = num_attention_heads UpperCamelCase = intermediate_size UpperCamelCase = hidden_act UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = max_position_embeddings UpperCamelCase = type_vocab_size UpperCamelCase = type_sequence_label_size UpperCamelCase = initializer_range UpperCamelCase = num_choices def UpperCAmelCase__ ( self) -> Optional[Any]: UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) UpperCamelCase = None if self.use_attention_mask: UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length]) UpperCamelCase = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , tie_weights_=lowerCamelCase_ , ) return config, input_ids, attention_mask def UpperCAmelCase__ ( self) -> str: UpperCamelCase = self.prepare_config_and_inputs() UpperCamelCase , UpperCamelCase , UpperCamelCase = config_and_inputs UpperCamelCase = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_flax class snake_case_ ( lowerCamelCase_ , unittest.TestCase ): """simple docstring""" A_ = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCAmelCase__ ( self) -> List[str]: UpperCamelCase = FlaxDistilBertModelTester(self) @slow def UpperCAmelCase__ ( self) -> Dict: for model_class_name in self.all_model_classes: UpperCamelCase = model_class_name.from_pretrained('''distilbert-base-uncased''') UpperCamelCase = model(np.ones((1, 1))) self.assertIsNotNone(lowerCamelCase_) @require_flax class snake_case_ ( unittest.TestCase ): """simple docstring""" @slow def UpperCAmelCase__ ( self) -> Optional[int]: UpperCamelCase = FlaxDistilBertModel.from_pretrained('''distilbert-base-uncased''') UpperCamelCase = np.array([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]]) UpperCamelCase = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]) UpperCamelCase = model(lowerCamelCase_ , attention_mask=lowerCamelCase_)[0] UpperCamelCase = (1, 1_1, 7_6_8) self.assertEqual(output.shape , lowerCamelCase_) UpperCamelCase = np.array([[[-0.1639, 0.3299, 0.1648], [-0.1746, 0.3289, 0.1710], [-0.1884, 0.3357, 0.1810]]]) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , lowerCamelCase_ , atol=1e-4))
34
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = {'vocab_file': 'spiece.model'} SCREAMING_SNAKE_CASE_ = { 'vocab_file': { 'xlnet-base-cased': 'https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model', 'xlnet-large-cased': 'https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model', } } SCREAMING_SNAKE_CASE_ = { 'xlnet-base-cased': None, 'xlnet-large-cased': None, } # Segments (not really needed) SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = 1 SCREAMING_SNAKE_CASE_ = 2 SCREAMING_SNAKE_CASE_ = 3 SCREAMING_SNAKE_CASE_ = 4 class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = VOCAB_FILES_NAMES A_ = PRETRAINED_VOCAB_FILES_MAP A_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ = '''left''' def __init__( self , lowerCamelCase_ , lowerCamelCase_=False , lowerCamelCase_=True , lowerCamelCase_=False , lowerCamelCase_="<s>" , lowerCamelCase_="</s>" , lowerCamelCase_="<unk>" , lowerCamelCase_="<sep>" , lowerCamelCase_="<pad>" , lowerCamelCase_="<cls>" , lowerCamelCase_="<mask>" , lowerCamelCase_=["<eop>", "<eod>"] , lowerCamelCase_ = None , **lowerCamelCase_ , ) -> None: # Mask token behave like a normal word, i.e. include the space before it UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_) if isinstance(lowerCamelCase_ , lowerCamelCase_) else mask_token UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=lowerCamelCase_ , remove_space=lowerCamelCase_ , keep_accents=lowerCamelCase_ , bos_token=lowerCamelCase_ , eos_token=lowerCamelCase_ , unk_token=lowerCamelCase_ , sep_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , cls_token=lowerCamelCase_ , mask_token=lowerCamelCase_ , additional_special_tokens=lowerCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase_ , ) UpperCamelCase = 3 UpperCamelCase = do_lower_case UpperCamelCase = remove_space UpperCamelCase = keep_accents UpperCamelCase = vocab_file UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(lowerCamelCase_) @property def UpperCAmelCase__ ( self) -> List[str]: return len(self.sp_model) def UpperCAmelCase__ ( self) -> Tuple: UpperCamelCase = {self.convert_ids_to_tokens(lowerCamelCase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__( self) -> Any: UpperCamelCase = self.__dict__.copy() UpperCamelCase = None return state def __setstate__( self , lowerCamelCase_) -> str: UpperCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs'''): UpperCamelCase = {} UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Union[str, Any]: if self.remove_space: UpperCamelCase = ''' '''.join(inputs.strip().split()) else: UpperCamelCase = inputs UpperCamelCase = outputs.replace('''``''' , '''"''').replace('''\'\'''' , '''"''') if not self.keep_accents: UpperCamelCase = unicodedata.normalize('''NFKD''' , lowerCamelCase_) UpperCamelCase = ''''''.join([c for c in outputs if not unicodedata.combining(lowerCamelCase_)]) if self.do_lower_case: UpperCamelCase = outputs.lower() return outputs def UpperCAmelCase__ ( self , lowerCamelCase_) -> List[str]: UpperCamelCase = self.preprocess_text(lowerCamelCase_) UpperCamelCase = self.sp_model.encode(lowerCamelCase_ , out_type=lowerCamelCase_) UpperCamelCase = [] for piece in pieces: if len(lowerCamelCase_) > 1 and piece[-1] == str(''',''') and piece[-2].isdigit(): UpperCamelCase = self.sp_model.EncodeAsPieces(piece[:-1].replace(lowerCamelCase_ , '''''')) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0]) == 1: UpperCamelCase = cur_pieces[1:] else: UpperCamelCase = cur_pieces[0][1:] cur_pieces.append(piece[-1]) new_pieces.extend(lowerCamelCase_) else: new_pieces.append(lowerCamelCase_) return new_pieces def UpperCAmelCase__ ( self , lowerCamelCase_) -> int: return self.sp_model.PieceToId(lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Optional[int]: return self.sp_model.IdToPiece(lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Dict: UpperCamelCase = ''''''.join(lowerCamelCase_).replace(lowerCamelCase_ , ''' ''').strip() return out_string def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = False , lowerCamelCase_ = None , lowerCamelCase_ = True , **lowerCamelCase_ , ) -> str: UpperCamelCase = kwargs.pop('''use_source_tokenizer''' , lowerCamelCase_) UpperCamelCase = self.convert_ids_to_tokens(lowerCamelCase_ , skip_special_tokens=lowerCamelCase_) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 UpperCamelCase = [] UpperCamelCase = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(lowerCamelCase_)) UpperCamelCase = [] sub_texts.append(lowerCamelCase_) else: current_sub_text.append(lowerCamelCase_) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(lowerCamelCase_)) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens UpperCamelCase = ''''''.join(lowerCamelCase_) UpperCamelCase = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: UpperCamelCase = self.clean_up_tokenization(lowerCamelCase_) return clean_text else: return text def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> List[int]: UpperCamelCase = [self.sep_token_id] UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None , lowerCamelCase_ = False) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ , token_ids_a=lowerCamelCase_ , already_has_special_tokens=lowerCamelCase_) if token_ids_a is not None: return ([0] * len(lowerCamelCase_)) + [1] + ([0] * len(lowerCamelCase_)) + [1, 1] return ([0] * len(lowerCamelCase_)) + [1, 1] def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> List[int]: UpperCamelCase = [self.sep_token_id] UpperCamelCase = [2] if token_ids_a is None: return len(token_ids_a + sep) * [0] + cls_segment_id return len(token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] + cls_segment_id def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> Tuple[str]: if not os.path.isdir(lowerCamelCase_): logger.error(F'Vocabulary path ({save_directory}) should be a directory') return UpperCamelCase = os.path.join( lowerCamelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowerCamelCase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , lowerCamelCase_) elif not os.path.isfile(self.vocab_file): with open(lowerCamelCase_ , '''wb''') as fi: UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase_) return (out_vocab_file,)
34
1
"""simple docstring""" from __future__ import annotations def __snake_case ( _lowercase ): """simple docstring""" return len(set(_lowercase ) ) == len(_lowercase ) if __name__ == "__main__": import doctest doctest.testmod()
34
"""simple docstring""" import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = {'vocab_file': 'vocab.txt'} SCREAMING_SNAKE_CASE_ = { 'vocab_file': { 'openbmb/cpm-ant-10b': 'https://huggingface.co/openbmb/cpm-ant-10b/blob/main/vocab.txt', }, } SCREAMING_SNAKE_CASE_ = { 'openbmb/cpm-ant-10b': 1024, } def __snake_case ( _lowercase ): """simple docstring""" UpperCamelCase = collections.OrderedDict() with open(_lowercase ,'''r''' ,encoding='''utf-8''' ) as reader: UpperCamelCase = reader.readlines() for index, token in enumerate(_lowercase ): UpperCamelCase = token.rstrip('''\n''' ) UpperCamelCase = index return vocab class snake_case_ ( lowerCamelCase_ ): """simple docstring""" def __init__( self , lowerCamelCase_ , lowerCamelCase_="<unk>" , lowerCamelCase_=2_0_0) -> Any: UpperCamelCase = vocab UpperCamelCase = unk_token UpperCamelCase = max_input_chars_per_word def UpperCAmelCase__ ( self , lowerCamelCase_) -> Union[str, Any]: UpperCamelCase = list(lowerCamelCase_) if len(lowerCamelCase_) > self.max_input_chars_per_word: return [self.unk_token] UpperCamelCase = 0 UpperCamelCase = [] while start < len(lowerCamelCase_): UpperCamelCase = len(lowerCamelCase_) UpperCamelCase = None while start < end: UpperCamelCase = ''''''.join(chars[start:end]) if substr in self.vocab: UpperCamelCase = substr break end -= 1 if cur_substr is None: sub_tokens.append(self.unk_token) start += 1 else: sub_tokens.append(lowerCamelCase_) UpperCamelCase = end return sub_tokens class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = VOCAB_FILES_NAMES A_ = PRETRAINED_VOCAB_FILES_MAP A_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ = ['''input_ids''', '''attention_mask'''] A_ = False def __init__( self , lowerCamelCase_ , lowerCamelCase_="<d>" , lowerCamelCase_="</d>" , lowerCamelCase_="<s>" , lowerCamelCase_="</s>" , lowerCamelCase_="<pad>" , lowerCamelCase_="<unk>" , lowerCamelCase_="</n>" , lowerCamelCase_="</_>" , lowerCamelCase_="left" , **lowerCamelCase_ , ) -> List[str]: requires_backends(self , ['''jieba''']) super().__init__( bod_token=lowerCamelCase_ , eod_token=lowerCamelCase_ , bos_token=lowerCamelCase_ , eos_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , unk_token=lowerCamelCase_ , line_token=lowerCamelCase_ , space_token=lowerCamelCase_ , padding_side=lowerCamelCase_ , **lowerCamelCase_ , ) UpperCamelCase = bod_token UpperCamelCase = eod_token UpperCamelCase = load_vocab(lowerCamelCase_) UpperCamelCase = self.encoder[space_token] UpperCamelCase = self.encoder[line_token] del self.encoder[space_token] del self.encoder[line_token] UpperCamelCase = collections.OrderedDict(sorted(self.encoder.items() , key=lambda lowerCamelCase_: x[1])) UpperCamelCase = {v: k for k, v in self.encoder.items()} UpperCamelCase = WordpieceTokenizer(vocab=self.encoder , unk_token=self.unk_token) @property def UpperCAmelCase__ ( self) -> Dict: return self.encoder[self.bod_token] @property def UpperCAmelCase__ ( self) -> str: return self.encoder[self.eod_token] @property def UpperCAmelCase__ ( self) -> List[Any]: return self.encoder["\n"] @property def UpperCAmelCase__ ( self) -> int: return len(self.encoder) def UpperCAmelCase__ ( self) -> Dict: return dict(self.encoder , **self.added_tokens_encoder) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Any: UpperCamelCase = [] for x in jieba.cut(lowerCamelCase_ , cut_all=lowerCamelCase_): output_tokens.extend(self.wordpiece_tokenizer.tokenize(lowerCamelCase_)) return output_tokens def UpperCAmelCase__ ( self , lowerCamelCase_ , **lowerCamelCase_) -> Tuple: UpperCamelCase = [i for i in token_ids if i >= 0] UpperCamelCase = [ x for x in token_ids if x != self.pad_token_id and x != self.eos_token_id and x != self.bos_token_id ] return super()._decode(lowerCamelCase_ , **lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Dict: return token in self.encoder def UpperCAmelCase__ ( self , lowerCamelCase_) -> str: return "".join(lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Optional[int]: return self.encoder.get(lowerCamelCase_ , self.encoder.get(self.unk_token)) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Dict: return self.decoder.get(lowerCamelCase_ , self.unk_token) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> Tuple[str]: if os.path.isdir(lowerCamelCase_): UpperCamelCase = os.path.join( lowerCamelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) else: UpperCamelCase = (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory UpperCamelCase = 0 if " " in self.encoder: UpperCamelCase = self.encoder[''' '''] del self.encoder[" "] if "\n" in self.encoder: UpperCamelCase = self.encoder['''\n'''] del self.encoder["\n"] UpperCamelCase = collections.OrderedDict(sorted(self.encoder.items() , key=lambda lowerCamelCase_: x[1])) with open(lowerCamelCase_ , '''w''' , encoding='''utf-8''') as writer: for token, token_index in self.encoder.items(): if index != token_index: logger.warning( F'Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.' ''' Please check that the vocabulary is not corrupted!''') UpperCamelCase = token_index writer.write(token + '''\n''') index += 1 return (vocab_file,) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> List[int]: if token_ids_a is None: return [self.bos_token_id] + token_ids_a return [self.bos_token_id] + token_ids_a + [self.bos_token_id] + token_ids_a def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None , lowerCamelCase_ = False) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ , token_ids_a=lowerCamelCase_ , already_has_special_tokens=lowerCamelCase_) if token_ids_a is not None: return [1] + ([0] * len(lowerCamelCase_)) + [1] + ([0] * len(lowerCamelCase_)) return [1] + ([0] * len(lowerCamelCase_))
34
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available SCREAMING_SNAKE_CASE_ = { 'configuration_instructblip': [ 'INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'InstructBlipConfig', 'InstructBlipQFormerConfig', 'InstructBlipVisionConfig', ], 'processing_instructblip': ['InstructBlipProcessor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ = [ 'INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST', 'InstructBlipQFormerModel', 'InstructBlipPreTrainedModel', 'InstructBlipForConditionalGeneration', 'InstructBlipVisionModel', ] if TYPE_CHECKING: from .configuration_instructblip import ( INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, InstructBlipConfig, InstructBlipQFormerConfig, InstructBlipVisionConfig, ) from .processing_instructblip import InstructBlipProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_instructblip import ( INSTRUCTBLIP_PRETRAINED_MODEL_ARCHIVE_LIST, InstructBlipForConditionalGeneration, InstructBlipPreTrainedModel, InstructBlipQFormerModel, InstructBlipVisionModel, ) else: import sys SCREAMING_SNAKE_CASE_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
34
"""simple docstring""" from typing import Callable, Dict, Optional, Tuple import torch from torch import nn from torch.distributions import ( AffineTransform, Distribution, Independent, NegativeBinomial, Normal, StudentT, TransformedDistribution, ) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" def __init__( self , lowerCamelCase_ , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=0) -> int: UpperCamelCase = 1.0 if scale is None else scale UpperCamelCase = 0.0 if loc is None else loc super().__init__(lowerCamelCase_ , [AffineTransform(loc=self.loc , scale=self.scale , event_dim=lowerCamelCase_)]) @property def UpperCAmelCase__ ( self) -> List[Any]: return self.base_dist.mean * self.scale + self.loc @property def UpperCAmelCase__ ( self) -> List[str]: return self.base_dist.variance * self.scale**2 @property def UpperCAmelCase__ ( self) -> Any: return self.variance.sqrt() class snake_case_ ( nn.Module ): """simple docstring""" def __init__( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , **lowerCamelCase_) -> None: super().__init__(**lowerCamelCase_) UpperCamelCase = args_dim UpperCamelCase = nn.ModuleList([nn.Linear(lowerCamelCase_ , lowerCamelCase_) for dim in args_dim.values()]) UpperCamelCase = domain_map def UpperCAmelCase__ ( self , lowerCamelCase_) -> Tuple[torch.Tensor]: UpperCamelCase = [proj(lowerCamelCase_) for proj in self.proj] return self.domain_map(*lowerCamelCase_) class snake_case_ ( nn.Module ): """simple docstring""" def __init__( self , lowerCamelCase_) -> int: super().__init__() UpperCamelCase = function def UpperCAmelCase__ ( self , lowerCamelCase_ , *lowerCamelCase_) -> Tuple: return self.function(lowerCamelCase_ , *lowerCamelCase_) class snake_case_ : """simple docstring""" A_ = 42 A_ = 42 A_ = 42 def __init__( self , lowerCamelCase_ = 1) -> None: UpperCamelCase = dim UpperCamelCase = {k: dim * self.args_dim[k] for k in self.args_dim} def UpperCAmelCase__ ( self , lowerCamelCase_) -> str: if self.dim == 1: return self.distribution_class(*lowerCamelCase_) else: return Independent(self.distribution_class(*lowerCamelCase_) , 1) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None , lowerCamelCase_ = None , ) -> Distribution: UpperCamelCase = self._base_distribution(lowerCamelCase_) if loc is None and scale is None: return distr else: return AffineTransformed(lowerCamelCase_ , loc=lowerCamelCase_ , scale=lowerCamelCase_ , event_dim=self.event_dim) @property def UpperCAmelCase__ ( self) -> Tuple: return () if self.dim == 1 else (self.dim,) @property def UpperCAmelCase__ ( self) -> int: return len(self.event_shape) @property def UpperCAmelCase__ ( self) -> float: return 0.0 def UpperCAmelCase__ ( self , lowerCamelCase_) -> nn.Module: return ParameterProjection( in_features=lowerCamelCase_ , args_dim=self.args_dim , domain_map=LambdaLayer(self.domain_map) , ) def UpperCAmelCase__ ( self , *lowerCamelCase_) -> List[str]: raise NotImplementedError() @staticmethod def UpperCAmelCase__ ( lowerCamelCase_) -> torch.Tensor: return (x + torch.sqrt(torch.square(lowerCamelCase_) + 4.0)) / 2.0 class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = {"df": 1, "loc": 1, "scale": 1} A_ = StudentT @classmethod def UpperCAmelCase__ ( cls , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_) -> Optional[int]: UpperCamelCase = cls.squareplus(lowerCamelCase_).clamp_min(torch.finfo(scale.dtype).eps) UpperCamelCase = 2.0 + cls.squareplus(lowerCamelCase_) return df.squeeze(-1), loc.squeeze(-1), scale.squeeze(-1) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = {"loc": 1, "scale": 1} A_ = Normal @classmethod def UpperCAmelCase__ ( cls , lowerCamelCase_ , lowerCamelCase_) -> str: UpperCamelCase = cls.squareplus(lowerCamelCase_).clamp_min(torch.finfo(scale.dtype).eps) return loc.squeeze(-1), scale.squeeze(-1) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = {"total_count": 1, "logits": 1} A_ = NegativeBinomial @classmethod def UpperCAmelCase__ ( cls , lowerCamelCase_ , lowerCamelCase_) -> List[Any]: UpperCamelCase = cls.squareplus(lowerCamelCase_) return total_count.squeeze(-1), logits.squeeze(-1) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Distribution: UpperCamelCase , UpperCamelCase = distr_args if self.dim == 1: return self.distribution_class(total_count=lowerCamelCase_ , logits=lowerCamelCase_) else: return Independent(self.distribution_class(total_count=lowerCamelCase_ , logits=lowerCamelCase_) , 1) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None , lowerCamelCase_ = None) -> Distribution: UpperCamelCase , UpperCamelCase = distr_args if scale is not None: # See scaling property of Gamma. logits += scale.log() return self._base_distribution((total_count, logits))
34
1
"""simple docstring""" from manim import * class snake_case_ ( lowerCamelCase_ ): """simple docstring""" def UpperCAmelCase__ ( self) -> Union[str, Any]: UpperCamelCase = Rectangle(height=0.5 , width=0.5) UpperCamelCase = Rectangle(height=0.46 , width=0.46).set_stroke(width=0) UpperCamelCase = [mem.copy() for i in range(6)] UpperCamelCase = [mem.copy() for i in range(6)] UpperCamelCase = VGroup(*lowerCamelCase_).arrange(lowerCamelCase_ , buff=0) UpperCamelCase = VGroup(*lowerCamelCase_).arrange(lowerCamelCase_ , buff=0) UpperCamelCase = VGroup(lowerCamelCase_ , lowerCamelCase_).arrange(lowerCamelCase_ , buff=0) UpperCamelCase = Text('''CPU''' , font_size=2_4) UpperCamelCase = Group(lowerCamelCase_ , lowerCamelCase_).arrange(lowerCamelCase_ , buff=0.5 , aligned_edge=lowerCamelCase_) cpu.move_to([-2.5, -0.5, 0]) self.add(lowerCamelCase_) UpperCamelCase = [mem.copy() for i in range(4)] UpperCamelCase = VGroup(*lowerCamelCase_).arrange(lowerCamelCase_ , buff=0) UpperCamelCase = Text('''GPU''' , font_size=2_4) UpperCamelCase = Group(lowerCamelCase_ , lowerCamelCase_).arrange(lowerCamelCase_ , buff=0.5 , aligned_edge=lowerCamelCase_) gpu.move_to([-1, -1, 0]) self.add(lowerCamelCase_) UpperCamelCase = [mem.copy() for i in range(6)] UpperCamelCase = VGroup(*lowerCamelCase_).arrange(lowerCamelCase_ , buff=0) UpperCamelCase = Text('''Model''' , font_size=2_4) UpperCamelCase = Group(lowerCamelCase_ , lowerCamelCase_).arrange(lowerCamelCase_ , buff=0.5 , aligned_edge=lowerCamelCase_) model.move_to([3, -1.0, 0]) self.add(lowerCamelCase_) UpperCamelCase = [] for i, rect in enumerate(lowerCamelCase_): rect.set_stroke(lowerCamelCase_) # target = fill.copy().set_fill(YELLOW, opacity=0.7) # target.move_to(rect) # self.add(target) UpperCamelCase = Rectangle(height=0.46 / 4 , width=0.46 / 3).set_stroke(width=0.0).set_fill(lowerCamelCase_ , opacity=0.7) if i == 0: cpu_target.next_to(cpu_left_col_base[0].get_corner(DOWN + LEFT) , buff=0.02 , direction=lowerCamelCase_) cpu_target.set_x(cpu_target.get_x() + 0.1) elif i == 3: cpu_target.next_to(cpu_targs[0] , direction=lowerCamelCase_ , buff=0.0) else: cpu_target.next_to(cpu_targs[i - 1] , direction=lowerCamelCase_ , buff=0.0) self.add(lowerCamelCase_) cpu_targs.append(lowerCamelCase_) UpperCamelCase = [mem.copy() for i in range(6)] UpperCamelCase = VGroup(*lowerCamelCase_).arrange(lowerCamelCase_ , buff=0) UpperCamelCase = Text('''Loaded Checkpoint''' , font_size=2_4) UpperCamelCase = Group(lowerCamelCase_ , lowerCamelCase_).arrange(lowerCamelCase_ , aligned_edge=lowerCamelCase_ , buff=0.4) checkpoint.move_to([3, 0.5, 0]) UpperCamelCase = Square(side_length=2.2) key.move_to([-5, 2, 0]) UpperCamelCase = MarkupText( F'<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model' , font_size=1_8 , ) key_text.move_to([-5, 2.4, 0]) self.add(lowerCamelCase_ , lowerCamelCase_) UpperCamelCase = MarkupText( F'<span fgcolor=\'{BLUE}\'>●</span> Checkpoint' , font_size=1_8 , ) blue_text.next_to(lowerCamelCase_ , DOWN * 2.4 , aligned_edge=key_text.get_left()) UpperCamelCase = MarkupText( F'Next, a <i><span fgcolor="{BLUE}">second</span></i> model is loaded into memory,\nwith the weights of a <span fgcolor="{BLUE}">single shard</span>.' , font_size=2_4 , ) step_a.move_to([2, 2, 0]) self.play(Write(lowerCamelCase_) , Write(lowerCamelCase_)) self.play(Write(lowerCamelCase_ , run_time=1) , Create(lowerCamelCase_ , run_time=1)) UpperCamelCase = [] UpperCamelCase = [] for i, rect in enumerate(lowerCamelCase_): UpperCamelCase = fill.copy().set_fill(lowerCamelCase_ , opacity=0.7) target.move_to(lowerCamelCase_) first_animations.append(GrowFromCenter(lowerCamelCase_ , run_time=1)) UpperCamelCase = target.copy() cpu_target.generate_target() if i < 5: cpu_target.target.move_to(cpu_left_col_base[i + 1]) else: cpu_target.target.move_to(cpu_right_col_base[i - 5]) second_animations.append(MoveToTarget(lowerCamelCase_ , run_time=1.5)) self.play(*lowerCamelCase_) self.play(*lowerCamelCase_) self.wait()
34
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. SCREAMING_SNAKE_CASE_ = abspath(join(dirname(dirname(__file__)), 'src')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='ignore', category=FutureWarning) def __snake_case ( _lowercase ): """simple docstring""" from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(_lowercase ) def __snake_case ( _lowercase ): """simple docstring""" from diffusers.utils.testing_utils import pytest_terminal_summary_main UpperCamelCase = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(_lowercase ,id=_lowercase )
34
1
"""simple docstring""" def __snake_case ( _lowercase = 6008_5147_5143 ): """simple docstring""" try: UpperCamelCase = int(_lowercase ) except (TypeError, ValueError): raise TypeError('''Parameter n must be int or castable to int.''' ) if n <= 0: raise ValueError('''Parameter n must be greater than or equal to one.''' ) UpperCamelCase = 1 UpperCamelCase = 2 while i * i <= n: while n % i == 0: UpperCamelCase = i n //= i i += 1 if n > 1: UpperCamelCase = n return int(_lowercase ) if __name__ == "__main__": print(f'{solution() = }')
34
"""simple docstring""" import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" def __init__( self , *lowerCamelCase_ , **lowerCamelCase_) -> None: warnings.warn( '''The class MobileViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use MobileViTImageProcessor instead.''' , lowerCamelCase_ , ) super().__init__(*lowerCamelCase_ , **lowerCamelCase_)
34
1
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.whisper import WhisperForConditionalGeneration, WhisperProcessor from .base import PipelineTool class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = '''openai/whisper-base''' A_ = ( '''This is a tool that transcribes an audio into text. It takes an input named `audio` and returns the ''' '''transcribed text.''' ) A_ = '''transcriber''' A_ = WhisperProcessor A_ = WhisperForConditionalGeneration A_ = ['''audio'''] A_ = ['''text'''] def UpperCAmelCase__ ( self , lowerCamelCase_) -> List[Any]: return self.pre_processor(lowerCamelCase_ , return_tensors='''pt''').input_features def UpperCAmelCase__ ( self , lowerCamelCase_) -> Any: return self.model.generate(inputs=lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> List[str]: return self.pre_processor.batch_decode(lowerCamelCase_ , skip_special_tokens=lowerCamelCase_)[0]
34
"""simple docstring""" def __snake_case ( _lowercase ): """simple docstring""" UpperCamelCase = [0 for i in range(len(_lowercase ) )] # initialize interval's left pointer and right pointer UpperCamelCase , UpperCamelCase = 0, 0 for i in range(1 ,len(_lowercase ) ): # case when current index is inside the interval if i <= right_pointer: UpperCamelCase = min(right_pointer - i + 1 ,z_result[i - left_pointer] ) UpperCamelCase = min_edge while go_next(_lowercase ,_lowercase ,_lowercase ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: UpperCamelCase , UpperCamelCase = i, i + z_result[i] - 1 return z_result def __snake_case ( _lowercase ,_lowercase ,_lowercase ): """simple docstring""" return i + z_result[i] < len(_lowercase ) and s[z_result[i]] == s[i + z_result[i]] def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" UpperCamelCase = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string UpperCamelCase = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(_lowercase ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
34
1
"""simple docstring""" import argparse import os import torch from transformers import ( XLNetConfig, XLNetForQuestionAnswering, XLNetForSequenceClassification, XLNetLMHeadModel, load_tf_weights_in_xlnet, ) from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging SCREAMING_SNAKE_CASE_ = { 'cola': 2, 'mnli': 3, 'mrpc': 2, 'sst-2': 2, 'sts-b': 1, 'qqp': 2, 'qnli': 2, 'rte': 2, 'wnli': 2, } logging.set_verbosity_info() def __snake_case ( _lowercase ,_lowercase ,_lowercase ,_lowercase=None ): """simple docstring""" UpperCamelCase = XLNetConfig.from_json_file(_lowercase ) UpperCamelCase = finetuning_task.lower() if finetuning_task is not None else '''''' if finetuning_task in GLUE_TASKS_NUM_LABELS: print(f'Building PyTorch XLNetForSequenceClassification model from configuration: {config}' ) UpperCamelCase = finetuning_task UpperCamelCase = GLUE_TASKS_NUM_LABELS[finetuning_task] UpperCamelCase = XLNetForSequenceClassification(_lowercase ) elif "squad" in finetuning_task: UpperCamelCase = finetuning_task UpperCamelCase = XLNetForQuestionAnswering(_lowercase ) else: UpperCamelCase = XLNetLMHeadModel(_lowercase ) # Load weights from tf checkpoint load_tf_weights_in_xlnet(_lowercase ,_lowercase ,_lowercase ) # Save pytorch-model UpperCamelCase = os.path.join(_lowercase ,_lowercase ) UpperCamelCase = os.path.join(_lowercase ,_lowercase ) print(f'Save PyTorch model to {os.path.abspath(_lowercase )}' ) torch.save(model.state_dict() ,_lowercase ) print(f'Save configuration file to {os.path.abspath(_lowercase )}' ) with open(_lowercase ,'''w''' ,encoding='''utf-8''' ) as f: f.write(config.to_json_string() ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--xlnet_config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained XLNet model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the folder to store the PyTorch model or dataset/vocab.', ) parser.add_argument( '--finetuning_task', default=None, type=str, help='Name of a task on which the XLNet TensorFlow model was fine-tuned', ) SCREAMING_SNAKE_CASE_ = parser.parse_args() print(args) convert_xlnet_checkpoint_to_pytorch( args.tf_checkpoint_path, args.xlnet_config_file, args.pytorch_dump_folder_path, args.finetuning_task )
34
"""simple docstring""" import importlib.metadata import warnings from copy import deepcopy from packaging import version from ..utils import logging from .import_utils import is_accelerate_available, is_bitsandbytes_available if is_bitsandbytes_available(): import bitsandbytes as bnb import torch import torch.nn as nn from ..pytorch_utils import ConvaD if is_accelerate_available(): from accelerate import init_empty_weights from accelerate.utils import find_tied_parameters SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) def __snake_case ( _lowercase ,_lowercase ,_lowercase ,_lowercase=None ,_lowercase=None ): """simple docstring""" if "." in tensor_name: UpperCamelCase = tensor_name.split('''.''' ) for split in splits[:-1]: UpperCamelCase = getattr(_lowercase ,_lowercase ) if new_module is None: raise ValueError(f'{module} has no attribute {split}.' ) UpperCamelCase = new_module UpperCamelCase = splits[-1] if tensor_name not in module._parameters and tensor_name not in module._buffers: raise ValueError(f'{module} does not have a parameter or a buffer named {tensor_name}.' ) UpperCamelCase = tensor_name in module._buffers UpperCamelCase = getattr(_lowercase ,_lowercase ) if old_value.device == torch.device('''meta''' ) and device not in ["meta", torch.device('''meta''' )] and value is None: raise ValueError(f'{tensor_name} is on the meta device, we need a `value` to put in on {device}.' ) UpperCamelCase = False UpperCamelCase = False if is_buffer or not is_bitsandbytes_available(): UpperCamelCase = False UpperCamelCase = False else: UpperCamelCase = hasattr(bnb.nn ,'''Params4bit''' ) and isinstance(module._parameters[tensor_name] ,bnb.nn.Paramsabit ) UpperCamelCase = isinstance(module._parameters[tensor_name] ,bnb.nn.IntaParams ) if is_abit or is_abit: UpperCamelCase = module._parameters[tensor_name] if param.device.type != "cuda": if value is None: UpperCamelCase = old_value.to(_lowercase ) elif isinstance(_lowercase ,torch.Tensor ): UpperCamelCase = value.to('''cpu''' ) if value.dtype == torch.inta: UpperCamelCase = version.parse(importlib.metadata.version('''bitsandbytes''' ) ) > version.parse( '''0.37.2''' ) if not is_abit_serializable: raise ValueError( '''Detected int8 weights but the version of bitsandbytes is not compatible with int8 serialization. ''' '''Make sure to download the latest `bitsandbytes` version. `pip install --upgrade bitsandbytes`.''' ) else: UpperCamelCase = torch.tensor(_lowercase ,device='''cpu''' ) # Support models using `Conv1D` in place of `nn.Linear` (e.g. gpt2) by transposing the weight matrix prior to quantization. # Since weights are saved in the correct "orientation", we skip transposing when loading. if issubclass(module.source_cls ,_lowercase ) and fpaa_statistics is None: UpperCamelCase = new_value.T UpperCamelCase = old_value.__dict__ if is_abit: UpperCamelCase = bnb.nn.IntaParams(_lowercase ,requires_grad=_lowercase ,**_lowercase ).to(_lowercase ) elif is_abit: UpperCamelCase = bnb.nn.Paramsabit(_lowercase ,requires_grad=_lowercase ,**_lowercase ).to(_lowercase ) UpperCamelCase = new_value if fpaa_statistics is not None: setattr(module.weight ,'''SCB''' ,fpaa_statistics.to(_lowercase ) ) else: if value is None: UpperCamelCase = old_value.to(_lowercase ) elif isinstance(_lowercase ,torch.Tensor ): UpperCamelCase = value.to(_lowercase ) else: UpperCamelCase = torch.tensor(_lowercase ,device=_lowercase ) if is_buffer: UpperCamelCase = new_value else: UpperCamelCase = nn.Parameter(_lowercase ,requires_grad=old_value.requires_grad ) UpperCamelCase = new_value def __snake_case ( _lowercase ,_lowercase=None ,_lowercase=None ,_lowercase=None ,_lowercase=False ): """simple docstring""" for name, module in model.named_children(): if current_key_name is None: UpperCamelCase = [] current_key_name.append(_lowercase ) if (isinstance(_lowercase ,nn.Linear ) or isinstance(_lowercase ,_lowercase )) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` if not any(key in '''.'''.join(_lowercase ) for key in modules_to_not_convert ): with init_empty_weights(): if isinstance(_lowercase ,_lowercase ): UpperCamelCase , UpperCamelCase = module.weight.shape else: UpperCamelCase = module.in_features UpperCamelCase = module.out_features if quantization_config.quantization_method() == "llm_int8": UpperCamelCase = bnb.nn.LinearabitLt( _lowercase ,_lowercase ,module.bias is not None ,has_fpaa_weights=quantization_config.llm_inta_has_fpaa_weight ,threshold=quantization_config.llm_inta_threshold ,) UpperCamelCase = True else: if ( quantization_config.llm_inta_skip_modules is not None and name in quantization_config.llm_inta_skip_modules ): pass else: UpperCamelCase = bnb.nn.Linearabit( _lowercase ,_lowercase ,module.bias is not None ,quantization_config.bnb_abit_compute_dtype ,compress_statistics=quantization_config.bnb_abit_use_double_quant ,quant_type=quantization_config.bnb_abit_quant_type ,) UpperCamelCase = True # Store the module class in case we need to transpose the weight later UpperCamelCase = type(_lowercase ) # Force requires grad to False to avoid unexpected errors model._modules[name].requires_grad_(_lowercase ) if len(list(module.children() ) ) > 0: UpperCamelCase , UpperCamelCase = _replace_with_bnb_linear( _lowercase ,_lowercase ,_lowercase ,_lowercase ,has_been_replaced=_lowercase ,) # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def __snake_case ( _lowercase ,_lowercase=None ,_lowercase=None ,_lowercase=None ): """simple docstring""" UpperCamelCase = ['''lm_head'''] if modules_to_not_convert is None else modules_to_not_convert UpperCamelCase , UpperCamelCase = _replace_with_bnb_linear( _lowercase ,_lowercase ,_lowercase ,_lowercase ) if not has_been_replaced: logger.warning( '''You are loading your model in 8bit or 4bit but no linear modules were found in your model.''' ''' Please double check your model architecture, or submit an issue on github if you think this is''' ''' a bug.''' ) return model def __snake_case ( *_lowercase ,**_lowercase ): """simple docstring""" warnings.warn( '''`replace_8bit_linear` will be deprecated in a future version, please use `replace_with_bnb_linear` instead''' ,_lowercase ,) return replace_with_bnb_linear(*_lowercase ,**_lowercase ) def __snake_case ( *_lowercase ,**_lowercase ): """simple docstring""" warnings.warn( '''`set_module_8bit_tensor_to_device` will be deprecated in a future version, please use `set_module_quantized_tensor_to_device` instead''' ,_lowercase ,) return set_module_quantized_tensor_to_device(*_lowercase ,**_lowercase ) def __snake_case ( _lowercase ): """simple docstring""" UpperCamelCase = deepcopy(_lowercase ) # this has 0 cost since it is done inside `init_empty_weights` context manager` tied_model.tie_weights() UpperCamelCase = find_tied_parameters(_lowercase ) # For compatibility with Accelerate < 0.18 if isinstance(_lowercase ,_lowercase ): UpperCamelCase = sum(list(tied_params.values() ) ,[] ) + list(tied_params.keys() ) else: UpperCamelCase = sum(_lowercase ,[] ) UpperCamelCase = len(_lowercase ) > 0 # Check if it is a base model UpperCamelCase = not hasattr(_lowercase ,model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head UpperCamelCase = list(model.named_children() ) UpperCamelCase = [list_modules[-1][0]] # add last module together with tied weights UpperCamelCase = set(_lowercase ) - set(_lowercase ) UpperCamelCase = list(set(_lowercase ) ) + list(_lowercase ) # remove ".weight" from the keys UpperCamelCase = ['''.weight''', '''.bias'''] UpperCamelCase = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: UpperCamelCase = name.replace(_lowercase ,'''''' ) filtered_module_names.append(_lowercase ) return filtered_module_names
34
1
"""simple docstring""" from unittest.mock import patch import pyspark from datasets.packaged_modules.spark.spark import ( Spark, SparkExamplesIterable, _generate_iterable_examples, ) from ..utils import ( require_dill_gt_0_3_2, require_not_windows, ) def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" UpperCamelCase = [] for part_id in partition_order: UpperCamelCase = df.where(f'SPARK_PARTITION_ID() = {part_id}' ).collect() for row_idx, row in enumerate(_lowercase ): expected_row_ids_and_row_dicts.append((f'{part_id}_{row_idx}', row.asDict()) ) return expected_row_ids_and_row_dicts @require_not_windows @require_dill_gt_0_3_2 def __snake_case ( ): """simple docstring""" UpperCamelCase = pyspark.sql.SparkSession.builder.master('''local[*]''' ).appName('''pyspark''' ).getOrCreate() UpperCamelCase = spark.range(100 ).repartition(1 ) UpperCamelCase = Spark(_lowercase ) # The id ints will be converted to Pyarrow int64s, so each row will be 8 bytes. Setting a max_shard_size of 16 means # that each partition can hold 2 rows. spark_builder._repartition_df_if_needed(max_shard_size=16 ) # Given that the dataframe has 100 rows and each partition has 2 rows, we expect 50 partitions. assert spark_builder.df.rdd.getNumPartitions() == 50 @require_not_windows @require_dill_gt_0_3_2 def __snake_case ( ): """simple docstring""" UpperCamelCase = pyspark.sql.SparkSession.builder.master('''local[*]''' ).appName('''pyspark''' ).getOrCreate() UpperCamelCase = spark.range(10 ).repartition(2 ) UpperCamelCase = [1, 0] UpperCamelCase = _generate_iterable_examples(_lowercase ,_lowercase ) # Reverse the partitions. UpperCamelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(_lowercase ,_lowercase ) for i, (row_id, row_dict) in enumerate(generate_fn() ): UpperCamelCase , UpperCamelCase = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def __snake_case ( ): """simple docstring""" UpperCamelCase = pyspark.sql.SparkSession.builder.master('''local[*]''' ).appName('''pyspark''' ).getOrCreate() UpperCamelCase = spark.range(10 ).repartition(1 ) UpperCamelCase = SparkExamplesIterable(_lowercase ) assert it.n_shards == 1 for i, (row_id, row_dict) in enumerate(_lowercase ): assert row_id == f'0_{i}' assert row_dict == {"id": i} @require_not_windows @require_dill_gt_0_3_2 def __snake_case ( ): """simple docstring""" UpperCamelCase = pyspark.sql.SparkSession.builder.master('''local[*]''' ).appName('''pyspark''' ).getOrCreate() UpperCamelCase = spark.range(30 ).repartition(3 ) # Mock the generator so that shuffle reverses the partition indices. with patch('''numpy.random.Generator''' ) as generator_mock: UpperCamelCase = lambda _lowercase : x.reverse() UpperCamelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(_lowercase ,[2, 1, 0] ) UpperCamelCase = SparkExamplesIterable(_lowercase ).shuffle_data_sources(_lowercase ) assert shuffled_it.n_shards == 3 for i, (row_id, row_dict) in enumerate(_lowercase ): UpperCamelCase , UpperCamelCase = expected_row_ids_and_row_dicts[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def __snake_case ( ): """simple docstring""" UpperCamelCase = pyspark.sql.SparkSession.builder.master('''local[*]''' ).appName('''pyspark''' ).getOrCreate() UpperCamelCase = spark.range(20 ).repartition(4 ) # Partitions 0 and 2 UpperCamelCase = SparkExamplesIterable(_lowercase ).shard_data_sources(worker_id=0 ,num_workers=2 ) assert shard_it_a.n_shards == 2 UpperCamelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(_lowercase ,[0, 2] ) for i, (row_id, row_dict) in enumerate(_lowercase ): UpperCamelCase , UpperCamelCase = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict # Partitions 1 and 3 UpperCamelCase = SparkExamplesIterable(_lowercase ).shard_data_sources(worker_id=1 ,num_workers=2 ) assert shard_it_a.n_shards == 2 UpperCamelCase = _get_expected_row_ids_and_row_dicts_for_partition_order(_lowercase ,[1, 3] ) for i, (row_id, row_dict) in enumerate(_lowercase ): UpperCamelCase , UpperCamelCase = expected_row_ids_and_row_dicts_a[i] assert row_id == expected_row_id assert row_dict == expected_row_dict @require_not_windows @require_dill_gt_0_3_2 def __snake_case ( ): """simple docstring""" UpperCamelCase = pyspark.sql.SparkSession.builder.master('''local[*]''' ).appName('''pyspark''' ).getOrCreate() UpperCamelCase = spark.range(100 ).repartition(1 ) UpperCamelCase = Spark(_lowercase ) # Choose a small max_shard_size for maximum partitioning. spark_builder._repartition_df_if_needed(max_shard_size=1 ) # The new number of partitions should not be greater than the number of rows. assert spark_builder.df.rdd.getNumPartitions() == 100
34
"""simple docstring""" from random import randint from tempfile import TemporaryFile import numpy as np def __snake_case ( _lowercase ,_lowercase ,_lowercase ): """simple docstring""" UpperCamelCase = 0 if start < end: UpperCamelCase = randint(_lowercase ,_lowercase ) UpperCamelCase = a[end] UpperCamelCase = a[pivot] UpperCamelCase = temp UpperCamelCase , UpperCamelCase = _in_place_partition(_lowercase ,_lowercase ,_lowercase ) count += _in_place_quick_sort(_lowercase ,_lowercase ,p - 1 ) count += _in_place_quick_sort(_lowercase ,p + 1 ,_lowercase ) return count def __snake_case ( _lowercase ,_lowercase ,_lowercase ): """simple docstring""" UpperCamelCase = 0 UpperCamelCase = randint(_lowercase ,_lowercase ) UpperCamelCase = a[end] UpperCamelCase = a[pivot] UpperCamelCase = temp UpperCamelCase = start - 1 for index in range(_lowercase ,_lowercase ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value UpperCamelCase = new_pivot_index + 1 UpperCamelCase = a[new_pivot_index] UpperCamelCase = a[index] UpperCamelCase = temp UpperCamelCase = a[new_pivot_index + 1] UpperCamelCase = a[end] UpperCamelCase = temp return new_pivot_index + 1, count SCREAMING_SNAKE_CASE_ = TemporaryFile() SCREAMING_SNAKE_CASE_ = 100 # 1000 elements are to be sorted SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0, 1 # mean and standard deviation SCREAMING_SNAKE_CASE_ = np.random.normal(mu, sigma, p) np.save(outfile, X) print('The array is') print(X) outfile.seek(0) # using the same array SCREAMING_SNAKE_CASE_ = np.load(outfile) SCREAMING_SNAKE_CASE_ = len(M) - 1 SCREAMING_SNAKE_CASE_ = _in_place_quick_sort(M, 0, r) print( 'No of Comparisons for 100 elements selected from a standard normal distribution' 'is :' ) print(z)
34
1
"""simple docstring""" import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class snake_case_ ( lowerCamelCase_ , unittest.TestCase ): """simple docstring""" A_ = BarthezTokenizer A_ = BarthezTokenizerFast A_ = True A_ = True def UpperCAmelCase__ ( self) -> Dict: super().setUp() UpperCamelCase = BarthezTokenizerFast.from_pretrained('''moussaKam/mbarthez''') tokenizer.save_pretrained(self.tmpdirname) tokenizer.save_pretrained(self.tmpdirname , legacy_format=lowerCamelCase_) UpperCamelCase = tokenizer def UpperCAmelCase__ ( self) -> Optional[Any]: UpperCamelCase = '''<pad>''' UpperCamelCase = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase_) , lowerCamelCase_) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase_) , lowerCamelCase_) def UpperCAmelCase__ ( self) -> Tuple: UpperCamelCase = list(self.get_tokenizer().get_vocab().keys()) self.assertEqual(vocab_keys[0] , '''<s>''') self.assertEqual(vocab_keys[1] , '''<pad>''') self.assertEqual(vocab_keys[-1] , '''<mask>''') self.assertEqual(len(lowerCamelCase_) , 1_0_1_1_2_2) def UpperCAmelCase__ ( self) -> Optional[int]: self.assertEqual(self.get_tokenizer().vocab_size , 1_0_1_1_2_2) @require_torch def UpperCAmelCase__ ( self) -> Optional[Any]: UpperCamelCase = ['''A long paragraph for summarization.''', '''Another paragraph for summarization.'''] UpperCamelCase = [0, 5_7, 3_0_1_8, 7_0_3_0_7, 9_1, 2] UpperCamelCase = self.tokenizer( lowerCamelCase_ , max_length=len(lowerCamelCase_) , padding=lowerCamelCase_ , truncation=lowerCamelCase_ , return_tensors='''pt''') self.assertIsInstance(lowerCamelCase_ , lowerCamelCase_) self.assertEqual((2, 6) , batch.input_ids.shape) self.assertEqual((2, 6) , batch.attention_mask.shape) UpperCamelCase = batch.input_ids.tolist()[0] self.assertListEqual(lowerCamelCase_ , lowerCamelCase_) def UpperCAmelCase__ ( self) -> List[str]: if not self.test_rust_tokenizer: return UpperCamelCase = self.get_tokenizer() UpperCamelCase = self.get_rust_tokenizer() UpperCamelCase = '''I was born in 92000, and this is falsé.''' UpperCamelCase = tokenizer.tokenize(lowerCamelCase_) UpperCamelCase = rust_tokenizer.tokenize(lowerCamelCase_) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_) UpperCamelCase = tokenizer.encode(lowerCamelCase_ , add_special_tokens=lowerCamelCase_) UpperCamelCase = rust_tokenizer.encode(lowerCamelCase_ , add_special_tokens=lowerCamelCase_) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_) UpperCamelCase = self.get_rust_tokenizer() UpperCamelCase = tokenizer.encode(lowerCamelCase_) UpperCamelCase = rust_tokenizer.encode(lowerCamelCase_) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_) @slow def UpperCAmelCase__ ( self) -> List[str]: # fmt: off UpperCamelCase = {'''input_ids''': [[0, 4_9_0, 1_4_3_2_8, 4_5_0_7, 3_5_4, 4_7, 4_3_6_6_9, 9_5, 2_5, 7_8_1_1_7, 2_0_2_1_5, 1_9_7_7_9, 1_9_0, 2_2, 4_0_0, 4, 3_5_3_4_3, 8_0_3_1_0, 6_0_3, 8_6, 2_4_9_3_7, 1_0_5, 3_3_4_3_8, 9_4_7_6_2, 1_9_6, 3_9_6_4_2, 7, 1_5, 1_5_9_3_3, 1_7_3, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_0_5_3_4, 8_7, 2_5, 6_6, 3_3_5_8, 1_9_6, 5_5_2_8_9, 8, 8_2_9_6_1, 8_1, 2_2_0_4, 7_5_2_0_3, 7, 1_5, 7_6_3, 1_2_9_5_6, 2_1_6, 1_7_8, 1_4_3_2_8, 9_5_9_5, 1_3_7_7, 6_9_6_9_3, 7, 4_4_8, 7_1_0_2_1, 1_9_6, 1_8_1_0_6, 1_4_3_7, 1_3_9_7_4, 1_0_8, 9_0_8_3, 4, 4_9_3_1_5, 7, 3_9, 8_6, 1_3_2_6, 2_7_9_3, 4_6_3_3_3, 4, 4_4_8, 1_9_6, 7_4_5_8_8, 7, 4_9_3_1_5, 7, 3_9, 2_1, 8_2_2, 3_8_4_7_0, 7_4, 2_1, 6_6_7_2_3, 6_2_4_8_0, 8, 2_2_0_5_0, 5, 2]], '''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. UpperCamelCase = [ '''Le transformeur est un modèle d\'apprentissage profond introduit en 2017, ''' '''utilisé principalement dans le domaine du traitement automatique des langues (TAL).''', '''À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus ''' '''pour gérer des données séquentielles, telles que le langage naturel, pour des tâches ''' '''telles que la traduction et la synthèse de texte.''', ] self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase_ , model_name='''moussaKam/mbarthez''' , revision='''c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6''' , sequences=lowerCamelCase_ , )
34
"""simple docstring""" import os import sys import unittest SCREAMING_SNAKE_CASE_ = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, 'utils')) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, create_dummy_object, find_backend, read_init # noqa: E402 # Align TRANSFORMERS_PATH in check_dummies with the current path SCREAMING_SNAKE_CASE_ = os.path.join(git_repo_path, 'src', 'transformers') SCREAMING_SNAKE_CASE_ = '\n{0} = None\n' SCREAMING_SNAKE_CASE_ = '\nclass {0}(metaclass=DummyObject):\n _backends = {1}\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, {1})\n' SCREAMING_SNAKE_CASE_ = '\ndef {0}(*args, **kwargs):\n requires_backends({0}, {1})\n' class snake_case_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self) -> List[Any]: UpperCamelCase = find_backend(''' _import_structure["models.albert"].append("AlbertTokenizerFast")''') self.assertIsNone(lowerCamelCase_) UpperCamelCase = find_backend(''' if not is_tokenizers_available():''') self.assertEqual(lowerCamelCase_ , '''tokenizers''') UpperCamelCase = find_backend(''' if not is_tensorflow_text_available():''') self.assertEqual(lowerCamelCase_ , '''tensorflow_text''') UpperCamelCase = find_backend(''' if not (is_sentencepiece_available() and is_tokenizers_available()):''') self.assertEqual(lowerCamelCase_ , '''sentencepiece_and_tokenizers''') UpperCamelCase = find_backend( ''' if not (is_sentencepiece_available() and is_tensorflow_text_available()):''') self.assertEqual(lowerCamelCase_ , '''sentencepiece_and_tensorflow_text''') UpperCamelCase = find_backend( ''' if not (is_sentencepiece_available() and is_tokenizers_available() and is_vision_available()):''') self.assertEqual(lowerCamelCase_ , '''sentencepiece_and_tokenizers_and_vision''') def UpperCAmelCase__ ( self) -> int: UpperCamelCase = read_init() # We don't assert on the exact list of keys to allow for smooth grow of backend-specific objects self.assertIn('''torch''' , lowerCamelCase_) self.assertIn('''tensorflow_text''' , lowerCamelCase_) self.assertIn('''sentencepiece_and_tokenizers''' , lowerCamelCase_) # Likewise, we can't assert on the exact content of a key self.assertIn('''BertModel''' , objects['''torch''']) self.assertIn('''TFBertModel''' , objects['''tf''']) self.assertIn('''FlaxBertModel''' , objects['''flax''']) self.assertIn('''BertModel''' , objects['''torch''']) self.assertIn('''TFBertTokenizer''' , objects['''tensorflow_text''']) self.assertIn('''convert_slow_tokenizer''' , objects['''sentencepiece_and_tokenizers''']) def UpperCAmelCase__ ( self) -> Optional[int]: UpperCamelCase = create_dummy_object('''CONSTANT''' , '''\'torch\'''') self.assertEqual(lowerCamelCase_ , '''\nCONSTANT = None\n''') UpperCamelCase = create_dummy_object('''function''' , '''\'torch\'''') self.assertEqual( lowerCamelCase_ , '''\ndef function(*args, **kwargs):\n requires_backends(function, \'torch\')\n''') UpperCamelCase = ''' class FakeClass(metaclass=DummyObject): _backends = \'torch\' def __init__(self, *args, **kwargs): requires_backends(self, \'torch\') ''' UpperCamelCase = create_dummy_object('''FakeClass''' , '''\'torch\'''') self.assertEqual(lowerCamelCase_ , lowerCamelCase_) def UpperCAmelCase__ ( self) -> int: UpperCamelCase = '''# This file is autogenerated by the command `make fix-copies`, do not edit. from ..utils import DummyObject, requires_backends CONSTANT = None def function(*args, **kwargs): requires_backends(function, ["torch"]) class FakeClass(metaclass=DummyObject): _backends = ["torch"] def __init__(self, *args, **kwargs): requires_backends(self, ["torch"]) ''' UpperCamelCase = create_dummy_files({'''torch''': ['''CONSTANT''', '''function''', '''FakeClass''']}) self.assertEqual(dummy_files['''torch'''] , lowerCamelCase_)
34
1
"""simple docstring""" import unittest from transformers import MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING, is_vision_available from transformers.pipelines import pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image else: class snake_case_ : """simple docstring""" @staticmethod def UpperCAmelCase__ ( *lowerCamelCase_ , **lowerCamelCase_) -> Optional[int]: pass @is_pipeline_test @require_torch @require_vision class snake_case_ ( unittest.TestCase ): """simple docstring""" A_ = MODEL_FOR_VISUAL_QUESTION_ANSWERING_MAPPING def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_) -> Dict: UpperCamelCase = pipeline('''visual-question-answering''' , model='''hf-internal-testing/tiny-vilt-random-vqa''') UpperCamelCase = [ { '''image''': Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png'''), '''question''': '''How many cats are there?''', }, { '''image''': '''./tests/fixtures/tests_samples/COCO/000000039769.png''', '''question''': '''How many cats are there?''', }, ] return vqa_pipeline, examples def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_) -> Dict: UpperCamelCase = vqa_pipeline(lowerCamelCase_ , top_k=1) self.assertEqual( lowerCamelCase_ , [ [{'''score''': ANY(lowerCamelCase_), '''answer''': ANY(lowerCamelCase_)}], [{'''score''': ANY(lowerCamelCase_), '''answer''': ANY(lowerCamelCase_)}], ] , ) @require_torch def UpperCAmelCase__ ( self) -> Union[str, Any]: UpperCamelCase = pipeline('''visual-question-answering''' , model='''hf-internal-testing/tiny-vilt-random-vqa''') UpperCamelCase = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' UpperCamelCase = '''How many cats are there?''' UpperCamelCase = vqa_pipeline(image=lowerCamelCase_ , question='''How many cats are there?''' , top_k=2) self.assertEqual( lowerCamelCase_ , [{'''score''': ANY(lowerCamelCase_), '''answer''': ANY(lowerCamelCase_)}, {'''score''': ANY(lowerCamelCase_), '''answer''': ANY(lowerCamelCase_)}]) UpperCamelCase = vqa_pipeline({'''image''': image, '''question''': question} , top_k=2) self.assertEqual( lowerCamelCase_ , [{'''score''': ANY(lowerCamelCase_), '''answer''': ANY(lowerCamelCase_)}, {'''score''': ANY(lowerCamelCase_), '''answer''': ANY(lowerCamelCase_)}]) @slow @require_torch def UpperCAmelCase__ ( self) -> Optional[int]: UpperCamelCase = pipeline('''visual-question-answering''' , model='''dandelin/vilt-b32-finetuned-vqa''') UpperCamelCase = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' UpperCamelCase = '''How many cats are there?''' UpperCamelCase = vqa_pipeline(image=lowerCamelCase_ , question=lowerCamelCase_ , top_k=2) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4) , [{'''score''': 0.8799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}]) UpperCamelCase = vqa_pipeline({'''image''': image, '''question''': question} , top_k=2) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4) , [{'''score''': 0.8799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}]) UpperCamelCase = vqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] , top_k=2) self.assertEqual( nested_simplify(lowerCamelCase_ , decimals=4) , [[{'''score''': 0.8799, '''answer''': '''2'''}, {'''score''': 0.296, '''answer''': '''1'''}]] * 2 , ) @require_tf @unittest.skip('''Visual question answering not implemented in TF''') def UpperCAmelCase__ ( self) -> Optional[int]: pass
34
"""simple docstring""" import argparse import requests import torch from PIL import Image from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor def __snake_case ( _lowercase ): """simple docstring""" if "cls_token" in name: UpperCamelCase = name.replace('''cls_token''' ,'''vit.embeddings.cls_token''' ) if "mask_token" in name: UpperCamelCase = name.replace('''mask_token''' ,'''decoder.mask_token''' ) if "decoder_pos_embed" in name: UpperCamelCase = name.replace('''decoder_pos_embed''' ,'''decoder.decoder_pos_embed''' ) if "pos_embed" in name and "decoder" not in name: UpperCamelCase = name.replace('''pos_embed''' ,'''vit.embeddings.position_embeddings''' ) if "patch_embed.proj" in name: UpperCamelCase = name.replace('''patch_embed.proj''' ,'''vit.embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: UpperCamelCase = name.replace('''patch_embed.norm''' ,'''vit.embeddings.norm''' ) if "decoder_blocks" in name: UpperCamelCase = name.replace('''decoder_blocks''' ,'''decoder.decoder_layers''' ) if "blocks" in name: UpperCamelCase = name.replace('''blocks''' ,'''vit.encoder.layer''' ) if "attn.proj" in name: UpperCamelCase = name.replace('''attn.proj''' ,'''attention.output.dense''' ) if "attn" in name: UpperCamelCase = name.replace('''attn''' ,'''attention.self''' ) if "norm1" in name: UpperCamelCase = name.replace('''norm1''' ,'''layernorm_before''' ) if "norm2" in name: UpperCamelCase = name.replace('''norm2''' ,'''layernorm_after''' ) if "mlp.fc1" in name: UpperCamelCase = name.replace('''mlp.fc1''' ,'''intermediate.dense''' ) if "mlp.fc2" in name: UpperCamelCase = name.replace('''mlp.fc2''' ,'''output.dense''' ) if "decoder_embed" in name: UpperCamelCase = name.replace('''decoder_embed''' ,'''decoder.decoder_embed''' ) if "decoder_norm" in name: UpperCamelCase = name.replace('''decoder_norm''' ,'''decoder.decoder_norm''' ) if "decoder_pred" in name: UpperCamelCase = name.replace('''decoder_pred''' ,'''decoder.decoder_pred''' ) if "norm.weight" in name and "decoder" not in name: UpperCamelCase = name.replace('''norm.weight''' ,'''vit.layernorm.weight''' ) if "norm.bias" in name and "decoder" not in name: UpperCamelCase = name.replace('''norm.bias''' ,'''vit.layernorm.bias''' ) return name def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" for key in orig_state_dict.copy().keys(): UpperCamelCase = orig_state_dict.pop(_lowercase ) if "qkv" in key: UpperCamelCase = key.split('''.''' ) UpperCamelCase = int(key_split[1] ) if "decoder_blocks" in key: UpperCamelCase = config.decoder_hidden_size UpperCamelCase = '''decoder.decoder_layers.''' if "weight" in key: UpperCamelCase = val[:dim, :] UpperCamelCase = val[dim : dim * 2, :] UpperCamelCase = val[-dim:, :] elif "bias" in key: UpperCamelCase = val[:dim] UpperCamelCase = val[dim : dim * 2] UpperCamelCase = val[-dim:] else: UpperCamelCase = config.hidden_size UpperCamelCase = '''vit.encoder.layer.''' if "weight" in key: UpperCamelCase = val[:dim, :] UpperCamelCase = val[dim : dim * 2, :] UpperCamelCase = val[-dim:, :] elif "bias" in key: UpperCamelCase = val[:dim] UpperCamelCase = val[dim : dim * 2] UpperCamelCase = val[-dim:] else: UpperCamelCase = val return orig_state_dict def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" UpperCamelCase = ViTMAEConfig() if "large" in checkpoint_url: UpperCamelCase = 1024 UpperCamelCase = 4096 UpperCamelCase = 24 UpperCamelCase = 16 elif "huge" in checkpoint_url: UpperCamelCase = 14 UpperCamelCase = 1280 UpperCamelCase = 5120 UpperCamelCase = 32 UpperCamelCase = 16 UpperCamelCase = ViTMAEForPreTraining(_lowercase ) UpperCamelCase = torch.hub.load_state_dict_from_url(_lowercase ,map_location='''cpu''' )['''model'''] UpperCamelCase = ViTMAEImageProcessor(size=config.image_size ) UpperCamelCase = convert_state_dict(_lowercase ,_lowercase ) model.load_state_dict(_lowercase ) model.eval() UpperCamelCase = '''https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg''' UpperCamelCase = Image.open(requests.get(_lowercase ,stream=_lowercase ).raw ) UpperCamelCase = ViTMAEImageProcessor(size=config.image_size ) UpperCamelCase = image_processor(images=_lowercase ,return_tensors='''pt''' ) # forward pass torch.manual_seed(2 ) UpperCamelCase = model(**_lowercase ) UpperCamelCase = outputs.logits if "large" in checkpoint_url: UpperCamelCase = torch.tensor( [[-0.7309, -0.7128, -1.0169], [-1.0161, -0.9058, -1.1878], [-1.0478, -0.9411, -1.1911]] ) elif "huge" in checkpoint_url: UpperCamelCase = torch.tensor( [[-1.1599, -0.9199, -1.2221], [-1.1952, -0.9269, -1.2307], [-1.2143, -0.9337, -1.2262]] ) else: UpperCamelCase = torch.tensor( [[-0.9192, -0.8481, -1.1259], [-1.1349, -1.0034, -1.2599], [-1.1757, -1.0429, -1.2726]] ) # verify logits assert torch.allclose(logits[0, :3, :3] ,_lowercase ,atol=1e-4 ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(_lowercase ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(_lowercase ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://dl.fbaipublicfiles.com/mae/visualize/mae_visualize_vit_base.pth', type=str, help='URL of the checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) SCREAMING_SNAKE_CASE_ = parser.parse_args() convert_vit_mae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
34
1
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class snake_case_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self) -> Union[str, Any]: UpperCamelCase = { '''task_specific_params''': { '''summarization''': {'''length_penalty''': 1.0, '''max_length''': 1_2_8, '''min_length''': 1_2, '''num_beams''': 4}, '''summarization_cnn''': {'''length_penalty''': 2.0, '''max_length''': 1_4_2, '''min_length''': 5_6, '''num_beams''': 4}, '''summarization_xsum''': {'''length_penalty''': 1.0, '''max_length''': 6_2, '''min_length''': 1_1, '''num_beams''': 6}, } } UpperCamelCase = { '''task_specific_params.summarization.length_penalty''': 1.0, '''task_specific_params.summarization.max_length''': 1_2_8, '''task_specific_params.summarization.min_length''': 1_2, '''task_specific_params.summarization.num_beams''': 4, '''task_specific_params.summarization_cnn.length_penalty''': 2.0, '''task_specific_params.summarization_cnn.max_length''': 1_4_2, '''task_specific_params.summarization_cnn.min_length''': 5_6, '''task_specific_params.summarization_cnn.num_beams''': 4, '''task_specific_params.summarization_xsum.length_penalty''': 1.0, '''task_specific_params.summarization_xsum.max_length''': 6_2, '''task_specific_params.summarization_xsum.min_length''': 1_1, '''task_specific_params.summarization_xsum.num_beams''': 6, } self.assertEqual(flatten_dict(lowerCamelCase_) , lowerCamelCase_) def UpperCAmelCase__ ( self) -> Optional[Any]: UpperCamelCase = np.random.randn(3 , 4) self.assertTrue(np.allclose(transpose(lowerCamelCase_) , x.transpose())) UpperCamelCase = np.random.randn(3 , 4 , 5) self.assertTrue(np.allclose(transpose(lowerCamelCase_ , axes=(1, 2, 0)) , x.transpose((1, 2, 0)))) @require_torch def UpperCAmelCase__ ( self) -> Dict: UpperCamelCase = np.random.randn(3 , 4) UpperCamelCase = torch.tensor(lowerCamelCase_) self.assertTrue(np.allclose(transpose(lowerCamelCase_) , transpose(lowerCamelCase_).numpy())) UpperCamelCase = np.random.randn(3 , 4 , 5) UpperCamelCase = torch.tensor(lowerCamelCase_) self.assertTrue(np.allclose(transpose(lowerCamelCase_ , axes=(1, 2, 0)) , transpose(lowerCamelCase_ , axes=(1, 2, 0)).numpy())) @require_tf def UpperCAmelCase__ ( self) -> Any: UpperCamelCase = np.random.randn(3 , 4) UpperCamelCase = tf.constant(lowerCamelCase_) self.assertTrue(np.allclose(transpose(lowerCamelCase_) , transpose(lowerCamelCase_).numpy())) UpperCamelCase = np.random.randn(3 , 4 , 5) UpperCamelCase = tf.constant(lowerCamelCase_) self.assertTrue(np.allclose(transpose(lowerCamelCase_ , axes=(1, 2, 0)) , transpose(lowerCamelCase_ , axes=(1, 2, 0)).numpy())) @require_flax def UpperCAmelCase__ ( self) -> Dict: UpperCamelCase = np.random.randn(3 , 4) UpperCamelCase = jnp.array(lowerCamelCase_) self.assertTrue(np.allclose(transpose(lowerCamelCase_) , np.asarray(transpose(lowerCamelCase_)))) UpperCamelCase = np.random.randn(3 , 4 , 5) UpperCamelCase = jnp.array(lowerCamelCase_) self.assertTrue(np.allclose(transpose(lowerCamelCase_ , axes=(1, 2, 0)) , np.asarray(transpose(lowerCamelCase_ , axes=(1, 2, 0))))) def UpperCAmelCase__ ( self) -> List[Any]: UpperCamelCase = np.random.randn(3 , 4) self.assertTrue(np.allclose(reshape(lowerCamelCase_ , (4, 3)) , np.reshape(lowerCamelCase_ , (4, 3)))) UpperCamelCase = np.random.randn(3 , 4 , 5) self.assertTrue(np.allclose(reshape(lowerCamelCase_ , (1_2, 5)) , np.reshape(lowerCamelCase_ , (1_2, 5)))) @require_torch def UpperCAmelCase__ ( self) -> List[Any]: UpperCamelCase = np.random.randn(3 , 4) UpperCamelCase = torch.tensor(lowerCamelCase_) self.assertTrue(np.allclose(reshape(lowerCamelCase_ , (4, 3)) , reshape(lowerCamelCase_ , (4, 3)).numpy())) UpperCamelCase = np.random.randn(3 , 4 , 5) UpperCamelCase = torch.tensor(lowerCamelCase_) self.assertTrue(np.allclose(reshape(lowerCamelCase_ , (1_2, 5)) , reshape(lowerCamelCase_ , (1_2, 5)).numpy())) @require_tf def UpperCAmelCase__ ( self) -> List[Any]: UpperCamelCase = np.random.randn(3 , 4) UpperCamelCase = tf.constant(lowerCamelCase_) self.assertTrue(np.allclose(reshape(lowerCamelCase_ , (4, 3)) , reshape(lowerCamelCase_ , (4, 3)).numpy())) UpperCamelCase = np.random.randn(3 , 4 , 5) UpperCamelCase = tf.constant(lowerCamelCase_) self.assertTrue(np.allclose(reshape(lowerCamelCase_ , (1_2, 5)) , reshape(lowerCamelCase_ , (1_2, 5)).numpy())) @require_flax def UpperCAmelCase__ ( self) -> Tuple: UpperCamelCase = np.random.randn(3 , 4) UpperCamelCase = jnp.array(lowerCamelCase_) self.assertTrue(np.allclose(reshape(lowerCamelCase_ , (4, 3)) , np.asarray(reshape(lowerCamelCase_ , (4, 3))))) UpperCamelCase = np.random.randn(3 , 4 , 5) UpperCamelCase = jnp.array(lowerCamelCase_) self.assertTrue(np.allclose(reshape(lowerCamelCase_ , (1_2, 5)) , np.asarray(reshape(lowerCamelCase_ , (1_2, 5))))) def UpperCAmelCase__ ( self) -> List[str]: UpperCamelCase = np.random.randn(1 , 3 , 4) self.assertTrue(np.allclose(squeeze(lowerCamelCase_) , np.squeeze(lowerCamelCase_))) UpperCamelCase = np.random.randn(1 , 4 , 1 , 5) self.assertTrue(np.allclose(squeeze(lowerCamelCase_ , axis=2) , np.squeeze(lowerCamelCase_ , axis=2))) @require_torch def UpperCAmelCase__ ( self) -> int: UpperCamelCase = np.random.randn(1 , 3 , 4) UpperCamelCase = torch.tensor(lowerCamelCase_) self.assertTrue(np.allclose(squeeze(lowerCamelCase_) , squeeze(lowerCamelCase_).numpy())) UpperCamelCase = np.random.randn(1 , 4 , 1 , 5) UpperCamelCase = torch.tensor(lowerCamelCase_) self.assertTrue(np.allclose(squeeze(lowerCamelCase_ , axis=2) , squeeze(lowerCamelCase_ , axis=2).numpy())) @require_tf def UpperCAmelCase__ ( self) -> List[Any]: UpperCamelCase = np.random.randn(1 , 3 , 4) UpperCamelCase = tf.constant(lowerCamelCase_) self.assertTrue(np.allclose(squeeze(lowerCamelCase_) , squeeze(lowerCamelCase_).numpy())) UpperCamelCase = np.random.randn(1 , 4 , 1 , 5) UpperCamelCase = tf.constant(lowerCamelCase_) self.assertTrue(np.allclose(squeeze(lowerCamelCase_ , axis=2) , squeeze(lowerCamelCase_ , axis=2).numpy())) @require_flax def UpperCAmelCase__ ( self) -> str: UpperCamelCase = np.random.randn(1 , 3 , 4) UpperCamelCase = jnp.array(lowerCamelCase_) self.assertTrue(np.allclose(squeeze(lowerCamelCase_) , np.asarray(squeeze(lowerCamelCase_)))) UpperCamelCase = np.random.randn(1 , 4 , 1 , 5) UpperCamelCase = jnp.array(lowerCamelCase_) self.assertTrue(np.allclose(squeeze(lowerCamelCase_ , axis=2) , np.asarray(squeeze(lowerCamelCase_ , axis=2)))) def UpperCAmelCase__ ( self) -> Optional[Any]: UpperCamelCase = np.random.randn(3 , 4) self.assertTrue(np.allclose(expand_dims(lowerCamelCase_ , axis=1) , np.expand_dims(lowerCamelCase_ , axis=1))) @require_torch def UpperCAmelCase__ ( self) -> List[str]: UpperCamelCase = np.random.randn(3 , 4) UpperCamelCase = torch.tensor(lowerCamelCase_) self.assertTrue(np.allclose(expand_dims(lowerCamelCase_ , axis=1) , expand_dims(lowerCamelCase_ , axis=1).numpy())) @require_tf def UpperCAmelCase__ ( self) -> List[Any]: UpperCamelCase = np.random.randn(3 , 4) UpperCamelCase = tf.constant(lowerCamelCase_) self.assertTrue(np.allclose(expand_dims(lowerCamelCase_ , axis=1) , expand_dims(lowerCamelCase_ , axis=1).numpy())) @require_flax def UpperCAmelCase__ ( self) -> Tuple: UpperCamelCase = np.random.randn(3 , 4) UpperCamelCase = jnp.array(lowerCamelCase_) self.assertTrue(np.allclose(expand_dims(lowerCamelCase_ , axis=1) , np.asarray(expand_dims(lowerCamelCase_ , axis=1))))
34
"""simple docstring""" import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def __snake_case ( ): """simple docstring""" raise RuntimeError('''CUDA out of memory.''' ) class snake_case_ ( nn.Module ): """simple docstring""" def __init__( self) -> Any: super().__init__() UpperCamelCase = nn.Linear(3 , 4) UpperCamelCase = nn.BatchNormad(4) UpperCamelCase = nn.Linear(4 , 5) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Union[str, Any]: return self.lineara(self.batchnorm(self.lineara(lowerCamelCase_))) class snake_case_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self) -> List[Any]: UpperCamelCase = [] @find_executable_batch_size(starting_batch_size=1_2_8) def mock_training_loop_function(lowerCamelCase_): nonlocal batch_sizes batch_sizes.append(lowerCamelCase_) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(lowerCamelCase_ , [1_2_8, 6_4, 3_2, 1_6, 8]) def UpperCAmelCase__ ( self) -> Optional[Any]: UpperCamelCase = [] @find_executable_batch_size(starting_batch_size=1_2_8) def mock_training_loop_function(lowerCamelCase_ , lowerCamelCase_): nonlocal batch_sizes batch_sizes.append(lowerCamelCase_) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga UpperCamelCase , UpperCamelCase = mock_training_loop_function('''hello''') self.assertListEqual(lowerCamelCase_ , [1_2_8, 6_4, 3_2, 1_6, 8]) self.assertListEqual([bs, arga] , [8, '''hello''']) def UpperCAmelCase__ ( self) -> Tuple: @find_executable_batch_size(starting_batch_size=0) def mock_training_loop_function(lowerCamelCase_): pass with self.assertRaises(lowerCamelCase_) as cm: mock_training_loop_function() self.assertIn('''No executable batch size found, reached zero.''' , cm.exception.args[0]) def UpperCAmelCase__ ( self) -> List[Any]: @find_executable_batch_size(starting_batch_size=1_6) def mock_training_loop_function(lowerCamelCase_): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(lowerCamelCase_) as cm: mock_training_loop_function() self.assertIn('''No executable batch size found, reached zero.''' , cm.exception.args[0]) def UpperCAmelCase__ ( self) -> Union[str, Any]: @find_executable_batch_size(starting_batch_size=1_2_8) def mock_training_loop_function(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(lowerCamelCase_) as cm: mock_training_loop_function(1_2_8 , '''hello''' , '''world''') self.assertIn('''Batch size was passed into `f`''' , cm.exception.args[0]) self.assertIn('''`f(arg1=\'hello\', arg2=\'world\')''' , cm.exception.args[0]) def UpperCAmelCase__ ( self) -> Dict: @find_executable_batch_size(starting_batch_size=1_6) def mock_training_loop_function(lowerCamelCase_): raise ValueError('''Oops, we had an error!''') with self.assertRaises(lowerCamelCase_) as cm: mock_training_loop_function() self.assertIn('''Oops, we had an error!''' , cm.exception.args[0]) @require_cuda def UpperCAmelCase__ ( self) -> Optional[int]: UpperCamelCase = torch.cuda.memory_allocated() UpperCamelCase = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , lowerCamelCase_) UpperCamelCase = release_memory(lowerCamelCase_) self.assertEqual(torch.cuda.memory_allocated() , lowerCamelCase_)
34
1
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = {'vocab_file': 'spiece.model'} SCREAMING_SNAKE_CASE_ = { 'vocab_file': { 'albert-base-v1': 'https://huggingface.co/albert-base-v1/resolve/main/spiece.model', 'albert-large-v1': 'https://huggingface.co/albert-large-v1/resolve/main/spiece.model', 'albert-xlarge-v1': 'https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model', 'albert-xxlarge-v1': 'https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model', 'albert-base-v2': 'https://huggingface.co/albert-base-v2/resolve/main/spiece.model', 'albert-large-v2': 'https://huggingface.co/albert-large-v2/resolve/main/spiece.model', 'albert-xlarge-v2': 'https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model', 'albert-xxlarge-v2': 'https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model', } } SCREAMING_SNAKE_CASE_ = { 'albert-base-v1': 512, 'albert-large-v1': 512, 'albert-xlarge-v1': 512, 'albert-xxlarge-v1': 512, 'albert-base-v2': 512, 'albert-large-v2': 512, 'albert-xlarge-v2': 512, 'albert-xxlarge-v2': 512, } SCREAMING_SNAKE_CASE_ = '▁' class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = VOCAB_FILES_NAMES A_ = PRETRAINED_VOCAB_FILES_MAP A_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , lowerCamelCase_ , lowerCamelCase_=True , lowerCamelCase_=True , lowerCamelCase_=False , lowerCamelCase_="[CLS]" , lowerCamelCase_="[SEP]" , lowerCamelCase_="<unk>" , lowerCamelCase_="[SEP]" , lowerCamelCase_="<pad>" , lowerCamelCase_="[CLS]" , lowerCamelCase_="[MASK]" , lowerCamelCase_ = None , **lowerCamelCase_ , ) -> None: # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. UpperCamelCase = ( AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ , normalized=lowerCamelCase_) if isinstance(lowerCamelCase_ , lowerCamelCase_) else mask_token ) UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=lowerCamelCase_ , remove_space=lowerCamelCase_ , keep_accents=lowerCamelCase_ , bos_token=lowerCamelCase_ , eos_token=lowerCamelCase_ , unk_token=lowerCamelCase_ , sep_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , cls_token=lowerCamelCase_ , mask_token=lowerCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase_ , ) UpperCamelCase = do_lower_case UpperCamelCase = remove_space UpperCamelCase = keep_accents UpperCamelCase = vocab_file UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(lowerCamelCase_) @property def UpperCAmelCase__ ( self) -> Union[str, Any]: return len(self.sp_model) def UpperCAmelCase__ ( self) -> Any: UpperCamelCase = {self.convert_ids_to_tokens(lowerCamelCase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__( self) -> Union[str, Any]: UpperCamelCase = self.__dict__.copy() UpperCamelCase = None return state def __setstate__( self , lowerCamelCase_) -> Optional[Any]: UpperCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs'''): UpperCamelCase = {} UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def UpperCAmelCase__ ( self , lowerCamelCase_) -> List[str]: if self.remove_space: UpperCamelCase = ''' '''.join(inputs.strip().split()) else: UpperCamelCase = inputs UpperCamelCase = outputs.replace('''``''' , '''"''').replace('''\'\'''' , '''"''') if not self.keep_accents: UpperCamelCase = unicodedata.normalize('''NFKD''' , lowerCamelCase_) UpperCamelCase = ''''''.join([c for c in outputs if not unicodedata.combining(lowerCamelCase_)]) if self.do_lower_case: UpperCamelCase = outputs.lower() return outputs def UpperCAmelCase__ ( self , lowerCamelCase_) -> List[str]: UpperCamelCase = self.preprocess_text(lowerCamelCase_) UpperCamelCase = self.sp_model.encode(lowerCamelCase_ , out_type=lowerCamelCase_) UpperCamelCase = [] for piece in pieces: if len(lowerCamelCase_) > 1 and piece[-1] == str(''',''') and piece[-2].isdigit(): UpperCamelCase = self.sp_model.EncodeAsPieces(piece[:-1].replace(lowerCamelCase_ , '''''')) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0]) == 1: UpperCamelCase = cur_pieces[1:] else: UpperCamelCase = cur_pieces[0][1:] cur_pieces.append(piece[-1]) new_pieces.extend(lowerCamelCase_) else: new_pieces.append(lowerCamelCase_) return new_pieces def UpperCAmelCase__ ( self , lowerCamelCase_) -> Union[str, Any]: return self.sp_model.PieceToId(lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Any: return self.sp_model.IdToPiece(lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> str: UpperCamelCase = [] UpperCamelCase = '''''' UpperCamelCase = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(lowerCamelCase_) + token UpperCamelCase = True UpperCamelCase = [] else: current_sub_tokens.append(lowerCamelCase_) UpperCamelCase = False out_string += self.sp_model.decode(lowerCamelCase_) return out_string.strip() def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> List[int]: UpperCamelCase = [self.sep_token_id] UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None , lowerCamelCase_ = False) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ , token_ids_a=lowerCamelCase_ , already_has_special_tokens=lowerCamelCase_) if token_ids_a is not None: return [1] + ([0] * len(lowerCamelCase_)) + [1] + ([0] * len(lowerCamelCase_)) + [1] return [1] + ([0] * len(lowerCamelCase_)) + [1] def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> List[int]: UpperCamelCase = [self.sep_token_id] UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> Tuple[str]: if not os.path.isdir(lowerCamelCase_): logger.error(F'Vocabulary path ({save_directory}) should be a directory') return UpperCamelCase = os.path.join( lowerCamelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowerCamelCase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , lowerCamelCase_) elif not os.path.isfile(self.vocab_file): with open(lowerCamelCase_ , '''wb''') as fi: UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase_) return (out_vocab_file,)
34
"""simple docstring""" from typing import Dict from transformers import EvalPrediction, HfArgumentParser, TrainingArguments, is_torch_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, get_torch_dist_unique_port, require_torch_multi_gpu, require_torch_neuroncore, ) from transformers.training_args import ParallelMode from transformers.utils import logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) if is_torch_available(): import torch from torch import nn from torch.utils.data import Dataset from transformers import Trainer class snake_case_ ( lowerCamelCase_ ): """simple docstring""" def __init__( self , lowerCamelCase_ = 1_0_1) -> Tuple: UpperCamelCase = length def __len__( self) -> List[str]: return self.length def __getitem__( self , lowerCamelCase_) -> int: return i class snake_case_ : """simple docstring""" def __call__( self , lowerCamelCase_) -> str: return {"input_ids": torch.tensor(lowerCamelCase_), "labels": torch.tensor(lowerCamelCase_)} class snake_case_ ( nn.Module ): """simple docstring""" def __init__( self) -> List[Any]: super().__init__() # Add some (unused) params otherwise DDP will complain. UpperCamelCase = nn.Linear(1_2_0 , 8_0) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_=None) -> Any: if labels is not None: return torch.tensor(0.0 , device=input_ids.device), input_ids else: return input_ids class snake_case_ ( lowerCamelCase_ ): """simple docstring""" @require_torch_neuroncore def UpperCAmelCase__ ( self) -> Tuple: UpperCamelCase = F'--nproc_per_node=2\n --master_port={get_torch_dist_unique_port()}\n {self.test_file_dir}/test_trainer_distributed.py\n '.split() UpperCamelCase = self.get_auto_remove_tmp_dir() UpperCamelCase = F'--output_dir {output_dir}'.split() UpperCamelCase = ['''torchrun'''] + distributed_args + args execute_subprocess_async(lowerCamelCase_ , env=self.get_env()) # successful return here == success - any errors would have caused an error in the sub-call class snake_case_ ( lowerCamelCase_ ): """simple docstring""" @require_torch_multi_gpu def UpperCAmelCase__ ( self) -> Union[str, Any]: UpperCamelCase = F'--nproc_per_node={torch.cuda.device_count()}\n --master_port={get_torch_dist_unique_port()}\n {self.test_file_dir}/test_trainer_distributed.py\n '.split() UpperCamelCase = self.get_auto_remove_tmp_dir() UpperCamelCase = F'--output_dir {output_dir}'.split() UpperCamelCase = ['''torchrun'''] + distributed_args + args execute_subprocess_async(lowerCamelCase_ , env=self.get_env()) # successful return here == success - any errors would have caused an error in the sub-call if __name__ == "__main__": # The script below is meant to be run under torch.distributed, on a machine with multiple GPUs: # # PYTHONPATH="src" python -m torch.distributed.run --nproc_per_node 2 --output_dir output_dir ./tests/test_trainer_distributed.py SCREAMING_SNAKE_CASE_ = HfArgumentParser((TrainingArguments,)) SCREAMING_SNAKE_CASE_ = parser.parse_args_into_dataclasses()[0] logger.warning( f'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}, ' f'distributed training: {training_args.parallel_mode != ParallelMode.NOT_DISTRIBUTED}' ) # Essentially, what we want to verify in the distributed case is that we get all samples back, # in the right order. (this is crucial for prediction for instance) for dataset_length in [101, 40, 7]: SCREAMING_SNAKE_CASE_ = DummyDataset(dataset_length) def __snake_case ( _lowercase ): """simple docstring""" UpperCamelCase = list(range(len(_lowercase ) ) ) UpperCamelCase = p.predictions.tolist() == sequential and p.label_ids.tolist() == sequential if not success and training_args.local_rank == 0: logger.warning( '''Predictions and/or labels do not match expected results:\n - predictions: ''' f'{p.predictions.tolist()}\n - labels: {p.label_ids.tolist()}\n - expected: {sequential}' ) return {"success": success} SCREAMING_SNAKE_CASE_ = Trainer( model=DummyModel(), args=training_args, data_collator=DummyDataCollator(), eval_dataset=dataset, compute_metrics=compute_metrics, ) SCREAMING_SNAKE_CASE_ = trainer.evaluate() logger.info(metrics) if metrics["eval_success"] is not True: logger.error(metrics) exit(1) SCREAMING_SNAKE_CASE_ = trainer.predict(dataset) logger.info(p.metrics) if p.metrics["test_success"] is not True: logger.error(p.metrics) exit(1) SCREAMING_SNAKE_CASE_ = 2 SCREAMING_SNAKE_CASE_ = trainer.evaluate() logger.info(metrics) if metrics["eval_success"] is not True: logger.error(metrics) exit(1) SCREAMING_SNAKE_CASE_ = trainer.predict(dataset) logger.info(p.metrics) if p.metrics["test_success"] is not True: logger.error(p.metrics) exit(1) SCREAMING_SNAKE_CASE_ = None
34
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = { 'junnyu/roformer_chinese_small': 'https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/config.json', 'junnyu/roformer_chinese_base': 'https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/config.json', 'junnyu/roformer_chinese_char_small': ( 'https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/config.json' ), 'junnyu/roformer_chinese_char_base': ( 'https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/config.json' ), 'junnyu/roformer_small_discriminator': ( 'https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/config.json' ), 'junnyu/roformer_small_generator': ( 'https://huggingface.co/junnyu/roformer_small_generator/resolve/main/config.json' ), # See all RoFormer models at https://huggingface.co/models?filter=roformer } class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = '''roformer''' def __init__( self , lowerCamelCase_=5_0_0_0_0 , lowerCamelCase_=None , lowerCamelCase_=7_6_8 , lowerCamelCase_=1_2 , lowerCamelCase_=1_2 , lowerCamelCase_=3_0_7_2 , lowerCamelCase_="gelu" , lowerCamelCase_=0.1 , lowerCamelCase_=0.1 , lowerCamelCase_=1_5_3_6 , lowerCamelCase_=2 , lowerCamelCase_=0.02 , lowerCamelCase_=1e-12 , lowerCamelCase_=0 , lowerCamelCase_=False , lowerCamelCase_=True , **lowerCamelCase_ , ) -> Optional[Any]: super().__init__(pad_token_id=lowerCamelCase_ , **lowerCamelCase_) UpperCamelCase = vocab_size UpperCamelCase = hidden_size if embedding_size is None else embedding_size UpperCamelCase = hidden_size UpperCamelCase = num_hidden_layers UpperCamelCase = num_attention_heads UpperCamelCase = hidden_act UpperCamelCase = intermediate_size UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = max_position_embeddings UpperCamelCase = type_vocab_size UpperCamelCase = initializer_range UpperCamelCase = layer_norm_eps UpperCamelCase = rotary_value UpperCamelCase = use_cache class snake_case_ ( lowerCamelCase_ ): """simple docstring""" @property def UpperCAmelCase__ ( self) -> Mapping[str, Mapping[int, str]]: if self.task == "multiple-choice": UpperCamelCase = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: UpperCamelCase = {0: '''batch''', 1: '''sequence'''} UpperCamelCase = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ('''token_type_ids''', dynamic_axis), ])
34
"""simple docstring""" import argparse from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import BigBirdPegasusConfig, BigBirdPegasusForConditionalGeneration SCREAMING_SNAKE_CASE_ = [ # tf -> hf ('/', '.'), ('layer_', 'layers.'), ('kernel', 'weight'), ('beta', 'bias'), ('gamma', 'weight'), ('pegasus', 'model'), ] SCREAMING_SNAKE_CASE_ = [ ('.output.dense', '.fc2'), ('intermediate.LayerNorm', 'final_layer_norm'), ('intermediate.dense', 'fc1'), ] SCREAMING_SNAKE_CASE_ = ( INIT_COMMON + [ ('attention.self.LayerNorm', 'self_attn_layer_norm'), ('attention.output.dense', 'self_attn.out_proj'), ('attention.self', 'self_attn'), ('attention.encdec.LayerNorm', 'encoder_attn_layer_norm'), ('attention.encdec_output.dense', 'encoder_attn.out_proj'), ('attention.encdec', 'encoder_attn'), ('key', 'k_proj'), ('value', 'v_proj'), ('query', 'q_proj'), ('decoder.LayerNorm', 'decoder.layernorm_embedding'), ] + END_COMMON ) SCREAMING_SNAKE_CASE_ = ( INIT_COMMON + [ ('embeddings.word_embeddings', 'shared.weight'), ('embeddings.position_embeddings', 'embed_positions.weight'), ('attention.self.LayerNorm', 'self_attn_layer_norm'), ('attention.output.dense', 'self_attn.output'), ('attention.self', 'self_attn.self'), ('encoder.LayerNorm', 'encoder.layernorm_embedding'), ] + END_COMMON ) SCREAMING_SNAKE_CASE_ = [ 'encdec/key/bias', 'encdec/query/bias', 'encdec/value/bias', 'self/key/bias', 'self/query/bias', 'self/value/bias', 'encdec_output/dense/bias', 'attention/output/dense/bias', ] def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" for tf_name, hf_name in patterns: UpperCamelCase = k.replace(_lowercase ,_lowercase ) return k def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" UpperCamelCase = BigBirdPegasusConfig(**_lowercase ) UpperCamelCase = BigBirdPegasusForConditionalGeneration(_lowercase ) UpperCamelCase = torch_model.state_dict() UpperCamelCase = {} # separating decoder weights UpperCamelCase = {k: tf_weights[k] for k in tf_weights if k.startswith('''pegasus/decoder''' )} UpperCamelCase = {k: tf_weights[k] for k in tf_weights if not k.startswith('''pegasus/decoder''' )} for k, v in tqdm(decoder_weights.items() ,'''tf -> hf conversion''' ): UpperCamelCase = [k.endswith(_lowercase ) for ending in KEYS_TO_IGNORE] if any(_lowercase ): continue UpperCamelCase = DECODER_PATTERNS UpperCamelCase = rename_state_dict_key(_lowercase ,_lowercase ) if new_k not in state_dict: raise ValueError(f'could not find new key {new_k} in state dict. (converted from {k})' ) if any(True if i in k else False for i in ['''dense''', '''query''', '''key''', '''value'''] ): UpperCamelCase = v.T UpperCamelCase = torch.from_numpy(_lowercase ) assert v.shape == state_dict[new_k].shape, f'{new_k}, {k}, {v.shape}, {state_dict[new_k].shape}' for k, v in tqdm(remaining_weights.items() ,'''tf -> hf conversion''' ): UpperCamelCase = [k.endswith(_lowercase ) for ending in KEYS_TO_IGNORE] if any(_lowercase ): continue UpperCamelCase = REMAINING_PATTERNS UpperCamelCase = rename_state_dict_key(_lowercase ,_lowercase ) if new_k not in state_dict and k != "pegasus/embeddings/position_embeddings": raise ValueError(f'could not find new key {new_k} in state dict. (converted from {k})' ) if any(True if i in k else False for i in ['''dense''', '''query''', '''key''', '''value'''] ): UpperCamelCase = v.T UpperCamelCase = torch.from_numpy(_lowercase ) if k != "pegasus/embeddings/position_embeddings": assert v.shape == state_dict[new_k].shape, f'{new_k}, {k}, {v.shape}, {state_dict[new_k].shape}' UpperCamelCase = mapping['''model.embed_positions.weight'''] UpperCamelCase = mapping.pop('''model.embed_positions.weight''' ) UpperCamelCase , UpperCamelCase = torch_model.load_state_dict(_lowercase ,strict=_lowercase ) UpperCamelCase = [ k for k in missing if k not in [ '''final_logits_bias''', '''model.encoder.embed_tokens.weight''', '''model.decoder.embed_tokens.weight''', '''lm_head.weight''', ] ] assert unexpected_missing == [], f'no matches found for the following torch keys {unexpected_missing}' assert extra == [], f'no matches found for the following tf keys {extra}' return torch_model def __snake_case ( _lowercase ): """simple docstring""" UpperCamelCase = tf.train.list_variables(_lowercase ) UpperCamelCase = {} UpperCamelCase = ['''global_step'''] for name, shape in tqdm(_lowercase ,desc='''converting tf checkpoint to dict''' ): UpperCamelCase = any(pat in name for pat in ignore_name ) if skip_key: continue UpperCamelCase = tf.train.load_variable(_lowercase ,_lowercase ) UpperCamelCase = array return tf_weights def __snake_case ( _lowercase ,_lowercase ,_lowercase ): """simple docstring""" UpperCamelCase = get_tf_weights_as_numpy(_lowercase ) UpperCamelCase = convert_bigbird_pegasus(_lowercase ,_lowercase ) torch_model.save_pretrained(_lowercase ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() parser.add_argument('--tf_ckpt_path', type=str, help='passed to tf.train.list_variables') parser.add_argument('--save_dir', default=None, type=str, help='Path to the output PyTorch model.') SCREAMING_SNAKE_CASE_ = parser.parse_args() SCREAMING_SNAKE_CASE_ = {} convert_bigbird_pegasus_ckpt_to_pytorch(args.tf_ckpt_path, args.save_dir, config_update=config_update)
34
1
"""simple docstring""" import argparse import collections import os import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_table.py SCREAMING_SNAKE_CASE_ = 'src/transformers' SCREAMING_SNAKE_CASE_ = 'docs/source/en' SCREAMING_SNAKE_CASE_ = '.' def __snake_case ( _lowercase ,_lowercase ,_lowercase ): """simple docstring""" with open(_lowercase ,'''r''' ,encoding='''utf-8''' ,newline='''\n''' ) as f: UpperCamelCase = f.readlines() # Find the start prompt. UpperCamelCase = 0 while not lines[start_index].startswith(_lowercase ): start_index += 1 start_index += 1 UpperCamelCase = start_index while not lines[end_index].startswith(_lowercase ): end_index += 1 end_index -= 1 while len(lines[start_index] ) <= 1: start_index += 1 while len(lines[end_index] ) <= 1: end_index -= 1 end_index += 1 return "".join(lines[start_index:end_index] ), start_index, end_index, lines # Add here suffixes that are used to identify models, separated by | SCREAMING_SNAKE_CASE_ = 'Model|Encoder|Decoder|ForConditionalGeneration' # Regexes that match TF/Flax/PT model names. SCREAMING_SNAKE_CASE_ = re.compile(R'TF(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)') SCREAMING_SNAKE_CASE_ = re.compile(R'Flax(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)') # Will match any TF or Flax model too so need to be in an else branch afterthe two previous regexes. SCREAMING_SNAKE_CASE_ = re.compile(R'(.*)(?:Model|Encoder|Decoder|ForConditionalGeneration)') # This is to make sure the transformers module imported is the one in the repo. SCREAMING_SNAKE_CASE_ = direct_transformers_import(TRANSFORMERS_PATH) def __snake_case ( _lowercase ): """simple docstring""" UpperCamelCase = re.finditer('''.+?(?:(?<=[a-z])(?=[A-Z])|(?<=[A-Z])(?=[A-Z][a-z])|$)''' ,_lowercase ) return [m.group(0 ) for m in matches] def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" UpperCamelCase = 2 if text == '''✅''' or text == '''❌''' else len(_lowercase ) UpperCamelCase = (width - text_length) // 2 UpperCamelCase = width - text_length - left_indent return " " * left_indent + text + " " * right_indent def __snake_case ( ): """simple docstring""" UpperCamelCase = transformers_module.models.auto.configuration_auto.CONFIG_MAPPING_NAMES UpperCamelCase = { name: config_maping_names[code] for code, name in transformers_module.MODEL_NAMES_MAPPING.items() if code in config_maping_names } UpperCamelCase = {name: config.replace('''Config''' ,'''''' ) for name, config in model_name_to_config.items()} # Dictionaries flagging if each model prefix has a slow/fast tokenizer, backend in PT/TF/Flax. UpperCamelCase = collections.defaultdict(_lowercase ) UpperCamelCase = collections.defaultdict(_lowercase ) UpperCamelCase = collections.defaultdict(_lowercase ) UpperCamelCase = collections.defaultdict(_lowercase ) UpperCamelCase = collections.defaultdict(_lowercase ) # Let's lookup through all transformers object (once). for attr_name in dir(_lowercase ): UpperCamelCase = None if attr_name.endswith('''Tokenizer''' ): UpperCamelCase = slow_tokenizers UpperCamelCase = attr_name[:-9] elif attr_name.endswith('''TokenizerFast''' ): UpperCamelCase = fast_tokenizers UpperCamelCase = attr_name[:-13] elif _re_tf_models.match(_lowercase ) is not None: UpperCamelCase = tf_models UpperCamelCase = _re_tf_models.match(_lowercase ).groups()[0] elif _re_flax_models.match(_lowercase ) is not None: UpperCamelCase = flax_models UpperCamelCase = _re_flax_models.match(_lowercase ).groups()[0] elif _re_pt_models.match(_lowercase ) is not None: UpperCamelCase = pt_models UpperCamelCase = _re_pt_models.match(_lowercase ).groups()[0] if lookup_dict is not None: while len(_lowercase ) > 0: if attr_name in model_name_to_prefix.values(): UpperCamelCase = True break # Try again after removing the last word in the name UpperCamelCase = ''''''.join(camel_case_split(_lowercase )[:-1] ) # Let's build that table! UpperCamelCase = list(model_name_to_config.keys() ) model_names.sort(key=str.lower ) UpperCamelCase = ['''Model''', '''Tokenizer slow''', '''Tokenizer fast''', '''PyTorch support''', '''TensorFlow support''', '''Flax Support'''] # We'll need widths to properly display everything in the center (+2 is to leave one extra space on each side). UpperCamelCase = [len(_lowercase ) + 2 for c in columns] UpperCamelCase = max([len(_lowercase ) for name in model_names] ) + 2 # Build the table per se UpperCamelCase = '''|''' + '''|'''.join([_center_text(_lowercase ,_lowercase ) for c, w in zip(_lowercase ,_lowercase )] ) + '''|\n''' # Use ":-----:" format to center-aligned table cell texts table += "|" + "|".join([''':''' + '''-''' * (w - 2) + ''':''' for w in widths] ) + "|\n" UpperCamelCase = {True: '''✅''', False: '''❌'''} for name in model_names: UpperCamelCase = model_name_to_prefix[name] UpperCamelCase = [ name, check[slow_tokenizers[prefix]], check[fast_tokenizers[prefix]], check[pt_models[prefix]], check[tf_models[prefix]], check[flax_models[prefix]], ] table += "|" + "|".join([_center_text(_lowercase ,_lowercase ) for l, w in zip(_lowercase ,_lowercase )] ) + "|\n" return table def __snake_case ( _lowercase=False ): """simple docstring""" UpperCamelCase , UpperCamelCase , UpperCamelCase , UpperCamelCase = _find_text_in_file( filename=os.path.join(_lowercase ,'''index.md''' ) ,start_prompt='''<!--This table is updated automatically from the auto modules''' ,end_prompt='''<!-- End table-->''' ,) UpperCamelCase = get_model_table_from_auto_modules() if current_table != new_table: if overwrite: with open(os.path.join(_lowercase ,'''index.md''' ) ,'''w''' ,encoding='''utf-8''' ,newline='''\n''' ) as f: f.writelines(lines[:start_index] + [new_table] + lines[end_index:] ) else: raise ValueError( '''The model table in the `index.md` has not been updated. Run `make fix-copies` to fix this.''' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() parser.add_argument('--fix_and_overwrite', action='store_true', help='Whether to fix inconsistencies.') SCREAMING_SNAKE_CASE_ = parser.parse_args() check_model_table(args.fix_and_overwrite)
34
"""simple docstring""" from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def __snake_case ( _lowercase ): """simple docstring""" UpperCamelCase , UpperCamelCase = analyze_text(_lowercase ) UpperCamelCase = list(''' ''' + ascii_lowercase ) # what is our total sum of probabilities. UpperCamelCase = sum(single_char_strings.values() ) # one length string UpperCamelCase = 0 # for each alpha we go in our dict and if it is in it we calculate entropy for ch in my_alphas: if ch in single_char_strings: UpperCamelCase = single_char_strings[ch] UpperCamelCase = my_str / all_sum my_fir_sum += prob * math.loga(_lowercase ) # entropy formula. # print entropy print(f'{round(-1 * my_fir_sum ):.1f}' ) # two len string UpperCamelCase = sum(two_char_strings.values() ) UpperCamelCase = 0 # for each alpha (two in size) calculate entropy. for cha in my_alphas: for cha in my_alphas: UpperCamelCase = cha + cha if sequence in two_char_strings: UpperCamelCase = two_char_strings[sequence] UpperCamelCase = int(_lowercase ) / all_sum my_sec_sum += prob * math.loga(_lowercase ) # print second entropy print(f'{round(-1 * my_sec_sum ):.1f}' ) # print the difference between them print(f'{round((-1 * my_sec_sum) - (-1 * my_fir_sum) ):.1f}' ) def __snake_case ( _lowercase ): """simple docstring""" UpperCamelCase = Counter() # type: ignore UpperCamelCase = Counter() # type: ignore single_char_strings[text[-1]] += 1 # first case when we have space at start. two_char_strings[" " + text[0]] += 1 for i in range(0 ,len(_lowercase ) - 1 ): single_char_strings[text[i]] += 1 two_char_strings[text[i : i + 2]] += 1 return single_char_strings, two_char_strings def __snake_case ( ): """simple docstring""" import doctest doctest.testmod() # text = ( # "Had repulsive dashwoods suspicion sincerity but advantage now him. Remark " # "easily garret nor nay. Civil those mrs enjoy shy fat merry. You greatest " # "jointure saw horrible. He private he on be imagine suppose. Fertile " # "beloved evident through no service elderly is. Blind there if every no so " # "at. Own neglected you preferred way sincerity delivered his attempted. To " # "of message cottage windows do besides against uncivil. Delightful " # "unreserved impossible few estimating men favourable see entreaties. She " # "propriety immediate was improving. He or entrance humoured likewise " # "moderate. Much nor game son say feel. Fat make met can must form into " # "gate. Me we offending prevailed discovery. " # ) # calculate_prob(text) if __name__ == "__main__": main()
34
1
"""simple docstring""" import copy import re class snake_case_ : """simple docstring""" A_ = '''hp''' A_ = {} A_ = None @classmethod def UpperCAmelCase__ ( cls , lowerCamelCase_ , lowerCamelCase_) -> Dict: UpperCamelCase = prefix UpperCamelCase = defaults cls.build_naming_info() @staticmethod def UpperCAmelCase__ ( lowerCamelCase_ , lowerCamelCase_) -> List[str]: if len(lowerCamelCase_) == 0: return "" UpperCamelCase = None if any(char.isdigit() for char in word): raise Exception(F'Parameters should not contain numbers: \'{word}\' contains a number') if word in info["short_word"]: return info["short_word"][word] for prefix_len in range(1 , len(lowerCamelCase_) + 1): UpperCamelCase = word[:prefix_len] if prefix in info["reverse_short_word"]: continue else: UpperCamelCase = prefix break if short_word is None: # Paranoid fallback def int_to_alphabetic(lowerCamelCase_): UpperCamelCase = '''''' while integer != 0: UpperCamelCase = chr(ord('''A''') + integer % 1_0) + s integer //= 1_0 return s UpperCamelCase = 0 while True: UpperCamelCase = word + '''#''' + int_to_alphabetic(lowerCamelCase_) if sword in info["reverse_short_word"]: continue else: UpperCamelCase = sword break UpperCamelCase = short_word UpperCamelCase = word return short_word @staticmethod def UpperCAmelCase__ ( lowerCamelCase_ , lowerCamelCase_) -> str: UpperCamelCase = param_name.split('''_''') UpperCamelCase = [TrialShortNamer.shortname_for_word(lowerCamelCase_ , lowerCamelCase_) for word in words] # We try to create a separatorless short name, but if there is a collision we have to fallback # to a separated short name UpperCamelCase = ['''''', '''_'''] for separator in separators: UpperCamelCase = separator.join(lowerCamelCase_) if shortname not in info["reverse_short_param"]: UpperCamelCase = shortname UpperCamelCase = param_name return shortname return param_name @staticmethod def UpperCAmelCase__ ( lowerCamelCase_ , lowerCamelCase_) -> Optional[int]: UpperCamelCase = TrialShortNamer.shortname_for_key(lowerCamelCase_ , lowerCamelCase_) UpperCamelCase = short_name UpperCamelCase = param_name @classmethod def UpperCAmelCase__ ( cls) -> Union[str, Any]: if cls.NAMING_INFO is not None: return UpperCamelCase = { '''short_word''': {}, '''reverse_short_word''': {}, '''short_param''': {}, '''reverse_short_param''': {}, } UpperCamelCase = list(cls.DEFAULTS.keys()) for k in field_keys: cls.add_new_param_name(lowerCamelCase_ , lowerCamelCase_) UpperCamelCase = info @classmethod def UpperCAmelCase__ ( cls , lowerCamelCase_) -> List[Any]: cls.build_naming_info() assert cls.PREFIX is not None UpperCamelCase = [copy.copy(cls.PREFIX)] for k, v in params.items(): if k not in cls.DEFAULTS: raise Exception(F'You should provide a default value for the param name {k} with value {v}') if v == cls.DEFAULTS[k]: # The default value is not added to the name continue UpperCamelCase = cls.NAMING_INFO['''short_param'''][k] if isinstance(lowerCamelCase_ , lowerCamelCase_): UpperCamelCase = 1 if v else 0 UpperCamelCase = '''''' if isinstance(lowerCamelCase_ , (int, float)) else '''-''' UpperCamelCase = F'{key}{sep}{v}' name.append(lowerCamelCase_) return "_".join(lowerCamelCase_) @classmethod def UpperCAmelCase__ ( cls , lowerCamelCase_) -> Dict: UpperCamelCase = repr[len(cls.PREFIX) + 1 :] if repr == "": UpperCamelCase = [] else: UpperCamelCase = repr.split('''_''') UpperCamelCase = {} for value in values: if "-" in value: UpperCamelCase , UpperCamelCase = value.split('''-''') else: UpperCamelCase = re.sub('''[0-9.]''' , '''''' , lowerCamelCase_) UpperCamelCase = float(re.sub('''[^0-9.]''' , '''''' , lowerCamelCase_)) UpperCamelCase = cls.NAMING_INFO['''reverse_short_param'''][p_k] UpperCamelCase = p_v for k in cls.DEFAULTS: if k not in parameters: UpperCamelCase = cls.DEFAULTS[k] return parameters
34
"""simple docstring""" import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class snake_case_ ( unittest.TestCase ): """simple docstring""" def __init__( self , lowerCamelCase_ , lowerCamelCase_=1_3 , lowerCamelCase_=7 , lowerCamelCase_=True , lowerCamelCase_=True , lowerCamelCase_=True , lowerCamelCase_=True , lowerCamelCase_=9_9 , lowerCamelCase_=3_2 , lowerCamelCase_=5 , lowerCamelCase_=4 , lowerCamelCase_=3_7 , lowerCamelCase_="gelu" , lowerCamelCase_=0.1 , lowerCamelCase_=0.1 , lowerCamelCase_=5_1_2 , lowerCamelCase_=1_6 , lowerCamelCase_=2 , lowerCamelCase_=0.02 , lowerCamelCase_=4 , ) -> Any: UpperCamelCase = parent UpperCamelCase = batch_size UpperCamelCase = seq_length UpperCamelCase = is_training UpperCamelCase = use_attention_mask UpperCamelCase = use_token_type_ids UpperCamelCase = use_labels UpperCamelCase = vocab_size UpperCamelCase = hidden_size UpperCamelCase = num_hidden_layers UpperCamelCase = num_attention_heads UpperCamelCase = intermediate_size UpperCamelCase = hidden_act UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = max_position_embeddings UpperCamelCase = type_vocab_size UpperCamelCase = type_sequence_label_size UpperCamelCase = initializer_range UpperCamelCase = num_choices def UpperCAmelCase__ ( self) -> Optional[Any]: UpperCamelCase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size) UpperCamelCase = None if self.use_attention_mask: UpperCamelCase = random_attention_mask([self.batch_size, self.seq_length]) UpperCamelCase = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , tie_weights_=lowerCamelCase_ , ) return config, input_ids, attention_mask def UpperCAmelCase__ ( self) -> str: UpperCamelCase = self.prepare_config_and_inputs() UpperCamelCase , UpperCamelCase , UpperCamelCase = config_and_inputs UpperCamelCase = {'''input_ids''': input_ids, '''attention_mask''': attention_mask} return config, inputs_dict @require_flax class snake_case_ ( lowerCamelCase_ , unittest.TestCase ): """simple docstring""" A_ = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCAmelCase__ ( self) -> List[str]: UpperCamelCase = FlaxDistilBertModelTester(self) @slow def UpperCAmelCase__ ( self) -> Dict: for model_class_name in self.all_model_classes: UpperCamelCase = model_class_name.from_pretrained('''distilbert-base-uncased''') UpperCamelCase = model(np.ones((1, 1))) self.assertIsNotNone(lowerCamelCase_) @require_flax class snake_case_ ( unittest.TestCase ): """simple docstring""" @slow def UpperCAmelCase__ ( self) -> Optional[int]: UpperCamelCase = FlaxDistilBertModel.from_pretrained('''distilbert-base-uncased''') UpperCamelCase = np.array([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]]) UpperCamelCase = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]) UpperCamelCase = model(lowerCamelCase_ , attention_mask=lowerCamelCase_)[0] UpperCamelCase = (1, 1_1, 7_6_8) self.assertEqual(output.shape , lowerCamelCase_) UpperCamelCase = np.array([[[-0.1639, 0.3299, 0.1648], [-0.1746, 0.3289, 0.1710], [-0.1884, 0.3357, 0.1810]]]) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , lowerCamelCase_ , atol=1e-4))
34
1
"""simple docstring""" # This model implementation is heavily inspired by https://github.com/haofanwang/ControlNet-for-Diffusers/ import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, ControlNetModel, DDIMScheduler, StableDiffusionControlNetImgaImgPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion_controlnet import MultiControlNetModel from diffusers.utils import floats_tensor, load_image, load_numpy, randn_tensor, slow, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import ( PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin, ) enable_full_determinism() class snake_case_ ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): """simple docstring""" A_ = StableDiffusionControlNetImgaImgPipeline A_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} A_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS A_ = IMAGE_TO_IMAGE_IMAGE_PARAMS.union({'''control_image'''} ) A_ = IMAGE_TO_IMAGE_IMAGE_PARAMS def UpperCAmelCase__ ( self) -> Tuple: torch.manual_seed(0) UpperCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=3_2 , ) torch.manual_seed(0) UpperCamelCase = ControlNetModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , in_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , cross_attention_dim=3_2 , conditioning_embedding_out_channels=(1_6, 3_2) , ) torch.manual_seed(0) UpperCamelCase = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=lowerCamelCase_ , set_alpha_to_one=lowerCamelCase_ , ) torch.manual_seed(0) UpperCamelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) torch.manual_seed(0) UpperCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) UpperCamelCase = CLIPTextModel(lowerCamelCase_) UpperCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') UpperCamelCase = { '''unet''': unet, '''controlnet''': controlnet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_=0) -> Union[str, Any]: if str(lowerCamelCase_).startswith('''mps'''): UpperCamelCase = torch.manual_seed(lowerCamelCase_) else: UpperCamelCase = torch.Generator(device=lowerCamelCase_).manual_seed(lowerCamelCase_) UpperCamelCase = 2 UpperCamelCase = randn_tensor( (1, 3, 3_2 * controlnet_embedder_scale_factor, 3_2 * controlnet_embedder_scale_factor) , generator=lowerCamelCase_ , device=torch.device(lowerCamelCase_) , ) UpperCamelCase = floats_tensor(control_image.shape , rng=random.Random(lowerCamelCase_)).to(lowerCamelCase_) UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1)[0] UpperCamelCase = Image.fromarray(np.uinta(lowerCamelCase_)).convert('''RGB''').resize((6_4, 6_4)) UpperCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', '''image''': image, '''control_image''': control_image, } return inputs def UpperCAmelCase__ ( self) -> Optional[int]: return self._test_attention_slicing_forward_pass(expected_max_diff=2e-3) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCAmelCase__ ( self) -> Optional[Any]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2e-3) def UpperCAmelCase__ ( self) -> Dict: self._test_inference_batch_single_identical(expected_max_diff=2e-3) class snake_case_ ( lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): """simple docstring""" A_ = StableDiffusionControlNetImgaImgPipeline A_ = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - {'''height''', '''width'''} A_ = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS A_ = frozenset([] ) # TO_DO: add image_params once refactored VaeImageProcessor.preprocess def UpperCAmelCase__ ( self) -> Dict: torch.manual_seed(0) UpperCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=3_2 , ) torch.manual_seed(0) def init_weights(lowerCamelCase_): if isinstance(lowerCamelCase_ , torch.nn.Convad): torch.nn.init.normal(m.weight) m.bias.data.fill_(1.0) UpperCamelCase = ControlNetModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , in_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , cross_attention_dim=3_2 , conditioning_embedding_out_channels=(1_6, 3_2) , ) controlneta.controlnet_down_blocks.apply(lowerCamelCase_) torch.manual_seed(0) UpperCamelCase = ControlNetModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , in_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , cross_attention_dim=3_2 , conditioning_embedding_out_channels=(1_6, 3_2) , ) controlneta.controlnet_down_blocks.apply(lowerCamelCase_) torch.manual_seed(0) UpperCamelCase = DDIMScheduler( beta_start=0.0_0085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , clip_sample=lowerCamelCase_ , set_alpha_to_one=lowerCamelCase_ , ) torch.manual_seed(0) UpperCamelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , ) torch.manual_seed(0) UpperCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) UpperCamelCase = CLIPTextModel(lowerCamelCase_) UpperCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') UpperCamelCase = MultiControlNetModel([controlneta, controlneta]) UpperCamelCase = { '''unet''': unet, '''controlnet''': controlnet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_=0) -> List[str]: if str(lowerCamelCase_).startswith('''mps'''): UpperCamelCase = torch.manual_seed(lowerCamelCase_) else: UpperCamelCase = torch.Generator(device=lowerCamelCase_).manual_seed(lowerCamelCase_) UpperCamelCase = 2 UpperCamelCase = [ randn_tensor( (1, 3, 3_2 * controlnet_embedder_scale_factor, 3_2 * controlnet_embedder_scale_factor) , generator=lowerCamelCase_ , device=torch.device(lowerCamelCase_) , ), randn_tensor( (1, 3, 3_2 * controlnet_embedder_scale_factor, 3_2 * controlnet_embedder_scale_factor) , generator=lowerCamelCase_ , device=torch.device(lowerCamelCase_) , ), ] UpperCamelCase = floats_tensor(control_image[0].shape , rng=random.Random(lowerCamelCase_)).to(lowerCamelCase_) UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1)[0] UpperCamelCase = Image.fromarray(np.uinta(lowerCamelCase_)).convert('''RGB''').resize((6_4, 6_4)) UpperCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', '''image''': image, '''control_image''': control_image, } return inputs def UpperCAmelCase__ ( self) -> List[Any]: UpperCamelCase = self.get_dummy_components() UpperCamelCase = self.pipeline_class(**lowerCamelCase_) pipe.to(lowerCamelCase_) UpperCamelCase = 10.0 UpperCamelCase = 4 UpperCamelCase = self.get_dummy_inputs(lowerCamelCase_) UpperCamelCase = steps UpperCamelCase = scale UpperCamelCase = pipe(**lowerCamelCase_)[0] UpperCamelCase = self.get_dummy_inputs(lowerCamelCase_) UpperCamelCase = steps UpperCamelCase = scale UpperCamelCase = pipe(**lowerCamelCase_ , control_guidance_start=0.1 , control_guidance_end=0.2)[0] UpperCamelCase = self.get_dummy_inputs(lowerCamelCase_) UpperCamelCase = steps UpperCamelCase = scale UpperCamelCase = pipe(**lowerCamelCase_ , control_guidance_start=[0.1, 0.3] , control_guidance_end=[0.2, 0.7])[0] UpperCamelCase = self.get_dummy_inputs(lowerCamelCase_) UpperCamelCase = steps UpperCamelCase = scale UpperCamelCase = pipe(**lowerCamelCase_ , control_guidance_start=0.4 , control_guidance_end=[0.5, 0.8])[0] # make sure that all outputs are different assert np.sum(np.abs(output_a - output_a)) > 1e-3 assert np.sum(np.abs(output_a - output_a)) > 1e-3 assert np.sum(np.abs(output_a - output_a)) > 1e-3 def UpperCAmelCase__ ( self) -> Optional[Any]: return self._test_attention_slicing_forward_pass(expected_max_diff=2e-3) @unittest.skipIf( torch_device != '''cuda''' or not is_xformers_available() , reason='''XFormers attention is only available with CUDA and `xformers` installed''' , ) def UpperCAmelCase__ ( self) -> Optional[Any]: self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=2e-3) def UpperCAmelCase__ ( self) -> int: self._test_inference_batch_single_identical(expected_max_diff=2e-3) def UpperCAmelCase__ ( self) -> Union[str, Any]: UpperCamelCase = self.get_dummy_components() UpperCamelCase = self.pipeline_class(**lowerCamelCase_) pipe.to(lowerCamelCase_) pipe.set_progress_bar_config(disable=lowerCamelCase_) with tempfile.TemporaryDirectory() as tmpdir: try: # save_pretrained is not implemented for Multi-ControlNet pipe.save_pretrained(lowerCamelCase_) except NotImplementedError: pass @slow @require_torch_gpu class snake_case_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self) -> Union[str, Any]: super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self) -> Optional[Any]: UpperCamelCase = ControlNetModel.from_pretrained('''lllyasviel/sd-controlnet-canny''') UpperCamelCase = StableDiffusionControlNetImgaImgPipeline.from_pretrained( '''runwayml/stable-diffusion-v1-5''' , safety_checker=lowerCamelCase_ , controlnet=lowerCamelCase_) pipe.enable_model_cpu_offload() pipe.set_progress_bar_config(disable=lowerCamelCase_) UpperCamelCase = torch.Generator(device='''cpu''').manual_seed(0) UpperCamelCase = '''evil space-punk bird''' UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/bird_canny.png''').resize((5_1_2, 5_1_2)) UpperCamelCase = load_image( '''https://huggingface.co/lllyasviel/sd-controlnet-canny/resolve/main/images/bird.png''').resize((5_1_2, 5_1_2)) UpperCamelCase = pipe( lowerCamelCase_ , lowerCamelCase_ , control_image=lowerCamelCase_ , generator=lowerCamelCase_ , output_type='''np''' , num_inference_steps=5_0 , strength=0.6 , ) UpperCamelCase = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) UpperCamelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd_controlnet/img2img.npy''') assert np.abs(expected_image - image).max() < 9e-2
34
"""simple docstring""" from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) @add_end_docstrings(lowerCamelCase_ ) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" def __init__( self , **lowerCamelCase_) -> Tuple: super().__init__(**lowerCamelCase_) requires_backends(self , '''vision''') self.check_model_type( TF_MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING if self.framework == '''tf''' else MODEL_FOR_ZERO_SHOT_IMAGE_CLASSIFICATION_MAPPING) def __call__( self , lowerCamelCase_ , **lowerCamelCase_) -> Optional[int]: return super().__call__(lowerCamelCase_ , **lowerCamelCase_) def UpperCAmelCase__ ( self , **lowerCamelCase_) -> Any: UpperCamelCase = {} if "candidate_labels" in kwargs: UpperCamelCase = kwargs['''candidate_labels'''] if "hypothesis_template" in kwargs: UpperCamelCase = kwargs['''hypothesis_template'''] return preprocess_params, {}, {} def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_=None , lowerCamelCase_="This is a photo of {}.") -> Union[str, Any]: UpperCamelCase = load_image(lowerCamelCase_) UpperCamelCase = self.image_processor(images=[image] , return_tensors=self.framework) UpperCamelCase = candidate_labels UpperCamelCase = [hypothesis_template.format(lowerCamelCase_) for x in candidate_labels] UpperCamelCase = self.tokenizer(lowerCamelCase_ , return_tensors=self.framework , padding=lowerCamelCase_) UpperCamelCase = [text_inputs] return inputs def UpperCAmelCase__ ( self , lowerCamelCase_) -> Any: UpperCamelCase = model_inputs.pop('''candidate_labels''') UpperCamelCase = model_inputs.pop('''text_inputs''') if isinstance(text_inputs[0] , lowerCamelCase_): UpperCamelCase = text_inputs[0] else: # Batching case. UpperCamelCase = text_inputs[0][0] UpperCamelCase = self.model(**lowerCamelCase_ , **lowerCamelCase_) UpperCamelCase = { '''candidate_labels''': candidate_labels, '''logits''': outputs.logits_per_image, } return model_outputs def UpperCAmelCase__ ( self , lowerCamelCase_) -> Any: UpperCamelCase = model_outputs.pop('''candidate_labels''') UpperCamelCase = model_outputs['''logits'''][0] if self.framework == "pt": UpperCamelCase = logits.softmax(dim=-1).squeeze(-1) UpperCamelCase = probs.tolist() if not isinstance(lowerCamelCase_ , lowerCamelCase_): UpperCamelCase = [scores] elif self.framework == "tf": UpperCamelCase = stable_softmax(lowerCamelCase_ , axis=-1) UpperCamelCase = probs.numpy().tolist() else: raise ValueError(F'Unsupported framework: {self.framework}') UpperCamelCase = [ {'''score''': score, '''label''': candidate_label} for score, candidate_label in sorted(zip(lowerCamelCase_ , lowerCamelCase_) , key=lambda lowerCamelCase_: -x[0]) ] return result
34
1
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = ['''pixel_values'''] def __init__( self , lowerCamelCase_ = True , lowerCamelCase_ = None , lowerCamelCase_ = PIL.Image.BICUBIC , lowerCamelCase_ = True , lowerCamelCase_ = None , lowerCamelCase_ = 1 / 2_5_5 , lowerCamelCase_ = True , lowerCamelCase_ = True , lowerCamelCase_ = None , lowerCamelCase_ = None , **lowerCamelCase_ , ) -> None: super().__init__(**lowerCamelCase_) UpperCamelCase = size if size is not None else {'''height''': 2_5_6, '''width''': 2_5_6} UpperCamelCase = get_size_dict(lowerCamelCase_) UpperCamelCase = crop_size if crop_size is not None else {'''height''': 2_2_4, '''width''': 2_2_4} UpperCamelCase = get_size_dict(lowerCamelCase_ , param_name='''crop_size''') UpperCamelCase = do_resize UpperCamelCase = size UpperCamelCase = resample UpperCamelCase = do_center_crop UpperCamelCase = crop_size UpperCamelCase = do_rescale UpperCamelCase = rescale_factor UpperCamelCase = do_normalize UpperCamelCase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN UpperCamelCase = image_std if image_std is not None else IMAGENET_STANDARD_STD def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = PIL.Image.BICUBIC , lowerCamelCase_ = None , **lowerCamelCase_ , ) -> np.ndarray: UpperCamelCase = get_size_dict(lowerCamelCase_) if "height" not in size or "width" not in size: raise ValueError(F'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}') return resize( lowerCamelCase_ , size=(size['''height'''], size['''width''']) , resample=lowerCamelCase_ , data_format=lowerCamelCase_ , **lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = None , **lowerCamelCase_ , ) -> np.ndarray: UpperCamelCase = get_size_dict(lowerCamelCase_) if "height" not in size or "width" not in size: raise ValueError(F'The size dictionary must have keys \'height\' and \'width\'. Got {size.keys()}') return center_crop(lowerCamelCase_ , size=(size['''height'''], size['''width''']) , data_format=lowerCamelCase_ , **lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = None , **lowerCamelCase_ , ) -> Tuple: return rescale(lowerCamelCase_ , scale=lowerCamelCase_ , data_format=lowerCamelCase_ , **lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = None , **lowerCamelCase_ , ) -> np.ndarray: return normalize(lowerCamelCase_ , mean=lowerCamelCase_ , std=lowerCamelCase_ , data_format=lowerCamelCase_ , **lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None , lowerCamelCase_ = None , lowerCamelCase_=None , lowerCamelCase_ = None , lowerCamelCase_ = None , lowerCamelCase_ = None , lowerCamelCase_ = None , lowerCamelCase_ = None , lowerCamelCase_ = None , lowerCamelCase_ = None , lowerCamelCase_ = None , lowerCamelCase_ = ChannelDimension.FIRST , **lowerCamelCase_ , ) -> PIL.Image.Image: UpperCamelCase = do_resize if do_resize is not None else self.do_resize UpperCamelCase = resample if resample is not None else self.resample UpperCamelCase = do_center_crop if do_center_crop is not None else self.do_center_crop UpperCamelCase = do_rescale if do_rescale is not None else self.do_rescale UpperCamelCase = rescale_factor if rescale_factor is not None else self.rescale_factor UpperCamelCase = do_normalize if do_normalize is not None else self.do_normalize UpperCamelCase = image_mean if image_mean is not None else self.image_mean UpperCamelCase = image_std if image_std is not None else self.image_std UpperCamelCase = size if size is not None else self.size UpperCamelCase = get_size_dict(lowerCamelCase_) UpperCamelCase = crop_size if crop_size is not None else self.crop_size UpperCamelCase = get_size_dict(lowerCamelCase_ , param_name='''crop_size''') UpperCamelCase = make_list_of_images(lowerCamelCase_) if not valid_images(lowerCamelCase_): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''') if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''') if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''') if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''') if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''') # All transformations expect numpy arrays. UpperCamelCase = [to_numpy_array(lowerCamelCase_) for image in images] if do_resize: UpperCamelCase = [self.resize(image=lowerCamelCase_ , size=lowerCamelCase_ , resample=lowerCamelCase_) for image in images] if do_center_crop: UpperCamelCase = [self.center_crop(image=lowerCamelCase_ , size=lowerCamelCase_) for image in images] if do_rescale: UpperCamelCase = [self.rescale(image=lowerCamelCase_ , scale=lowerCamelCase_) for image in images] if do_normalize: UpperCamelCase = [self.normalize(image=lowerCamelCase_ , mean=lowerCamelCase_ , std=lowerCamelCase_) for image in images] UpperCamelCase = [to_channel_dimension_format(lowerCamelCase_ , lowerCamelCase_) for image in images] UpperCamelCase = {'''pixel_values''': images} return BatchFeature(data=lowerCamelCase_ , tensor_type=lowerCamelCase_)
34
"""simple docstring""" import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class snake_case_ ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , unittest.TestCase ): """simple docstring""" A_ = StableDiffusionInpaintPipeline A_ = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS A_ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS A_ = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess A_ = frozenset([] ) def UpperCAmelCase__ ( self) -> List[Any]: torch.manual_seed(0) UpperCamelCase = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=9 , out_channels=4 , down_block_types=('''DownBlock2D''', '''CrossAttnDownBlock2D''') , up_block_types=('''CrossAttnUpBlock2D''', '''UpBlock2D''') , cross_attention_dim=3_2 , attention_head_dim=(2, 4) , use_linear_projection=lowerCamelCase_ , ) UpperCamelCase = PNDMScheduler(skip_prk_steps=lowerCamelCase_) torch.manual_seed(0) UpperCamelCase = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=['''DownEncoderBlock2D''', '''DownEncoderBlock2D'''] , up_block_types=['''UpDecoderBlock2D''', '''UpDecoderBlock2D'''] , latent_channels=4 , sample_size=1_2_8 , ) torch.manual_seed(0) UpperCamelCase = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , hidden_act='''gelu''' , projection_dim=5_1_2 , ) UpperCamelCase = CLIPTextModel(lowerCamelCase_) UpperCamelCase = CLIPTokenizer.from_pretrained('''hf-internal-testing/tiny-random-clip''') UpperCamelCase = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, '''safety_checker''': None, '''feature_extractor''': None, } return components def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_=0) -> Dict: # TODO: use tensor inputs instead of PIL, this is here just to leave the old expected_slices untouched UpperCamelCase = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(lowerCamelCase_)).to(lowerCamelCase_) UpperCamelCase = image.cpu().permute(0 , 2 , 3 , 1)[0] UpperCamelCase = Image.fromarray(np.uinta(lowerCamelCase_)).convert('''RGB''').resize((6_4, 6_4)) UpperCamelCase = Image.fromarray(np.uinta(image + 4)).convert('''RGB''').resize((6_4, 6_4)) if str(lowerCamelCase_).startswith('''mps'''): UpperCamelCase = torch.manual_seed(lowerCamelCase_) else: UpperCamelCase = torch.Generator(device=lowerCamelCase_).manual_seed(lowerCamelCase_) UpperCamelCase = { '''prompt''': '''A painting of a squirrel eating a burger''', '''image''': init_image, '''mask_image''': mask_image, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''numpy''', } return inputs def UpperCAmelCase__ ( self) -> Optional[Any]: UpperCamelCase = '''cpu''' # ensure determinism for the device-dependent torch.Generator UpperCamelCase = self.get_dummy_components() UpperCamelCase = StableDiffusionInpaintPipeline(**lowerCamelCase_) UpperCamelCase = sd_pipe.to(lowerCamelCase_) sd_pipe.set_progress_bar_config(disable=lowerCamelCase_) UpperCamelCase = self.get_dummy_inputs(lowerCamelCase_) UpperCamelCase = sd_pipe(**lowerCamelCase_).images UpperCamelCase = image[0, -3:, -3:, -1] assert image.shape == (1, 6_4, 6_4, 3) UpperCamelCase = np.array([0.4727, 0.5735, 0.3941, 0.5446, 0.5926, 0.4394, 0.5062, 0.4654, 0.4476]) assert np.abs(image_slice.flatten() - expected_slice).max() < 1e-2 def UpperCAmelCase__ ( self) -> Union[str, Any]: super().test_inference_batch_single_identical(expected_max_diff=3e-3) @slow @require_torch_gpu class snake_case_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self) -> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__ ( self) -> List[Any]: UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''') UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''') UpperCamelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench.npy''') UpperCamelCase = '''stabilityai/stable-diffusion-2-inpainting''' UpperCamelCase = StableDiffusionInpaintPipeline.from_pretrained(lowerCamelCase_ , safety_checker=lowerCamelCase_) pipe.to(lowerCamelCase_) pipe.set_progress_bar_config(disable=lowerCamelCase_) pipe.enable_attention_slicing() UpperCamelCase = '''Face of a yellow cat, high resolution, sitting on a park bench''' UpperCamelCase = torch.manual_seed(0) UpperCamelCase = pipe( prompt=lowerCamelCase_ , image=lowerCamelCase_ , mask_image=lowerCamelCase_ , generator=lowerCamelCase_ , output_type='''np''' , ) UpperCamelCase = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image).max() < 9e-3 def UpperCAmelCase__ ( self) -> Optional[Any]: UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''') UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''') UpperCamelCase = load_numpy( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint''' '''/yellow_cat_sitting_on_a_park_bench_fp16.npy''') UpperCamelCase = '''stabilityai/stable-diffusion-2-inpainting''' UpperCamelCase = StableDiffusionInpaintPipeline.from_pretrained( lowerCamelCase_ , torch_dtype=torch.floataa , safety_checker=lowerCamelCase_ , ) pipe.to(lowerCamelCase_) pipe.set_progress_bar_config(disable=lowerCamelCase_) pipe.enable_attention_slicing() UpperCamelCase = '''Face of a yellow cat, high resolution, sitting on a park bench''' UpperCamelCase = torch.manual_seed(0) UpperCamelCase = pipe( prompt=lowerCamelCase_ , image=lowerCamelCase_ , mask_image=lowerCamelCase_ , generator=lowerCamelCase_ , output_type='''np''' , ) UpperCamelCase = output.images[0] assert image.shape == (5_1_2, 5_1_2, 3) assert np.abs(expected_image - image).max() < 5e-1 def UpperCAmelCase__ ( self) -> List[str]: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main''' '''/sd2-inpaint/init_image.png''') UpperCamelCase = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png''') UpperCamelCase = '''stabilityai/stable-diffusion-2-inpainting''' UpperCamelCase = PNDMScheduler.from_pretrained(lowerCamelCase_ , subfolder='''scheduler''') UpperCamelCase = StableDiffusionInpaintPipeline.from_pretrained( lowerCamelCase_ , safety_checker=lowerCamelCase_ , scheduler=lowerCamelCase_ , torch_dtype=torch.floataa , ) pipe.to(lowerCamelCase_) pipe.set_progress_bar_config(disable=lowerCamelCase_) pipe.enable_attention_slicing(1) pipe.enable_sequential_cpu_offload() UpperCamelCase = '''Face of a yellow cat, high resolution, sitting on a park bench''' UpperCamelCase = torch.manual_seed(0) UpperCamelCase = pipe( prompt=lowerCamelCase_ , image=lowerCamelCase_ , mask_image=lowerCamelCase_ , generator=lowerCamelCase_ , num_inference_steps=2 , output_type='''np''' , ) UpperCamelCase = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 1_0**9
34
1
"""simple docstring""" from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=lowerCamelCase_ ) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = field(default='''summarization''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) A_ = Features({'''text''': Value('''string''' )} ) A_ = Features({'''summary''': Value('''string''' )} ) A_ = "text" A_ = "summary" @property def UpperCAmelCase__ ( self) -> Dict[str, str]: return {self.text_column: "text", self.summary_column: "summary"}
34
"""simple docstring""" import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('3.8'): import importlib_metadata else: import importlib.metadata as importlib_metadata def __snake_case ( _lowercase ,_lowercase=False ): """simple docstring""" try: UpperCamelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. UpperCamelCase = default else: # KEY is set, convert it to True or False. try: UpperCamelCase = strtobool(_lowercase ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f'If set, {key} must be yes or no.' ) return _value SCREAMING_SNAKE_CASE_ = parse_flag_from_env('RUN_SLOW', default=False) SCREAMING_SNAKE_CASE_ = parse_flag_from_env('RUN_REMOTE', default=False) SCREAMING_SNAKE_CASE_ = parse_flag_from_env('RUN_LOCAL', default=True) SCREAMING_SNAKE_CASE_ = parse_flag_from_env('RUN_PACKAGED', default=True) # Compression SCREAMING_SNAKE_CASE_ = pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='test requires lz4') SCREAMING_SNAKE_CASE_ = pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='test requires py7zr') SCREAMING_SNAKE_CASE_ = pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='test requires zstandard') # Audio SCREAMING_SNAKE_CASE_ = pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('soundfile') is None or version.parse(importlib_metadata.version('soundfile')) < version.parse('0.12.0'), reason='test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ', ) # Beam SCREAMING_SNAKE_CASE_ = pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('0.3.2'), reason='test requires apache-beam and a compatible dill version', ) # Dill-cloudpickle compatibility SCREAMING_SNAKE_CASE_ = pytest.mark.skipif( config.DILL_VERSION <= version.parse('0.3.2'), reason='test requires dill>0.3.2 for cloudpickle compatibility', ) # Windows SCREAMING_SNAKE_CASE_ = pytest.mark.skipif( sys.platform == 'win32', reason='test should not be run on Windows', ) def __snake_case ( _lowercase ): """simple docstring""" try: import faiss # noqa except ImportError: UpperCamelCase = unittest.skip('''test requires faiss''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" try: import regex # noqa except ImportError: UpperCamelCase = unittest.skip('''test requires regex''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" try: import elasticsearch # noqa except ImportError: UpperCamelCase = unittest.skip('''test requires elasticsearch''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" try: import sqlalchemy # noqa except ImportError: UpperCamelCase = unittest.skip('''test requires sqlalchemy''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" if not config.TORCH_AVAILABLE: UpperCamelCase = unittest.skip('''test requires PyTorch''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" if not config.TF_AVAILABLE: UpperCamelCase = unittest.skip('''test requires TensorFlow''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" if not config.JAX_AVAILABLE: UpperCamelCase = unittest.skip('''test requires JAX''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" if not config.PIL_AVAILABLE: UpperCamelCase = unittest.skip('''test requires Pillow''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" try: import transformers # noqa F401 except ImportError: return unittest.skip('''test requires transformers''' )(_lowercase ) else: return test_case def __snake_case ( _lowercase ): """simple docstring""" try: import tiktoken # noqa F401 except ImportError: return unittest.skip('''test requires tiktoken''' )(_lowercase ) else: return test_case def __snake_case ( _lowercase ): """simple docstring""" try: import spacy # noqa F401 except ImportError: return unittest.skip('''test requires spacy''' )(_lowercase ) else: return test_case def __snake_case ( _lowercase ): """simple docstring""" def _require_spacy_model(_lowercase ): try: import spacy # noqa F401 spacy.load(_lowercase ) except ImportError: return unittest.skip('''test requires spacy''' )(_lowercase ) except OSError: return unittest.skip('''test requires spacy model \'{}\''''.format(_lowercase ) )(_lowercase ) else: return test_case return _require_spacy_model def __snake_case ( _lowercase ): """simple docstring""" try: import pyspark # noqa F401 except ImportError: return unittest.skip('''test requires pyspark''' )(_lowercase ) else: return test_case def __snake_case ( _lowercase ): """simple docstring""" try: import joblibspark # noqa F401 except ImportError: return unittest.skip('''test requires joblibspark''' )(_lowercase ) else: return test_case def __snake_case ( _lowercase ): """simple docstring""" if not _run_slow_tests or _run_slow_tests == 0: UpperCamelCase = unittest.skip('''test is slow''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" if not _run_local_tests or _run_local_tests == 0: UpperCamelCase = unittest.skip('''test is local''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" if not _run_packaged_tests or _run_packaged_tests == 0: UpperCamelCase = unittest.skip('''test is packaged''' )(_lowercase ) return test_case def __snake_case ( _lowercase ): """simple docstring""" if not _run_remote_tests or _run_remote_tests == 0: UpperCamelCase = unittest.skip('''test requires remote''' )(_lowercase ) return test_case def __snake_case ( *_lowercase ): """simple docstring""" def decorate(cls ): for name, fn in cls.__dict__.items(): if callable(_lowercase ) and name.startswith('''test''' ): for decorator in decorators: UpperCamelCase = decorator(_lowercase ) setattr(cls ,_lowercase ,_lowercase ) return cls return decorate class snake_case_ ( lowerCamelCase_ ): """simple docstring""" pass class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = 0 A_ = 1 A_ = 2 @contextmanager def __snake_case ( _lowercase=OfflineSimulationMode.CONNECTION_FAILS ,_lowercase=1e-16 ): """simple docstring""" UpperCamelCase = requests.Session().request def timeout_request(_lowercase ,_lowercase ,_lowercase ,**_lowercase ): # Change the url to an invalid url so that the connection hangs UpperCamelCase = '''https://10.255.255.1''' if kwargs.get('''timeout''' ) is None: raise RequestWouldHangIndefinitelyError( f'Tried a call to {url} in offline mode with no timeout set. Please set a timeout.' ) UpperCamelCase = timeout try: return online_request(_lowercase ,_lowercase ,**_lowercase ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier UpperCamelCase = url UpperCamelCase = e.args[0] UpperCamelCase = (max_retry_error.args[0].replace('''10.255.255.1''' ,f'OfflineMock[{url}]' ),) UpperCamelCase = (max_retry_error,) raise def raise_connection_error(_lowercase ,_lowercase ,**_lowercase ): raise requests.ConnectionError('''Offline mode is enabled.''' ,request=_lowercase ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch('''requests.Session.send''' ,_lowercase ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch('''requests.Session.request''' ,_lowercase ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch('''datasets.config.HF_DATASETS_OFFLINE''' ,_lowercase ): yield else: raise ValueError('''Please use a value from the OfflineSimulationMode enum.''' ) @contextmanager def __snake_case ( *_lowercase ,**_lowercase ): """simple docstring""" UpperCamelCase = str(Path().resolve() ) with tempfile.TemporaryDirectory(*_lowercase ,**_lowercase ) as tmp_dir: try: os.chdir(_lowercase ) yield finally: os.chdir(_lowercase ) @contextmanager def __snake_case ( ): """simple docstring""" import gc gc.collect() UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def __snake_case ( ): """simple docstring""" import gc gc.collect() UpperCamelCase = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" return deepcopy(_lowercase ).integers(0 ,100 ,10 ).tolist() == deepcopy(_lowercase ).integers(0 ,100 ,10 ).tolist() def __snake_case ( _lowercase ): """simple docstring""" import decorator from requests.exceptions import HTTPError def _wrapper(_lowercase ,*_lowercase ,**_lowercase ): try: return func(*_lowercase ,**_lowercase ) except HTTPError as err: if str(_lowercase ).startswith('''500''' ) or str(_lowercase ).startswith('''502''' ): pytest.xfail(str(_lowercase ) ) raise err return decorator.decorator(_wrapper ,_lowercase ) class snake_case_ : """simple docstring""" def __init__( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_) -> Dict: UpperCamelCase = returncode UpperCamelCase = stdout UpperCamelCase = stderr async def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" while True: UpperCamelCase = await stream.readline() if line: callback(_lowercase ) else: break async def __snake_case ( _lowercase ,_lowercase=None ,_lowercase=None ,_lowercase=None ,_lowercase=False ,_lowercase=False ): """simple docstring""" if echo: print('''\nRunning: ''' ,''' '''.join(_lowercase ) ) UpperCamelCase = await asyncio.create_subprocess_exec( cmd[0] ,*cmd[1:] ,stdin=_lowercase ,stdout=asyncio.subprocess.PIPE ,stderr=asyncio.subprocess.PIPE ,env=_lowercase ,) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) UpperCamelCase = [] UpperCamelCase = [] def tee(_lowercase ,_lowercase ,_lowercase ,_lowercase="" ): UpperCamelCase = line.decode('''utf-8''' ).rstrip() sink.append(_lowercase ) if not quiet: print(_lowercase ,_lowercase ,file=_lowercase ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout ,lambda _lowercase : tee(_lowercase ,_lowercase ,sys.stdout ,label='''stdout:''' ) ), _read_stream(p.stderr ,lambda _lowercase : tee(_lowercase ,_lowercase ,sys.stderr ,label='''stderr:''' ) ), ] ,timeout=_lowercase ,) return _RunOutput(await p.wait() ,_lowercase ,_lowercase ) def __snake_case ( _lowercase ,_lowercase=None ,_lowercase=None ,_lowercase=180 ,_lowercase=False ,_lowercase=True ): """simple docstring""" UpperCamelCase = asyncio.get_event_loop() UpperCamelCase = loop.run_until_complete( _stream_subprocess(_lowercase ,env=_lowercase ,stdin=_lowercase ,timeout=_lowercase ,quiet=_lowercase ,echo=_lowercase ) ) UpperCamelCase = ''' '''.join(_lowercase ) if result.returncode > 0: UpperCamelCase = '''\n'''.join(result.stderr ) raise RuntimeError( f'\'{cmd_str}\' failed with returncode {result.returncode}\n\n' f'The combined stderr from workers follows:\n{stderr}' ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(f'\'{cmd_str}\' produced no output.' ) return result def __snake_case ( ): """simple docstring""" UpperCamelCase = os.environ.get('''PYTEST_XDIST_WORKER''' ,'''gw0''' ) UpperCamelCase = re.sub(r'''^gw''' ,'''''' ,_lowercase ,0 ,re.M ) return int(_lowercase ) def __snake_case ( ): """simple docstring""" UpperCamelCase = 2_9500 UpperCamelCase = pytest_xdist_worker_id() return port + uniq_delta
34
1
"""simple docstring""" # HF Trainer benchmarking tool # # This tool can be used to run and compare multiple dimensions of the HF Trainers args. # # It then prints a report once in github format with all the information that needs to be shared # with others and second time in a console-friendly format, so it's easier to use for tuning things up. # # The main idea is: # # ./trainer-benchmark.py --base-cmd '<cmd args that don't change>' \ # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' \ # --target-metric-key train_samples_per_second # # The variations can be any command line argument that you want to compare and not just dtype as in # the example. # # --variations allows you to compare variations in multiple dimensions. # # as the first dimention has 2 options and the second 3 in our example, this will run the trainer 6 # times adding one of: # # 1. --tf32 0 --fp16 0 # 2. --tf32 0 --fp16 1 # 3. --tf32 0 --bf16 1 # 4. --tf32 1 --fp16 0 # 5. --tf32 1 --fp16 1 # 6. --tf32 1 --bf16 1 # # and print the results. This is just a cartesian product - and more than 2 dimensions can be used. # # If you want to rely on defaults, this: # --variations '--tf32 0|--tf32 1' '--fp16 0|--fp16 1|--bf16 1' # is identical to this: # --variations '--tf32 0|--tf32 1' '|--fp16|--bf16' # # the leading empty variation in the 2nd dimension is a valid variation. # # So here we get the following 6 variations: # # 1. --tf32 0 # 2. --tf32 0 --fp16 # 3. --tf32 0 --bf16 # 4. --tf32 1 # 5. --tf32 1 --fp16 # 6. --tf32 1 --bf16 # # In this particular case we don't know what the default tf32 setting is as it's normally # pytorch-version dependent). That's why it's best to do an explicit setting of each variation: # `--tf32 0|--tf32 1` # # Here is a full example of a train: # # CUDA_VISIBLE_DEVICES=0 python ./scripts/benchmark/trainer-benchmark.py \ # --base-cmd \ # ' examples/pytorch/translation/run_translation.py --model_name_or_path t5-small \ # --output_dir output_dir --do_train --label_smoothing 0.1 --logging_strategy no \ # --save_strategy no --per_device_train_batch_size 32 --max_source_length 512 \ # --max_target_length 512 --num_train_epochs 1 --overwrite_output_dir \ # --source_lang en --target_lang ro --dataset_name wmt16 --dataset_config "ro-en" \ # --source_prefix "translate English to Romanian: " --warmup_steps 50 \ # --max_train_samples 20000 --dataloader_num_workers 2 ' \ # --target-metric-key train_samples_per_second --repeat-times 1 --variations \ # '|--fp16|--bf16' '--tf32 0|--tf32 1' --report-metric-keys train_loss \ # --repeat-times 1 --base-variation '--tf32 0' # # and here is a possible output: # # # | Variation | Train | Diff | Train | # | | samples | % | loss | # | | per | | | # | | second | | | # |:----------------|----------:|-------:|--------:| # | --tf32 0 | 285.11 | 0 | 2.51 | # | --tf32 1 | 342.09 | 20 | 2.51 | # | --fp16 --tf32 0 | 423.49 | 49 | 2.51 | # | --fp16 --tf32 1 | 423.13 | 48 | 2.51 | # | --bf16 --tf32 0 | 416.80 | 46 | 2.52 | # | --bf16 --tf32 1 | 415.87 | 46 | 2.52 | # # # So you can quickly compare the different outcomes. # # Typically running each experiment once is enough, but if the environment is unstable you can # re-run each multiple times, e.g., 3 using --repeat-times 3 and it will report the averaged results. # # By default it'll use the lowest result as the base line to use as 100% and then compare the rest to # it as can be seen from the table above, but you can also specify which combination is the one to use as # the baseline, e.g., to change to another entry use: --base-variation '--tf32 1 --fp16 0' # # --target-metric-key is there to tell the program which metrics to compare - the different metric keys are # inside output_dir/all_results.json. e.g., to measure eval performance instead of train use: # --target-metric-key eval_samples_per_second # but of course you will need to adjust the --base-cmd value in the example to perform evaluation as # well (as currently it doesn't) # import argparse import datetime import io import itertools import json import math import os import platform import re import shlex import subprocess import sys from pathlib import Path from statistics import fmean import pandas as pd import torch from tqdm import tqdm import transformers SCREAMING_SNAKE_CASE_ = float('nan') class snake_case_ : """simple docstring""" def __init__( self , lowerCamelCase_) -> Union[str, Any]: UpperCamelCase = sys.stdout UpperCamelCase = open(lowerCamelCase_ , '''a''') def __getattr__( self , lowerCamelCase_) -> Optional[Any]: return getattr(self.stdout , lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Optional[int]: self.stdout.write(lowerCamelCase_) # strip tqdm codes self.file.write(re.sub(R'''^.*\r''' , '''''' , lowerCamelCase_ , 0 , re.M)) def __snake_case ( _lowercase=80 ,_lowercase=False ): """simple docstring""" UpperCamelCase = [] # deal with critical env vars UpperCamelCase = ['''CUDA_VISIBLE_DEVICES'''] for key in env_keys: UpperCamelCase = os.environ.get(_lowercase ,_lowercase ) if val is not None: cmd.append(f'{key}={val}' ) # python executable (not always needed if the script is executable) UpperCamelCase = sys.executable if full_python_path else sys.executable.split('''/''' )[-1] cmd.append(_lowercase ) # now the normal args cmd += list(map(shlex.quote ,sys.argv ) ) # split up into up to MAX_WIDTH lines with shell multi-line escapes UpperCamelCase = [] UpperCamelCase = '''''' while len(_lowercase ) > 0: current_line += f'{cmd.pop(0 )} ' if len(_lowercase ) == 0 or len(_lowercase ) + len(cmd[0] ) + 1 > max_width - 1: lines.append(_lowercase ) UpperCamelCase = '''''' return "\\\n".join(_lowercase ) def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" UpperCamelCase = re.sub(r'''[\\\n]+''' ,''' ''' ,args.base_cmd ) # remove --output_dir if any and set our own UpperCamelCase = re.sub('''--output_dir\s+[^\s]+''' ,'''''' ,args.base_cmd ) args.base_cmd += f' --output_dir {output_dir}' # ensure we have --overwrite_output_dir UpperCamelCase = re.sub('''--overwrite_output_dir\s+''' ,'''''' ,args.base_cmd ) args.base_cmd += " --overwrite_output_dir" return [sys.executable] + shlex.split(args.base_cmd ) def __snake_case ( _lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ): """simple docstring""" if 0: import random from time import sleep sleep(0 ) return dict( {k: random.uniform(0 ,100 ) for k in metric_keys} ,**{target_metric_key: random.choice([nan, 10.31, 100.2, 55.6666, 222.22222222] )} ,) UpperCamelCase = subprocess.run(_lowercase ,capture_output=_lowercase ,text=_lowercase ) if verbose: print('''STDOUT''' ,result.stdout ) print('''STDERR''' ,result.stderr ) # save the streams UpperCamelCase = variation.replace(''' ''' ,'''-''' ) with open(Path(_lowercase ) / f'log.{prefix}.stdout.txt' ,'''w''' ) as f: f.write(result.stdout ) with open(Path(_lowercase ) / f'log.{prefix}.stderr.txt' ,'''w''' ) as f: f.write(result.stderr ) if result.returncode != 0: if verbose: print('''failed''' ) return {target_metric_key: nan} with io.open(f'{output_dir}/all_results.json' ,'''r''' ,encoding='''utf-8''' ) as f: UpperCamelCase = json.load(_lowercase ) # filter out just the keys we want return {k: v for k, v in metrics.items() if k in metric_keys} def __snake_case ( _lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ,): """simple docstring""" UpperCamelCase = [] UpperCamelCase = [] UpperCamelCase = f'{id}: {variation:<{longest_variation_len}}' UpperCamelCase = f'{preamble}: ' UpperCamelCase = set(report_metric_keys + [target_metric_key] ) for i in tqdm(range(_lowercase ) ,desc=_lowercase ,leave=_lowercase ): UpperCamelCase = process_run_single( _lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ) UpperCamelCase = single_run_metrics[target_metric_key] if not math.isnan(_lowercase ): metrics.append(_lowercase ) results.append(_lowercase ) outcome += "✓" else: outcome += "✘" UpperCamelCase = f'\33[2K\r{outcome}' if len(_lowercase ) > 0: UpperCamelCase = {k: fmean([x[k] for x in metrics] ) for k in metrics[0].keys()} UpperCamelCase = round(mean_metrics[target_metric_key] ,2 ) UpperCamelCase = f'{outcome} {mean_target}' if len(_lowercase ) > 1: results_str += f' {tuple(round(_lowercase ,2 ) for x in results )}' print(_lowercase ) UpperCamelCase = variation return mean_metrics else: print(_lowercase ) return {variation_key: variation, target_metric_key: nan} def __snake_case ( ): """simple docstring""" UpperCamelCase = torch.cuda.get_device_properties(torch.device('''cuda''' ) ) return f'\nDatetime : {datetime.datetime.now().strftime("%Y-%m-%d %H:%M:%S" )}\n\nSoftware:\ntransformers: {transformers.__version__}\ntorch : {torch.__version__}\ncuda : {torch.version.cuda}\npython : {platform.python_version()}\n\nHardware:\n{torch.cuda.device_count()} GPUs : {properties.name}, {properties.total_memory/2**30:0.2f}GB\n' def __snake_case ( _lowercase ,_lowercase ,_lowercase ,_lowercase ,_lowercase ): """simple docstring""" UpperCamelCase = pd.DataFrame(_lowercase ) UpperCamelCase = '''variation''' UpperCamelCase = '''diff_%''' UpperCamelCase = nan if base_variation is not None and len(df[df[variation_key] == base_variation] ): # this may still return nan UpperCamelCase = df.loc[df[variation_key] == base_variation][target_metric_key].item() if math.isnan(_lowercase ): # as a fallback, use the minimal value as the sentinel UpperCamelCase = df.loc[df[target_metric_key] != nan][target_metric_key].min() # create diff column if possible if not math.isnan(_lowercase ): UpperCamelCase = df.apply( lambda _lowercase : round(100 * (r[target_metric_key] - sentinel_value) / sentinel_value ) if not math.isnan(r[target_metric_key] ) else 0 ,axis='''columns''' ,) # re-order columns UpperCamelCase = [variation_key, target_metric_key, diff_key, *report_metric_keys] UpperCamelCase = df.reindex(_lowercase ,axis='''columns''' ) # reorder cols # capitalize UpperCamelCase = df.rename(str.capitalize ,axis='''columns''' ) # make the cols as narrow as possible UpperCamelCase = df.rename(lambda _lowercase : c.replace('''_''' ,'''<br>''' ) ,axis='''columns''' ) UpperCamelCase = df.rename(lambda _lowercase : c.replace('''_''' ,'''\n''' ) ,axis='''columns''' ) UpperCamelCase = ['''''', '''Copy between the cut-here-lines and paste as is to github or a forum'''] report += ["----------8<-----------------8<--------"] report += ["*** Results:", df_github.to_markdown(index=_lowercase ,floatfmt='''.2f''' )] report += ["```"] report += ["*** Setup:", get_versions()] report += ["*** The benchmark command line was:", get_original_command()] report += ["```"] report += ["----------8<-----------------8<--------"] report += ["*** Results (console):", df_console.to_markdown(index=_lowercase ,floatfmt='''.2f''' )] print('''\n\n'''.join(_lowercase ) ) def __snake_case ( ): """simple docstring""" UpperCamelCase = argparse.ArgumentParser() parser.add_argument( '''--base-cmd''' ,default=_lowercase ,type=_lowercase ,required=_lowercase ,help='''Base cmd''' ,) parser.add_argument( '''--variations''' ,default=_lowercase ,type=_lowercase ,nargs='''+''' ,required=_lowercase ,help='''Multi-dimensional variations, example: \'|--fp16|--bf16\' \'|--tf32\'''' ,) parser.add_argument( '''--base-variation''' ,default=_lowercase ,type=_lowercase ,help='''Baseline variation to compare to. if None the minimal target value will be used to compare against''' ,) parser.add_argument( '''--target-metric-key''' ,default=_lowercase ,type=_lowercase ,required=_lowercase ,help='''Target metric key in output_dir/all_results.json, e.g., train_samples_per_second''' ,) parser.add_argument( '''--report-metric-keys''' ,default='''''' ,type=_lowercase ,help='''Report metric keys - other metric keys from output_dir/all_results.json to report, e.g., train_loss. Use a single argument e.g., \'train_loss train_samples''' ,) parser.add_argument( '''--repeat-times''' ,default=1 ,type=_lowercase ,help='''How many times to re-run each variation - an average will be reported''' ,) parser.add_argument( '''--output_dir''' ,default='''output_benchmark''' ,type=_lowercase ,help='''The output directory where all the benchmark reports will go to and additionally this directory will be used to override --output_dir in the script that is being benchmarked''' ,) parser.add_argument( '''--verbose''' ,default=_lowercase ,action='''store_true''' ,help='''Whether to show the outputs of each run or just the benchmark progress''' ,) UpperCamelCase = parser.parse_args() UpperCamelCase = args.output_dir Path(_lowercase ).mkdir(exist_ok=_lowercase ) UpperCamelCase = get_base_command(_lowercase ,_lowercase ) # split each dimension into its --foo variations UpperCamelCase = [list(map(str.strip ,re.split(r'''\|''' ,_lowercase ) ) ) for x in args.variations] # build a cartesian product of dimensions and convert those back into cmd-line arg strings, # while stripping white space for inputs that were empty UpperCamelCase = list(map(str.strip ,map(''' '''.join ,itertools.product(*_lowercase ) ) ) ) UpperCamelCase = max(len(_lowercase ) for x in variations ) # split wanted keys UpperCamelCase = args.report_metric_keys.split() # capture prints into a log file for convenience UpperCamelCase = f'benchmark-report-{datetime.datetime.now().strftime("%Y-%m-%d-%H-%M-%S" )}.txt' print(f'\nNote: each run\'s output is also logged under {output_dir}/log.*.std*.txt' ) print(f'and this script\'s output is also piped into {report_fn}' ) UpperCamelCase = Tee(_lowercase ) print(f'\n*** Running {len(_lowercase )} benchmarks:' ) print(f'Base command: {" ".join(_lowercase )}' ) UpperCamelCase = '''variation''' UpperCamelCase = [] for id, variation in enumerate(tqdm(_lowercase ,desc='''Total completion: ''' ,leave=_lowercase ) ): UpperCamelCase = base_cmd + variation.split() results.append( process_run( id + 1 ,_lowercase ,_lowercase ,_lowercase ,_lowercase ,args.target_metric_key ,_lowercase ,args.repeat_times ,_lowercase ,args.verbose ,) ) process_results(_lowercase ,args.target_metric_key ,_lowercase ,args.base_variation ,_lowercase ) if __name__ == "__main__": main()
34
"""simple docstring""" import operator def __snake_case ( _lowercase ,_lowercase = False ,_lowercase = None ): """simple docstring""" UpperCamelCase = operator.lt if reverse else operator.gt UpperCamelCase = solution or [] if not arr: return solution UpperCamelCase = [arr.pop(0 )] for i, item in enumerate(_lowercase ): if _operator(_lowercase ,sublist[-1] ): sublist.append(_lowercase ) arr.pop(_lowercase ) # merging sublist into solution list if not solution: solution.extend(_lowercase ) else: while sublist: UpperCamelCase = sublist.pop(0 ) for i, xx in enumerate(_lowercase ): if not _operator(_lowercase ,_lowercase ): solution.insert(_lowercase ,_lowercase ) break else: solution.append(_lowercase ) strand_sort(_lowercase ,_lowercase ,_lowercase ) return solution if __name__ == "__main__": assert strand_sort([4, 3, 5, 1, 2]) == [1, 2, 3, 4, 5] assert strand_sort([4, 3, 5, 1, 2], reverse=True) == [5, 4, 3, 2, 1]
34
1
"""simple docstring""" import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / 'utils')) from test_module.custom_configuration import CustomConfig # noqa E402 SCREAMING_SNAKE_CASE_ = { 'return_dict': False, 'output_hidden_states': True, 'output_attentions': True, 'torchscript': True, 'torch_dtype': 'float16', 'use_bfloat16': True, 'tf_legacy_loss': True, 'pruned_heads': {'a': 1}, 'tie_word_embeddings': False, 'is_decoder': True, 'cross_attention_hidden_size': 128, 'add_cross_attention': True, 'tie_encoder_decoder': True, 'max_length': 50, 'min_length': 3, 'do_sample': True, 'early_stopping': True, 'num_beams': 3, 'num_beam_groups': 3, 'diversity_penalty': 0.5, 'temperature': 2.0, 'top_k': 10, 'top_p': 0.7, 'typical_p': 0.2, 'repetition_penalty': 0.8, 'length_penalty': 0.8, 'no_repeat_ngram_size': 5, 'encoder_no_repeat_ngram_size': 5, 'bad_words_ids': [1, 2, 3], 'num_return_sequences': 3, 'chunk_size_feed_forward': 5, 'output_scores': True, 'return_dict_in_generate': True, 'forced_bos_token_id': 2, 'forced_eos_token_id': 3, 'remove_invalid_values': True, 'architectures': ['BertModel'], 'finetuning_task': 'translation', 'id2label': {0: 'label'}, 'label2id': {'label': '0'}, 'tokenizer_class': 'BertTokenizerFast', 'prefix': 'prefix', 'bos_token_id': 6, 'pad_token_id': 7, 'eos_token_id': 8, 'sep_token_id': 9, 'decoder_start_token_id': 10, 'exponential_decay_length_penalty': (5, 1.01), 'suppress_tokens': [0, 1], 'begin_suppress_tokens': 2, 'task_specific_params': {'translation': 'some_params'}, 'problem_type': 'regression', } @is_staging_test class snake_case_ ( unittest.TestCase ): """simple docstring""" @classmethod def UpperCAmelCase__ ( cls) -> Dict: UpperCamelCase = TOKEN HfFolder.save_token(lowerCamelCase_) @classmethod def UpperCAmelCase__ ( cls) -> Union[str, Any]: try: delete_repo(token=cls._token , repo_id='''test-config''') except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''valid_org/test-config-org''') except HTTPError: pass try: delete_repo(token=cls._token , repo_id='''test-dynamic-config''') except HTTPError: pass def UpperCAmelCase__ ( self) -> int: UpperCamelCase = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7) config.push_to_hub('''test-config''' , use_auth_token=self._token) UpperCamelCase = BertConfig.from_pretrained(F'{USER}/test-config') for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase_ , getattr(lowerCamelCase_ , lowerCamelCase_)) # Reset repo delete_repo(token=self._token , repo_id='''test-config''') # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(lowerCamelCase_ , repo_id='''test-config''' , push_to_hub=lowerCamelCase_ , use_auth_token=self._token) UpperCamelCase = BertConfig.from_pretrained(F'{USER}/test-config') for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase_ , getattr(lowerCamelCase_ , lowerCamelCase_)) def UpperCAmelCase__ ( self) -> List[str]: UpperCamelCase = BertConfig( vocab_size=9_9 , hidden_size=3_2 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=3_7) config.push_to_hub('''valid_org/test-config-org''' , use_auth_token=self._token) UpperCamelCase = BertConfig.from_pretrained('''valid_org/test-config-org''') for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase_ , getattr(lowerCamelCase_ , lowerCamelCase_)) # Reset repo delete_repo(token=self._token , repo_id='''valid_org/test-config-org''') # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( lowerCamelCase_ , repo_id='''valid_org/test-config-org''' , push_to_hub=lowerCamelCase_ , use_auth_token=self._token) UpperCamelCase = BertConfig.from_pretrained('''valid_org/test-config-org''') for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(lowerCamelCase_ , getattr(lowerCamelCase_ , lowerCamelCase_)) def UpperCAmelCase__ ( self) -> Tuple: CustomConfig.register_for_auto_class() UpperCamelCase = CustomConfig(attribute=4_2) config.push_to_hub('''test-dynamic-config''' , use_auth_token=self._token) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {'''AutoConfig''': '''custom_configuration.CustomConfig'''}) UpperCamelCase = AutoConfig.from_pretrained(F'{USER}/test-dynamic-config' , trust_remote_code=lowerCamelCase_) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , '''CustomConfig''') self.assertEqual(new_config.attribute , 4_2) class snake_case_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self) -> Dict: UpperCamelCase = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated UpperCamelCase = c.n_embd + 1 # int UpperCamelCase = c.resid_pdrop + 1.0 # float UpperCamelCase = not c.scale_attn_weights # bool UpperCamelCase = c.summary_type + '''foo''' # str c.update_from_string( F'n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}') self.assertEqual(lowerCamelCase_ , c.n_embd , '''mismatch for key: n_embd''') self.assertEqual(lowerCamelCase_ , c.resid_pdrop , '''mismatch for key: resid_pdrop''') self.assertEqual(lowerCamelCase_ , c.scale_attn_weights , '''mismatch for key: scale_attn_weights''') self.assertEqual(lowerCamelCase_ , c.summary_type , '''mismatch for key: summary_type''') def UpperCAmelCase__ ( self) -> str: UpperCamelCase = PretrainedConfig() UpperCamelCase = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( lowerCamelCase_ , ['''is_encoder_decoder''', '''_name_or_path''', '''_commit_hash''', '''transformers_version''']) UpperCamelCase = [key for key, value in config_common_kwargs.items() if value == getattr(lowerCamelCase_ , lowerCamelCase_)] if len(lowerCamelCase_) > 0: raise ValueError( '''The following keys are set with the default values in''' ''' `test_configuration_common.config_common_kwargs` pick another value for them:''' F' {", ".join(lowerCamelCase_)}.') def UpperCAmelCase__ ( self) -> str: with self.assertRaises(lowerCamelCase_): # config is in subfolder, the following should not work without specifying the subfolder UpperCamelCase = BertConfig.from_pretrained('''hf-internal-testing/tiny-random-bert-subfolder''') UpperCamelCase = BertConfig.from_pretrained('''hf-internal-testing/tiny-random-bert-subfolder''' , subfolder='''bert''') self.assertIsNotNone(lowerCamelCase_) def UpperCAmelCase__ ( self) -> Union[str, Any]: # A mock response for an HTTP head request to emulate server down UpperCamelCase = mock.Mock() UpperCamelCase = 5_0_0 UpperCamelCase = {} UpperCamelCase = HTTPError UpperCamelCase = {} # Download this model to make sure it's in the cache. UpperCamelCase = BertConfig.from_pretrained('''hf-internal-testing/tiny-random-bert''') # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch('''requests.Session.request''' , return_value=lowerCamelCase_) as mock_head: UpperCamelCase = BertConfig.from_pretrained('''hf-internal-testing/tiny-random-bert''') # This check we did call the fake head request mock_head.assert_called() def UpperCAmelCase__ ( self) -> Optional[int]: # This test is for deprecated behavior and can be removed in v5 UpperCamelCase = BertConfig.from_pretrained( '''https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json''') def UpperCAmelCase__ ( self) -> Union[str, Any]: UpperCamelCase = AutoConfig.from_pretrained('''bert-base-cased''') UpperCamelCase = ['''config.4.0.0.json'''] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(lowerCamelCase_) UpperCamelCase = 2 json.dump(configuration.to_dict() , open(os.path.join(lowerCamelCase_ , '''config.4.0.0.json''') , '''w''')) # This should pick the new configuration file as the version of Transformers is > 4.0.0 UpperCamelCase = AutoConfig.from_pretrained(lowerCamelCase_) self.assertEqual(new_configuration.hidden_size , 2) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 UpperCamelCase = ['''config.42.0.0.json'''] UpperCamelCase = 7_6_8 configuration.save_pretrained(lowerCamelCase_) shutil.move(os.path.join(lowerCamelCase_ , '''config.4.0.0.json''') , os.path.join(lowerCamelCase_ , '''config.42.0.0.json''')) UpperCamelCase = AutoConfig.from_pretrained(lowerCamelCase_) self.assertEqual(new_configuration.hidden_size , 7_6_8) def UpperCAmelCase__ ( self) -> Optional[Any]: # This repo has two configuration files, one for v4.0.0 and above with a different hidden size. UpperCamelCase = '''hf-internal-testing/test-two-configs''' import transformers as new_transformers UpperCamelCase = '''v4.0.0''' UpperCamelCase , UpperCamelCase = new_transformers.models.auto.AutoConfig.from_pretrained( lowerCamelCase_ , return_unused_kwargs=lowerCamelCase_) self.assertEqual(new_configuration.hidden_size , 2) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(lowerCamelCase_ , {}) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers UpperCamelCase = '''v3.0.0''' UpperCamelCase = old_transformers.models.auto.AutoConfig.from_pretrained(lowerCamelCase_) self.assertEqual(old_configuration.hidden_size , 7_6_8)
34
"""simple docstring""" from scipy.stats import pearsonr import datasets SCREAMING_SNAKE_CASE_ = '\nPearson correlation coefficient and p-value for testing non-correlation.\nThe Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each dataset is normally distributed. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Correlations of -1 or +1 imply an exact linear relationship. Positive correlations imply that as x increases, so does y. Negative correlations imply that as x increases, y decreases.\nThe p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets.\n' SCREAMING_SNAKE_CASE_ = '\nArgs:\n predictions (`list` of `int`): Predicted class labels, as returned by a model.\n references (`list` of `int`): Ground truth labels.\n return_pvalue (`boolean`): If `True`, returns the p-value, along with the correlation coefficient. If `False`, returns only the correlation coefficient. Defaults to `False`.\n\nReturns:\n pearsonr (`float`): Pearson correlation coefficient. Minimum possible value is -1. Maximum possible value is 1. Values of 1 and -1 indicate exact linear positive and negative relationships, respectively. A value of 0 implies no correlation.\n p-value (`float`): P-value, which roughly indicates the probability of an The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Pearson correlation at least as extreme as the one computed from these datasets. Minimum possible value is 0. Maximum possible value is 1. Higher values indicate higher probabilities.\n\nExamples:\n\n Example 1-A simple example using only predictions and references.\n >>> pearsonr_metric = datasets.load_metric("pearsonr")\n >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5])\n >>> print(round(results[\'pearsonr\'], 2))\n -0.74\n\n Example 2-The same as Example 1, but that also returns the `p-value`.\n >>> pearsonr_metric = datasets.load_metric("pearsonr")\n >>> results = pearsonr_metric.compute(predictions=[10, 9, 2.5, 6, 4], references=[1, 2, 3, 4, 5], return_pvalue=True)\n >>> print(sorted(list(results.keys())))\n [\'p-value\', \'pearsonr\']\n >>> print(round(results[\'pearsonr\'], 2))\n -0.74\n >>> print(round(results[\'p-value\'], 2))\n 0.15\n' SCREAMING_SNAKE_CASE_ = '\n@article{2020SciPy-NMeth,\nauthor = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and\n Haberland, Matt and Reddy, Tyler and Cournapeau, David and\n Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and\n Bright, Jonathan and {van der Walt}, St{\'e}fan J. and\n Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and\n Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and\n Kern, Robert and Larson, Eric and Carey, C J and\n Polat, Ilhan and Feng, Yu and Moore, Eric W. and\n {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and\n Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and\n Harris, Charles R. and Archibald, Anne M. and\n Ribeiro, Antonio H. and Pedregosa, Fabian and\n {van Mulbregt}, Paul and {SciPy 1.0 Contributors}},\ntitle = {{{SciPy} 1.0: Fundamental Algorithms for Scientific\n Computing in Python}},\njournal = {Nature Methods},\nyear = {2020},\nvolume = {17},\npages = {261--272},\nadsurl = {https://rdcu.be/b08Wh},\ndoi = {10.1038/s41592-019-0686-2},\n}\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class snake_case_ ( datasets.Metric ): """simple docstring""" def UpperCAmelCase__ ( self) -> Tuple: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''float'''), '''references''': datasets.Value('''float'''), }) , reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.pearsonr.html'''] , ) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=False) -> Any: if return_pvalue: UpperCamelCase = pearsonr(lowerCamelCase_ , lowerCamelCase_) return {"pearsonr": results[0], "p-value": results[1]} else: return {"pearsonr": float(pearsonr(lowerCamelCase_ , lowerCamelCase_)[0])}
34
1
"""simple docstring""" from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = { 'linear': get_linear_schedule_with_warmup, 'cosine': get_cosine_schedule_with_warmup, 'cosine_w_restarts': get_cosine_with_hard_restarts_schedule_with_warmup, 'polynomial': get_polynomial_decay_schedule_with_warmup, 'constant': get_constant_schedule, 'constant_w_warmup': get_constant_schedule_with_warmup, } class snake_case_ ( lowerCamelCase_ ): """simple docstring""" def __init__( self , lowerCamelCase_=None , lowerCamelCase_=None , *lowerCamelCase_ , **lowerCamelCase_) -> int: super().__init__(*lowerCamelCase_ , **lowerCamelCase_) if config is None: assert isinstance(self.model , lowerCamelCase_), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F' {self.model.__class__}' ) UpperCamelCase = self.model.config else: UpperCamelCase = config UpperCamelCase = data_args UpperCamelCase = self.config.tgt_vocab_size if isinstance(self.config , lowerCamelCase_) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F'The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for' ''' padding..''') if self.args.label_smoothing == 0: UpperCamelCase = torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss UpperCamelCase = label_smoothed_nll_loss def UpperCAmelCase__ ( self , lowerCamelCase_) -> Optional[int]: if self.optimizer is None: UpperCamelCase = ['''bias''', '''LayerNorm.weight'''] UpperCamelCase = [ { '''params''': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay)], '''weight_decay''': self.args.weight_decay, }, { '''params''': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay)], '''weight_decay''': 0.0, }, ] UpperCamelCase = Adafactor if self.args.adafactor else AdamW if self.args.adafactor: UpperCamelCase = Adafactor UpperCamelCase = {'''scale_parameter''': False, '''relative_step''': False} else: UpperCamelCase = AdamW UpperCamelCase = { '''betas''': (self.args.adam_betaa, self.args.adam_betaa), '''eps''': self.args.adam_epsilon, } UpperCamelCase = self.args.learning_rate if self.sharded_ddp: UpperCamelCase = OSS( params=lowerCamelCase_ , optim=lowerCamelCase_ , **lowerCamelCase_ , ) else: UpperCamelCase = optimizer_cls(lowerCamelCase_ , **lowerCamelCase_) if self.lr_scheduler is None: UpperCamelCase = self._get_lr_scheduler(lowerCamelCase_) else: # ignoring --lr_scheduler logger.warning('''scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.''') def UpperCAmelCase__ ( self , lowerCamelCase_) -> List[Any]: UpperCamelCase = arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": UpperCamelCase = schedule_func(self.optimizer) elif self.args.lr_scheduler == "constant_w_warmup": UpperCamelCase = schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps) else: UpperCamelCase = schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=lowerCamelCase_) return scheduler def UpperCAmelCase__ ( self) -> Optional[torch.utils.data.Sampler]: if isinstance(self.train_dataset , torch.utils.data.IterableDataset): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset) ) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_) -> Optional[Any]: if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token UpperCamelCase = model(**lowerCamelCase_ , use_cache=lowerCamelCase_)[0] UpperCamelCase = self.loss_fn(logits.view(-1 , logits.shape[-1]) , labels.view(-1)) else: # compute usual loss via models UpperCamelCase , UpperCamelCase = model(**lowerCamelCase_ , labels=lowerCamelCase_ , use_cache=lowerCamelCase_)[:2] else: # compute label smoothed loss UpperCamelCase = model(**lowerCamelCase_ , use_cache=lowerCamelCase_)[0] UpperCamelCase = torch.nn.functional.log_softmax(lowerCamelCase_ , dim=-1) UpperCamelCase , UpperCamelCase = self.loss_fn(lowerCamelCase_ , lowerCamelCase_ , self.args.label_smoothing , ignore_index=self.config.pad_token_id) return loss, logits def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_) -> Any: UpperCamelCase = inputs.pop('''labels''') UpperCamelCase , UpperCamelCase = self._compute_loss(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_) return loss def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = None , ) -> Tuple[Optional[float], Optional[torch.Tensor], Optional[torch.Tensor]]: UpperCamelCase = self._prepare_inputs(lowerCamelCase_) UpperCamelCase = { '''max_length''': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, '''num_beams''': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: UpperCamelCase = self.model.generate( inputs['''input_ids'''] , attention_mask=inputs['''attention_mask'''] , **lowerCamelCase_ , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: UpperCamelCase = self._pad_tensors_to_max_len(lowerCamelCase_ , gen_kwargs['''max_length''']) UpperCamelCase = inputs.pop('''labels''') with torch.no_grad(): # compute loss on predict data UpperCamelCase , UpperCamelCase = self._compute_loss(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_) UpperCamelCase = loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) UpperCamelCase = generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: UpperCamelCase = self._pad_tensors_to_max_len(lowerCamelCase_ , gen_kwargs['''max_length''']) return (loss, logits, labels) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_) -> List[Any]: # If PAD token is not defined at least EOS token has to be defined UpperCamelCase = self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( '''Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be''' F' padded to `max_length`={max_length}') UpperCamelCase = pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device) UpperCamelCase = tensor return padded_tensor
34
"""simple docstring""" import unittest from dataclasses import dataclass import pytest from accelerate.commands.config.config_args import SageMakerConfig from accelerate.utils import ComputeEnvironment from accelerate.utils.launch import _convert_nargs_to_dict @dataclass class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = ComputeEnvironment.AMAZON_SAGEMAKER A_ = True A_ = '''ml.p3.2xlarge''' A_ = '''accelerate_sagemaker_execution_role''' A_ = '''hf-sm''' A_ = '''us-east-1''' A_ = 1 A_ = '''accelerate-sagemaker-1''' A_ = '''1.6''' A_ = '''4.4''' A_ = '''train.py''' A_ = [ '''--model_name_or_path''', '''bert''', '''--do_train''', '''False''', '''--epochs''', '''3''', '''--learning_rate''', '''5e-5''', '''--max_steps''', '''50.5''', ] A_ = [ '''--model_name_or_path''', '''bert''', '''--do_train''', '''--do_test''', '''False''', '''--do_predict''', '''--epochs''', '''3''', '''--learning_rate''', '''5e-5''', '''--max_steps''', '''50.5''', ] class snake_case_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self) -> List[Any]: # If no defaults are changed, `to_kwargs` returns an empty dict. UpperCamelCase = _convert_nargs_to_dict(MockLaunchConfig.success_training_script_args) assert isinstance(converted_args['''model_name_or_path'''] , lowerCamelCase_) assert isinstance(converted_args['''do_train'''] , lowerCamelCase_) assert isinstance(converted_args['''epochs'''] , lowerCamelCase_) assert isinstance(converted_args['''learning_rate'''] , lowerCamelCase_) assert isinstance(converted_args['''max_steps'''] , lowerCamelCase_) with pytest.raises(lowerCamelCase_): _convert_nargs_to_dict(MockLaunchConfig.fail_training_script_args)
34
1
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Audio, Features, Value from .base import TaskTemplate @dataclass(frozen=lowerCamelCase_ ) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = field(default='''automatic-speech-recognition''' , metadata={'''include_in_asdict_even_if_is_default''': True} ) A_ = Features({'''audio''': Audio()} ) A_ = Features({'''transcription''': Value('''string''' )} ) A_ = "audio" A_ = "transcription" def UpperCAmelCase__ ( self , lowerCamelCase_) -> str: if self.audio_column not in features: raise ValueError(F'Column {self.audio_column} is not present in features.') if not isinstance(features[self.audio_column] , lowerCamelCase_): raise ValueError(F'Column {self.audio_column} is not an Audio type.') UpperCamelCase = copy.deepcopy(self) UpperCamelCase = self.input_schema.copy() UpperCamelCase = features[self.audio_column] UpperCamelCase = input_schema return task_template @property def UpperCAmelCase__ ( self) -> Dict[str, str]: return {self.audio_column: "audio", self.transcription_column: "transcription"}
34
"""simple docstring""" from typing import List import jiwer import jiwer.transforms as tr from packaging import version import datasets from datasets.config import PY_VERSION if PY_VERSION < version.parse('3.8'): import importlib_metadata else: import importlib.metadata as importlib_metadata SCREAMING_SNAKE_CASE_ = '' if version.parse(importlib_metadata.version('jiwer')) < version.parse('2.3.0'): class snake_case_ ( tr.AbstractTransform ): """simple docstring""" def __init__( self , lowerCamelCase_ = " ") -> List[str]: UpperCamelCase = sentence_delimiter def UpperCAmelCase__ ( self , lowerCamelCase_) -> Tuple: return list(lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Optional[Any]: UpperCamelCase = [] for sent_idx, sentence in enumerate(lowerCamelCase_): chars.extend(self.process_string(lowerCamelCase_)) if self.sentence_delimiter is not None and self.sentence_delimiter != "" and sent_idx < len(lowerCamelCase_) - 1: chars.append(self.sentence_delimiter) return chars SCREAMING_SNAKE_CASE_ = tr.Compose( [tr.RemoveMultipleSpaces(), tr.Strip(), SentencesToListOfCharacters(SENTENCE_DELIMITER)] ) else: SCREAMING_SNAKE_CASE_ = tr.Compose( [ tr.RemoveMultipleSpaces(), tr.Strip(), tr.ReduceToSingleSentence(SENTENCE_DELIMITER), tr.ReduceToListOfListOfChars(), ] ) SCREAMING_SNAKE_CASE_ = '\\n@inproceedings{inproceedings,\n author = {Morris, Andrew and Maier, Viktoria and Green, Phil},\n year = {2004},\n month = {01},\n pages = {},\n title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.}\n}\n' SCREAMING_SNAKE_CASE_ = '\\nCharacter error rate (CER) is a common metric of the performance of an automatic speech recognition system.\n\nCER is similar to Word Error Rate (WER), but operates on character instead of word. Please refer to docs of WER for further information.\n\nCharacter error rate can be computed as:\n\nCER = (S + D + I) / N = (S + D + I) / (S + D + C)\n\nwhere\n\nS is the number of substitutions,\nD is the number of deletions,\nI is the number of insertions,\nC is the number of correct characters,\nN is the number of characters in the reference (N=S+D+C).\n\nCER\'s output is not always a number between 0 and 1, in particular when there is a high number of insertions. This value is often associated to the percentage of characters that were incorrectly predicted. The lower the value, the better the\nperformance of the ASR system with a CER of 0 being a perfect score.\n' SCREAMING_SNAKE_CASE_ = '\nComputes CER score of transcribed segments against references.\nArgs:\n references: list of references for each speech input.\n predictions: list of transcribtions to score.\n concatenate_texts: Whether or not to concatenate sentences before evaluation, set to True for more accurate result.\nReturns:\n (float): the character error rate\n\nExamples:\n\n >>> predictions = ["this is the prediction", "there is an other sample"]\n >>> references = ["this is the reference", "there is another one"]\n >>> cer = datasets.load_metric("cer")\n >>> cer_score = cer.compute(predictions=predictions, references=references)\n >>> print(cer_score)\n 0.34146341463414637\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class snake_case_ ( datasets.Metric ): """simple docstring""" def UpperCAmelCase__ ( self) -> Dict: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence'''), '''references''': datasets.Value('''string''' , id='''sequence'''), }) , codebase_urls=['''https://github.com/jitsi/jiwer/'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/Word_error_rate''', '''https://sites.google.com/site/textdigitisation/qualitymeasures/computingerrorrates''', ] , ) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=False) -> List[Any]: if concatenate_texts: return jiwer.compute_measures( lowerCamelCase_ , lowerCamelCase_ , truth_transform=lowerCamelCase_ , hypothesis_transform=lowerCamelCase_ , )["wer"] UpperCamelCase = 0 UpperCamelCase = 0 for prediction, reference in zip(lowerCamelCase_ , lowerCamelCase_): UpperCamelCase = jiwer.compute_measures( lowerCamelCase_ , lowerCamelCase_ , truth_transform=lowerCamelCase_ , hypothesis_transform=lowerCamelCase_ , ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
34
1
"""simple docstring""" import logging import os from logging import ( CRITICAL, # NOQA DEBUG, # NOQA ERROR, # NOQA FATAL, # NOQA INFO, # NOQA NOTSET, # NOQA WARN, # NOQA WARNING, # NOQA ) from typing import Optional from tqdm import auto as tqdm_lib SCREAMING_SNAKE_CASE_ = { 'debug': logging.DEBUG, 'info': logging.INFO, 'warning': logging.WARNING, 'error': logging.ERROR, 'critical': logging.CRITICAL, } SCREAMING_SNAKE_CASE_ = logging.WARNING def __snake_case ( ): """simple docstring""" UpperCamelCase = os.getenv('''DATASETS_VERBOSITY''' ,_lowercase ) if env_level_str: if env_level_str in log_levels: return log_levels[env_level_str] else: logging.getLogger().warning( f'Unknown option DATASETS_VERBOSITY={env_level_str}, ' f'has to be one of: { ", ".join(log_levels.keys() ) }' ) return _default_log_level def __snake_case ( ): """simple docstring""" return __name__.split('''.''' )[0] def __snake_case ( ): """simple docstring""" return logging.getLogger(_get_library_name() ) def __snake_case ( ): """simple docstring""" UpperCamelCase = _get_library_root_logger() library_root_logger.setLevel(_get_default_logging_level() ) def __snake_case ( ): """simple docstring""" UpperCamelCase = _get_library_root_logger() library_root_logger.setLevel(logging.NOTSET ) def __snake_case ( _lowercase = None ): """simple docstring""" if name is None: UpperCamelCase = _get_library_name() return logging.getLogger(_lowercase ) def __snake_case ( ): """simple docstring""" return _get_library_root_logger().getEffectiveLevel() def __snake_case ( _lowercase ): """simple docstring""" _get_library_root_logger().setLevel(_lowercase ) def __snake_case ( ): """simple docstring""" return set_verbosity(_lowercase ) def __snake_case ( ): """simple docstring""" return set_verbosity(_lowercase ) def __snake_case ( ): """simple docstring""" return set_verbosity(_lowercase ) def __snake_case ( ): """simple docstring""" return set_verbosity(_lowercase ) def __snake_case ( ): """simple docstring""" UpperCamelCase = False def __snake_case ( ): """simple docstring""" UpperCamelCase = True # Configure the library root logger at the module level (singleton-like) _configure_library_root_logger() class snake_case_ : """simple docstring""" def __init__( self , *lowerCamelCase_ , **lowerCamelCase_) -> int: # pylint: disable=unused-argument UpperCamelCase = args[0] if args else None def __iter__( self) -> Optional[int]: return iter(self._iterator) def __getattr__( self , lowerCamelCase_) -> Union[str, Any]: def empty_fn(*lowerCamelCase_ , **lowerCamelCase_): # pylint: disable=unused-argument return return empty_fn def __enter__( self) -> Optional[Any]: return self def __exit__( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_) -> str: return SCREAMING_SNAKE_CASE_ = True class snake_case_ : """simple docstring""" def __call__( self , *lowerCamelCase_ , lowerCamelCase_=False , **lowerCamelCase_) -> Union[str, Any]: if _tqdm_active and not disable: return tqdm_lib.tqdm(*lowerCamelCase_ , **lowerCamelCase_) else: return EmptyTqdm(*lowerCamelCase_ , **lowerCamelCase_) def UpperCAmelCase__ ( self , *lowerCamelCase_ , **lowerCamelCase_) -> Tuple: UpperCamelCase = None if _tqdm_active: return tqdm_lib.tqdm.set_lock(*lowerCamelCase_ , **lowerCamelCase_) def UpperCAmelCase__ ( self) -> Tuple: if _tqdm_active: return tqdm_lib.tqdm.get_lock() SCREAMING_SNAKE_CASE_ = _tqdm_cls() def __snake_case ( ): """simple docstring""" global _tqdm_active return bool(_tqdm_active ) def __snake_case ( ): """simple docstring""" global _tqdm_active UpperCamelCase = True def __snake_case ( ): """simple docstring""" global _tqdm_active UpperCamelCase = False
34
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = {'vocab_file': 'spiece.model'} SCREAMING_SNAKE_CASE_ = { 'vocab_file': { 'xlnet-base-cased': 'https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model', 'xlnet-large-cased': 'https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model', } } SCREAMING_SNAKE_CASE_ = { 'xlnet-base-cased': None, 'xlnet-large-cased': None, } # Segments (not really needed) SCREAMING_SNAKE_CASE_ = 0 SCREAMING_SNAKE_CASE_ = 1 SCREAMING_SNAKE_CASE_ = 2 SCREAMING_SNAKE_CASE_ = 3 SCREAMING_SNAKE_CASE_ = 4 class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = VOCAB_FILES_NAMES A_ = PRETRAINED_VOCAB_FILES_MAP A_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ = '''left''' def __init__( self , lowerCamelCase_ , lowerCamelCase_=False , lowerCamelCase_=True , lowerCamelCase_=False , lowerCamelCase_="<s>" , lowerCamelCase_="</s>" , lowerCamelCase_="<unk>" , lowerCamelCase_="<sep>" , lowerCamelCase_="<pad>" , lowerCamelCase_="<cls>" , lowerCamelCase_="<mask>" , lowerCamelCase_=["<eop>", "<eod>"] , lowerCamelCase_ = None , **lowerCamelCase_ , ) -> None: # Mask token behave like a normal word, i.e. include the space before it UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_) if isinstance(lowerCamelCase_ , lowerCamelCase_) else mask_token UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=lowerCamelCase_ , remove_space=lowerCamelCase_ , keep_accents=lowerCamelCase_ , bos_token=lowerCamelCase_ , eos_token=lowerCamelCase_ , unk_token=lowerCamelCase_ , sep_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , cls_token=lowerCamelCase_ , mask_token=lowerCamelCase_ , additional_special_tokens=lowerCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase_ , ) UpperCamelCase = 3 UpperCamelCase = do_lower_case UpperCamelCase = remove_space UpperCamelCase = keep_accents UpperCamelCase = vocab_file UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(lowerCamelCase_) @property def UpperCAmelCase__ ( self) -> List[str]: return len(self.sp_model) def UpperCAmelCase__ ( self) -> Tuple: UpperCamelCase = {self.convert_ids_to_tokens(lowerCamelCase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def __getstate__( self) -> Any: UpperCamelCase = self.__dict__.copy() UpperCamelCase = None return state def __setstate__( self , lowerCamelCase_) -> str: UpperCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs'''): UpperCamelCase = {} UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(self.vocab_file) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Union[str, Any]: if self.remove_space: UpperCamelCase = ''' '''.join(inputs.strip().split()) else: UpperCamelCase = inputs UpperCamelCase = outputs.replace('''``''' , '''"''').replace('''\'\'''' , '''"''') if not self.keep_accents: UpperCamelCase = unicodedata.normalize('''NFKD''' , lowerCamelCase_) UpperCamelCase = ''''''.join([c for c in outputs if not unicodedata.combining(lowerCamelCase_)]) if self.do_lower_case: UpperCamelCase = outputs.lower() return outputs def UpperCAmelCase__ ( self , lowerCamelCase_) -> List[str]: UpperCamelCase = self.preprocess_text(lowerCamelCase_) UpperCamelCase = self.sp_model.encode(lowerCamelCase_ , out_type=lowerCamelCase_) UpperCamelCase = [] for piece in pieces: if len(lowerCamelCase_) > 1 and piece[-1] == str(''',''') and piece[-2].isdigit(): UpperCamelCase = self.sp_model.EncodeAsPieces(piece[:-1].replace(lowerCamelCase_ , '''''')) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0]) == 1: UpperCamelCase = cur_pieces[1:] else: UpperCamelCase = cur_pieces[0][1:] cur_pieces.append(piece[-1]) new_pieces.extend(lowerCamelCase_) else: new_pieces.append(lowerCamelCase_) return new_pieces def UpperCAmelCase__ ( self , lowerCamelCase_) -> int: return self.sp_model.PieceToId(lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Optional[int]: return self.sp_model.IdToPiece(lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Dict: UpperCamelCase = ''''''.join(lowerCamelCase_).replace(lowerCamelCase_ , ''' ''').strip() return out_string def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = False , lowerCamelCase_ = None , lowerCamelCase_ = True , **lowerCamelCase_ , ) -> str: UpperCamelCase = kwargs.pop('''use_source_tokenizer''' , lowerCamelCase_) UpperCamelCase = self.convert_ids_to_tokens(lowerCamelCase_ , skip_special_tokens=lowerCamelCase_) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 UpperCamelCase = [] UpperCamelCase = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(lowerCamelCase_)) UpperCamelCase = [] sub_texts.append(lowerCamelCase_) else: current_sub_text.append(lowerCamelCase_) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(lowerCamelCase_)) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens UpperCamelCase = ''''''.join(lowerCamelCase_) UpperCamelCase = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: UpperCamelCase = self.clean_up_tokenization(lowerCamelCase_) return clean_text else: return text def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> List[int]: UpperCamelCase = [self.sep_token_id] UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None , lowerCamelCase_ = False) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ , token_ids_a=lowerCamelCase_ , already_has_special_tokens=lowerCamelCase_) if token_ids_a is not None: return ([0] * len(lowerCamelCase_)) + [1] + ([0] * len(lowerCamelCase_)) + [1, 1] return ([0] * len(lowerCamelCase_)) + [1, 1] def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> List[int]: UpperCamelCase = [self.sep_token_id] UpperCamelCase = [2] if token_ids_a is None: return len(token_ids_a + sep) * [0] + cls_segment_id return len(token_ids_a + sep) * [0] + len(token_ids_a + sep) * [1] + cls_segment_id def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> Tuple[str]: if not os.path.isdir(lowerCamelCase_): logger.error(F'Vocabulary path ({save_directory}) should be a directory') return UpperCamelCase = os.path.join( lowerCamelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) if os.path.abspath(self.vocab_file) != os.path.abspath(lowerCamelCase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , lowerCamelCase_) elif not os.path.isfile(self.vocab_file): with open(lowerCamelCase_ , '''wb''') as fi: UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase_) return (out_vocab_file,)
34
1
"""simple docstring""" import gc import unittest from parameterized import parameterized from diffusers import FlaxUNetaDConditionModel from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import load_hf_numpy, require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp @slow @require_flax class snake_case_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_) -> str: return F'gaussian_noise_s={seed}_shape={"_".join([str(lowerCamelCase_) for s in shape])}.npy' def UpperCAmelCase__ ( self) -> List[str]: # clean up the VRAM after each test super().tearDown() gc.collect() def UpperCAmelCase__ ( self , lowerCamelCase_=0 , lowerCamelCase_=(4, 4, 6_4, 6_4) , lowerCamelCase_=False) -> Union[str, Any]: UpperCamelCase = jnp.bfloataa if fpaa else jnp.floataa UpperCamelCase = jnp.array(load_hf_numpy(self.get_file_format(lowerCamelCase_ , lowerCamelCase_)) , dtype=lowerCamelCase_) return image def UpperCAmelCase__ ( self , lowerCamelCase_=False , lowerCamelCase_="CompVis/stable-diffusion-v1-4") -> Tuple: UpperCamelCase = jnp.bfloataa if fpaa else jnp.floataa UpperCamelCase = '''bf16''' if fpaa else None UpperCamelCase , UpperCamelCase = FlaxUNetaDConditionModel.from_pretrained( lowerCamelCase_ , subfolder='''unet''' , dtype=lowerCamelCase_ , revision=lowerCamelCase_) return model, params def UpperCAmelCase__ ( self , lowerCamelCase_=0 , lowerCamelCase_=(4, 7_7, 7_6_8) , lowerCamelCase_=False) -> Dict: UpperCamelCase = jnp.bfloataa if fpaa else jnp.floataa UpperCamelCase = jnp.array(load_hf_numpy(self.get_file_format(lowerCamelCase_ , lowerCamelCase_)) , dtype=lowerCamelCase_) return hidden_states @parameterized.expand( [ # fmt: off [8_3, 4, [-0.2323, -0.1304, 0.0813, -0.3093, -0.0919, -0.1571, -0.1125, -0.5806]], [1_7, 0.55, [-0.0831, -0.2443, 0.0901, -0.0919, 0.3396, 0.0103, -0.3743, 0.0701]], [8, 0.89, [-0.4863, 0.0859, 0.0875, -0.1658, 0.9199, -0.0114, 0.4839, 0.4639]], [3, 1_0_0_0, [-0.5649, 0.2402, -0.5518, 0.1248, 1.1328, -0.2443, -0.0325, -1.0078]], # fmt: on ]) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_) -> Union[str, Any]: UpperCamelCase , UpperCamelCase = self.get_unet_model(model_id='''CompVis/stable-diffusion-v1-4''' , fpaa=lowerCamelCase_) UpperCamelCase = self.get_latents(lowerCamelCase_ , fpaa=lowerCamelCase_) UpperCamelCase = self.get_encoder_hidden_states(lowerCamelCase_ , fpaa=lowerCamelCase_) UpperCamelCase = model.apply( {'''params''': params} , lowerCamelCase_ , jnp.array(lowerCamelCase_ , dtype=jnp.intaa) , encoder_hidden_states=lowerCamelCase_ , ).sample assert sample.shape == latents.shape UpperCamelCase = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten())) , dtype=jnp.floataa) UpperCamelCase = jnp.array(lowerCamelCase_ , dtype=jnp.floataa) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, in the same hardware assert jnp.allclose(lowerCamelCase_ , lowerCamelCase_ , atol=1e-2) @parameterized.expand( [ # fmt: off [8_3, 4, [0.1514, 0.0807, 0.1624, 0.1016, -0.1896, 0.0263, 0.0677, 0.2310]], [1_7, 0.55, [0.1164, -0.0216, 0.0170, 0.1589, -0.3120, 0.1005, -0.0581, -0.1458]], [8, 0.89, [-0.1758, -0.0169, 0.1004, -0.1411, 0.1312, 0.1103, -0.1996, 0.2139]], [3, 1_0_0_0, [0.1214, 0.0352, -0.0731, -0.1562, -0.0994, -0.0906, -0.2340, -0.0539]], # fmt: on ]) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_) -> Optional[int]: UpperCamelCase , UpperCamelCase = self.get_unet_model(model_id='''stabilityai/stable-diffusion-2''' , fpaa=lowerCamelCase_) UpperCamelCase = self.get_latents(lowerCamelCase_ , shape=(4, 4, 9_6, 9_6) , fpaa=lowerCamelCase_) UpperCamelCase = self.get_encoder_hidden_states(lowerCamelCase_ , shape=(4, 7_7, 1_0_2_4) , fpaa=lowerCamelCase_) UpperCamelCase = model.apply( {'''params''': params} , lowerCamelCase_ , jnp.array(lowerCamelCase_ , dtype=jnp.intaa) , encoder_hidden_states=lowerCamelCase_ , ).sample assert sample.shape == latents.shape UpperCamelCase = jnp.asarray(jax.device_get((sample[-1, -2:, -2:, :2].flatten())) , dtype=jnp.floataa) UpperCamelCase = jnp.array(lowerCamelCase_ , dtype=jnp.floataa) # Found torch (float16) and flax (bfloat16) outputs to be within this tolerance, on the same hardware assert jnp.allclose(lowerCamelCase_ , lowerCamelCase_ , atol=1e-2)
34
"""simple docstring""" import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = {'vocab_file': 'vocab.txt'} SCREAMING_SNAKE_CASE_ = { 'vocab_file': { 'openbmb/cpm-ant-10b': 'https://huggingface.co/openbmb/cpm-ant-10b/blob/main/vocab.txt', }, } SCREAMING_SNAKE_CASE_ = { 'openbmb/cpm-ant-10b': 1024, } def __snake_case ( _lowercase ): """simple docstring""" UpperCamelCase = collections.OrderedDict() with open(_lowercase ,'''r''' ,encoding='''utf-8''' ) as reader: UpperCamelCase = reader.readlines() for index, token in enumerate(_lowercase ): UpperCamelCase = token.rstrip('''\n''' ) UpperCamelCase = index return vocab class snake_case_ ( lowerCamelCase_ ): """simple docstring""" def __init__( self , lowerCamelCase_ , lowerCamelCase_="<unk>" , lowerCamelCase_=2_0_0) -> Any: UpperCamelCase = vocab UpperCamelCase = unk_token UpperCamelCase = max_input_chars_per_word def UpperCAmelCase__ ( self , lowerCamelCase_) -> Union[str, Any]: UpperCamelCase = list(lowerCamelCase_) if len(lowerCamelCase_) > self.max_input_chars_per_word: return [self.unk_token] UpperCamelCase = 0 UpperCamelCase = [] while start < len(lowerCamelCase_): UpperCamelCase = len(lowerCamelCase_) UpperCamelCase = None while start < end: UpperCamelCase = ''''''.join(chars[start:end]) if substr in self.vocab: UpperCamelCase = substr break end -= 1 if cur_substr is None: sub_tokens.append(self.unk_token) start += 1 else: sub_tokens.append(lowerCamelCase_) UpperCamelCase = end return sub_tokens class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = VOCAB_FILES_NAMES A_ = PRETRAINED_VOCAB_FILES_MAP A_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ = ['''input_ids''', '''attention_mask'''] A_ = False def __init__( self , lowerCamelCase_ , lowerCamelCase_="<d>" , lowerCamelCase_="</d>" , lowerCamelCase_="<s>" , lowerCamelCase_="</s>" , lowerCamelCase_="<pad>" , lowerCamelCase_="<unk>" , lowerCamelCase_="</n>" , lowerCamelCase_="</_>" , lowerCamelCase_="left" , **lowerCamelCase_ , ) -> List[str]: requires_backends(self , ['''jieba''']) super().__init__( bod_token=lowerCamelCase_ , eod_token=lowerCamelCase_ , bos_token=lowerCamelCase_ , eos_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , unk_token=lowerCamelCase_ , line_token=lowerCamelCase_ , space_token=lowerCamelCase_ , padding_side=lowerCamelCase_ , **lowerCamelCase_ , ) UpperCamelCase = bod_token UpperCamelCase = eod_token UpperCamelCase = load_vocab(lowerCamelCase_) UpperCamelCase = self.encoder[space_token] UpperCamelCase = self.encoder[line_token] del self.encoder[space_token] del self.encoder[line_token] UpperCamelCase = collections.OrderedDict(sorted(self.encoder.items() , key=lambda lowerCamelCase_: x[1])) UpperCamelCase = {v: k for k, v in self.encoder.items()} UpperCamelCase = WordpieceTokenizer(vocab=self.encoder , unk_token=self.unk_token) @property def UpperCAmelCase__ ( self) -> Dict: return self.encoder[self.bod_token] @property def UpperCAmelCase__ ( self) -> str: return self.encoder[self.eod_token] @property def UpperCAmelCase__ ( self) -> List[Any]: return self.encoder["\n"] @property def UpperCAmelCase__ ( self) -> int: return len(self.encoder) def UpperCAmelCase__ ( self) -> Dict: return dict(self.encoder , **self.added_tokens_encoder) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Any: UpperCamelCase = [] for x in jieba.cut(lowerCamelCase_ , cut_all=lowerCamelCase_): output_tokens.extend(self.wordpiece_tokenizer.tokenize(lowerCamelCase_)) return output_tokens def UpperCAmelCase__ ( self , lowerCamelCase_ , **lowerCamelCase_) -> Tuple: UpperCamelCase = [i for i in token_ids if i >= 0] UpperCamelCase = [ x for x in token_ids if x != self.pad_token_id and x != self.eos_token_id and x != self.bos_token_id ] return super()._decode(lowerCamelCase_ , **lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Dict: return token in self.encoder def UpperCAmelCase__ ( self , lowerCamelCase_) -> str: return "".join(lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Optional[int]: return self.encoder.get(lowerCamelCase_ , self.encoder.get(self.unk_token)) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Dict: return self.decoder.get(lowerCamelCase_ , self.unk_token) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> Tuple[str]: if os.path.isdir(lowerCamelCase_): UpperCamelCase = os.path.join( lowerCamelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) else: UpperCamelCase = (filename_prefix + '''-''' if filename_prefix else '''''') + save_directory UpperCamelCase = 0 if " " in self.encoder: UpperCamelCase = self.encoder[''' '''] del self.encoder[" "] if "\n" in self.encoder: UpperCamelCase = self.encoder['''\n'''] del self.encoder["\n"] UpperCamelCase = collections.OrderedDict(sorted(self.encoder.items() , key=lambda lowerCamelCase_: x[1])) with open(lowerCamelCase_ , '''w''' , encoding='''utf-8''') as writer: for token, token_index in self.encoder.items(): if index != token_index: logger.warning( F'Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.' ''' Please check that the vocabulary is not corrupted!''') UpperCamelCase = token_index writer.write(token + '''\n''') index += 1 return (vocab_file,) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> List[int]: if token_ids_a is None: return [self.bos_token_id] + token_ids_a return [self.bos_token_id] + token_ids_a + [self.bos_token_id] + token_ids_a def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None , lowerCamelCase_ = False) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ , token_ids_a=lowerCamelCase_ , already_has_special_tokens=lowerCamelCase_) if token_ids_a is not None: return [1] + ([0] * len(lowerCamelCase_)) + [1] + ([0] * len(lowerCamelCase_)) return [1] + ([0] * len(lowerCamelCase_))
34
1
"""simple docstring""" import numpy as np def __snake_case ( _lowercase ): """simple docstring""" return 1 / (1 + np.exp(-vector )) def __snake_case ( _lowercase ): """simple docstring""" return vector * sigmoid(1.702 * vector ) if __name__ == "__main__": import doctest doctest.testmod()
34
"""simple docstring""" from typing import Callable, Dict, Optional, Tuple import torch from torch import nn from torch.distributions import ( AffineTransform, Distribution, Independent, NegativeBinomial, Normal, StudentT, TransformedDistribution, ) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" def __init__( self , lowerCamelCase_ , lowerCamelCase_=None , lowerCamelCase_=None , lowerCamelCase_=0) -> int: UpperCamelCase = 1.0 if scale is None else scale UpperCamelCase = 0.0 if loc is None else loc super().__init__(lowerCamelCase_ , [AffineTransform(loc=self.loc , scale=self.scale , event_dim=lowerCamelCase_)]) @property def UpperCAmelCase__ ( self) -> List[Any]: return self.base_dist.mean * self.scale + self.loc @property def UpperCAmelCase__ ( self) -> List[str]: return self.base_dist.variance * self.scale**2 @property def UpperCAmelCase__ ( self) -> Any: return self.variance.sqrt() class snake_case_ ( nn.Module ): """simple docstring""" def __init__( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ , **lowerCamelCase_) -> None: super().__init__(**lowerCamelCase_) UpperCamelCase = args_dim UpperCamelCase = nn.ModuleList([nn.Linear(lowerCamelCase_ , lowerCamelCase_) for dim in args_dim.values()]) UpperCamelCase = domain_map def UpperCAmelCase__ ( self , lowerCamelCase_) -> Tuple[torch.Tensor]: UpperCamelCase = [proj(lowerCamelCase_) for proj in self.proj] return self.domain_map(*lowerCamelCase_) class snake_case_ ( nn.Module ): """simple docstring""" def __init__( self , lowerCamelCase_) -> int: super().__init__() UpperCamelCase = function def UpperCAmelCase__ ( self , lowerCamelCase_ , *lowerCamelCase_) -> Tuple: return self.function(lowerCamelCase_ , *lowerCamelCase_) class snake_case_ : """simple docstring""" A_ = 42 A_ = 42 A_ = 42 def __init__( self , lowerCamelCase_ = 1) -> None: UpperCamelCase = dim UpperCamelCase = {k: dim * self.args_dim[k] for k in self.args_dim} def UpperCAmelCase__ ( self , lowerCamelCase_) -> str: if self.dim == 1: return self.distribution_class(*lowerCamelCase_) else: return Independent(self.distribution_class(*lowerCamelCase_) , 1) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None , lowerCamelCase_ = None , ) -> Distribution: UpperCamelCase = self._base_distribution(lowerCamelCase_) if loc is None and scale is None: return distr else: return AffineTransformed(lowerCamelCase_ , loc=lowerCamelCase_ , scale=lowerCamelCase_ , event_dim=self.event_dim) @property def UpperCAmelCase__ ( self) -> Tuple: return () if self.dim == 1 else (self.dim,) @property def UpperCAmelCase__ ( self) -> int: return len(self.event_shape) @property def UpperCAmelCase__ ( self) -> float: return 0.0 def UpperCAmelCase__ ( self , lowerCamelCase_) -> nn.Module: return ParameterProjection( in_features=lowerCamelCase_ , args_dim=self.args_dim , domain_map=LambdaLayer(self.domain_map) , ) def UpperCAmelCase__ ( self , *lowerCamelCase_) -> List[str]: raise NotImplementedError() @staticmethod def UpperCAmelCase__ ( lowerCamelCase_) -> torch.Tensor: return (x + torch.sqrt(torch.square(lowerCamelCase_) + 4.0)) / 2.0 class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = {"df": 1, "loc": 1, "scale": 1} A_ = StudentT @classmethod def UpperCAmelCase__ ( cls , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_) -> Optional[int]: UpperCamelCase = cls.squareplus(lowerCamelCase_).clamp_min(torch.finfo(scale.dtype).eps) UpperCamelCase = 2.0 + cls.squareplus(lowerCamelCase_) return df.squeeze(-1), loc.squeeze(-1), scale.squeeze(-1) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = {"loc": 1, "scale": 1} A_ = Normal @classmethod def UpperCAmelCase__ ( cls , lowerCamelCase_ , lowerCamelCase_) -> str: UpperCamelCase = cls.squareplus(lowerCamelCase_).clamp_min(torch.finfo(scale.dtype).eps) return loc.squeeze(-1), scale.squeeze(-1) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = {"total_count": 1, "logits": 1} A_ = NegativeBinomial @classmethod def UpperCAmelCase__ ( cls , lowerCamelCase_ , lowerCamelCase_) -> List[Any]: UpperCamelCase = cls.squareplus(lowerCamelCase_) return total_count.squeeze(-1), logits.squeeze(-1) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Distribution: UpperCamelCase , UpperCamelCase = distr_args if self.dim == 1: return self.distribution_class(total_count=lowerCamelCase_ , logits=lowerCamelCase_) else: return Independent(self.distribution_class(total_count=lowerCamelCase_ , logits=lowerCamelCase_) , 1) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None , lowerCamelCase_ = None) -> Distribution: UpperCamelCase , UpperCamelCase = distr_args if scale is not None: # See scaling property of Gamma. logits += scale.log() return self._base_distribution((total_count, logits))
34
1
"""simple docstring""" import re import string import numpy as np import datasets SCREAMING_SNAKE_CASE_ = '\nReturns the rate at which the input predicted strings exactly match their references, ignoring any strings input as part of the regexes_to_ignore list.\n' SCREAMING_SNAKE_CASE_ = '\nArgs:\n predictions: List of predicted texts.\n references: List of reference texts.\n regexes_to_ignore: List, defaults to None. Regex expressions of characters to\n ignore when calculating the exact matches. Note: these regexes are removed\n from the input data before the changes based on the options below (e.g. ignore_case,\n ignore_punctuation, ignore_numbers) are applied.\n ignore_case: Boolean, defaults to False. If true, turns everything\n to lowercase so that capitalization differences are ignored.\n ignore_punctuation: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\n ignore_numbers: Boolean, defaults to False. If true, removes all punctuation before\n comparing predictions and references.\nReturns:\n exact_match: Dictionary containing exact_match rate. Possible values are between 0.0 and 100.0, inclusive.\nExamples:\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results["exact_match"], 1))\n 25.0\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results["exact_match"], 1))\n 50.0\n\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell", "YELL"], ignore_case=True, ignore_punctuation=True)\n >>> print(round(results["exact_match"], 1))\n 75.0\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["the cat", "theater", "YELLING", "agent007"]\n >>> preds = ["cat?", "theater", "yelling", "agent"]\n >>> results = exact_match.compute(references=refs, predictions=preds, regexes_to_ignore=["the ", "yell", "YELL"], ignore_case=True, ignore_punctuation=True, ignore_numbers=True)\n >>> print(round(results["exact_match"], 1))\n 100.0\n\n >>> exact_match = datasets.load_metric("exact_match")\n >>> refs = ["The cat sat on the mat.", "Theaters are great.", "It\'s like comparing oranges and apples."]\n >>> preds = ["The cat sat on the mat?", "Theaters are great.", "It\'s like comparing apples and oranges."]\n >>> results = exact_match.compute(references=refs, predictions=preds)\n >>> print(round(results["exact_match"], 1))\n 33.3\n\n' SCREAMING_SNAKE_CASE_ = '\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class snake_case_ ( datasets.Metric ): """simple docstring""" def UpperCAmelCase__ ( self) -> str: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence'''), '''references''': datasets.Value('''string''' , id='''sequence'''), }) , reference_urls=[] , ) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_=None , lowerCamelCase_=False , lowerCamelCase_=False , lowerCamelCase_=False , ) -> Any: if regexes_to_ignore is not None: for s in regexes_to_ignore: UpperCamelCase = np.array([re.sub(lowerCamelCase_ , '''''' , lowerCamelCase_) for x in predictions]) UpperCamelCase = np.array([re.sub(lowerCamelCase_ , '''''' , lowerCamelCase_) for x in references]) else: UpperCamelCase = np.asarray(lowerCamelCase_) UpperCamelCase = np.asarray(lowerCamelCase_) if ignore_case: UpperCamelCase = np.char.lower(lowerCamelCase_) UpperCamelCase = np.char.lower(lowerCamelCase_) if ignore_punctuation: UpperCamelCase = string.punctuation.maketrans('''''' , '''''' , string.punctuation) UpperCamelCase = np.char.translate(lowerCamelCase_ , table=lowerCamelCase_) UpperCamelCase = np.char.translate(lowerCamelCase_ , table=lowerCamelCase_) if ignore_numbers: UpperCamelCase = string.digits.maketrans('''''' , '''''' , string.digits) UpperCamelCase = np.char.translate(lowerCamelCase_ , table=lowerCamelCase_) UpperCamelCase = np.char.translate(lowerCamelCase_ , table=lowerCamelCase_) UpperCamelCase = predictions == references return {"exact_match": np.mean(lowerCamelCase_) * 1_0_0}
34
"""simple docstring""" # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. SCREAMING_SNAKE_CASE_ = abspath(join(dirname(dirname(__file__)), 'src')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='ignore', category=FutureWarning) def __snake_case ( _lowercase ): """simple docstring""" from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(_lowercase ) def __snake_case ( _lowercase ): """simple docstring""" from diffusers.utils.testing_utils import pytest_terminal_summary_main UpperCamelCase = terminalreporter.config.getoption('''--make-reports''' ) if make_reports: pytest_terminal_summary_main(_lowercase ,id=_lowercase )
34
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" def __init__( self , *lowerCamelCase_ , **lowerCamelCase_) -> None: warnings.warn( '''The class MobileViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use MobileViTImageProcessor instead.''' , lowerCamelCase_ , ) super().__init__(*lowerCamelCase_ , **lowerCamelCase_)
34
"""simple docstring""" import warnings from ...utils import logging from .image_processing_mobilevit import MobileViTImageProcessor SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" def __init__( self , *lowerCamelCase_ , **lowerCamelCase_) -> None: warnings.warn( '''The class MobileViTFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use MobileViTImageProcessor instead.''' , lowerCamelCase_ , ) super().__init__(*lowerCamelCase_ , **lowerCamelCase_)
34
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available SCREAMING_SNAKE_CASE_ = { 'configuration_data2vec_audio': ['DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecAudioConfig'], 'configuration_data2vec_text': [ 'DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecTextConfig', 'Data2VecTextOnnxConfig', ], 'configuration_data2vec_vision': [ 'DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecVisionConfig', 'Data2VecVisionOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: SCREAMING_SNAKE_CASE_ = [ 'DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecAudioForAudioFrameClassification', 'Data2VecAudioForCTC', 'Data2VecAudioForSequenceClassification', 'Data2VecAudioForXVector', 'Data2VecAudioModel', 'Data2VecAudioPreTrainedModel', ] SCREAMING_SNAKE_CASE_ = [ 'DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecTextForCausalLM', 'Data2VecTextForMaskedLM', 'Data2VecTextForMultipleChoice', 'Data2VecTextForQuestionAnswering', 'Data2VecTextForSequenceClassification', 'Data2VecTextForTokenClassification', 'Data2VecTextModel', 'Data2VecTextPreTrainedModel', ] SCREAMING_SNAKE_CASE_ = [ 'DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecVisionForImageClassification', 'Data2VecVisionForMaskedImageModeling', 'Data2VecVisionForSemanticSegmentation', 'Data2VecVisionModel', 'Data2VecVisionPreTrainedModel', ] if is_tf_available(): SCREAMING_SNAKE_CASE_ = [ 'TFData2VecVisionForImageClassification', 'TFData2VecVisionForSemanticSegmentation', 'TFData2VecVisionModel', 'TFData2VecVisionPreTrainedModel', ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys SCREAMING_SNAKE_CASE_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
34
"""simple docstring""" def __snake_case ( _lowercase ): """simple docstring""" UpperCamelCase = [0 for i in range(len(_lowercase ) )] # initialize interval's left pointer and right pointer UpperCamelCase , UpperCamelCase = 0, 0 for i in range(1 ,len(_lowercase ) ): # case when current index is inside the interval if i <= right_pointer: UpperCamelCase = min(right_pointer - i + 1 ,z_result[i - left_pointer] ) UpperCamelCase = min_edge while go_next(_lowercase ,_lowercase ,_lowercase ): z_result[i] += 1 # if new index's result gives us more right interval, # we've to update left_pointer and right_pointer if i + z_result[i] - 1 > right_pointer: UpperCamelCase , UpperCamelCase = i, i + z_result[i] - 1 return z_result def __snake_case ( _lowercase ,_lowercase ,_lowercase ): """simple docstring""" return i + z_result[i] < len(_lowercase ) and s[z_result[i]] == s[i + z_result[i]] def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" UpperCamelCase = 0 # concatenate 'pattern' and 'input_str' and call z_function # with concatenated string UpperCamelCase = z_function(pattern + input_str ) for val in z_result: # if value is greater then length of the pattern string # that means this index is starting position of substring # which is equal to pattern string if val >= len(_lowercase ): answer += 1 return answer if __name__ == "__main__": import doctest doctest.testmod()
34
1
"""simple docstring""" import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class snake_case_ ( lowerCamelCase_ , unittest.TestCase ): """simple docstring""" A_ = CTRLTokenizer A_ = False A_ = False def UpperCAmelCase__ ( self) -> Tuple: super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt UpperCamelCase = ['''adapt''', '''re@@''', '''a@@''', '''apt''', '''c@@''', '''t''', '''<unk>'''] UpperCamelCase = dict(zip(lowerCamelCase_ , range(len(lowerCamelCase_)))) UpperCamelCase = ['''#version: 0.2''', '''a p''', '''ap t</w>''', '''r e''', '''a d''', '''ad apt</w>''', ''''''] UpperCamelCase = {'''unk_token''': '''<unk>'''} UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''vocab_file''']) UpperCamelCase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['''merges_file''']) with open(self.vocab_file , '''w''' , encoding='''utf-8''') as fp: fp.write(json.dumps(lowerCamelCase_) + '''\n''') with open(self.merges_file , '''w''' , encoding='''utf-8''') as fp: fp.write('''\n'''.join(lowerCamelCase_)) def UpperCAmelCase__ ( self , **lowerCamelCase_) -> List[str]: kwargs.update(self.special_tokens_map) return CTRLTokenizer.from_pretrained(self.tmpdirname , **lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Union[str, Any]: UpperCamelCase = '''adapt react readapt apt''' UpperCamelCase = '''adapt react readapt apt''' return input_text, output_text def UpperCAmelCase__ ( self) -> str: UpperCamelCase = CTRLTokenizer(self.vocab_file , self.merges_file , **self.special_tokens_map) UpperCamelCase = '''adapt react readapt apt''' UpperCamelCase = '''adapt re@@ a@@ c@@ t re@@ adapt apt'''.split() UpperCamelCase = tokenizer.tokenize(lowerCamelCase_) self.assertListEqual(lowerCamelCase_ , lowerCamelCase_) UpperCamelCase = tokens + [tokenizer.unk_token] UpperCamelCase = [0, 1, 2, 4, 5, 1, 0, 3, 6] self.assertListEqual(tokenizer.convert_tokens_to_ids(lowerCamelCase_) , lowerCamelCase_)
34
"""simple docstring""" import importlib.metadata import warnings from copy import deepcopy from packaging import version from ..utils import logging from .import_utils import is_accelerate_available, is_bitsandbytes_available if is_bitsandbytes_available(): import bitsandbytes as bnb import torch import torch.nn as nn from ..pytorch_utils import ConvaD if is_accelerate_available(): from accelerate import init_empty_weights from accelerate.utils import find_tied_parameters SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) def __snake_case ( _lowercase ,_lowercase ,_lowercase ,_lowercase=None ,_lowercase=None ): """simple docstring""" if "." in tensor_name: UpperCamelCase = tensor_name.split('''.''' ) for split in splits[:-1]: UpperCamelCase = getattr(_lowercase ,_lowercase ) if new_module is None: raise ValueError(f'{module} has no attribute {split}.' ) UpperCamelCase = new_module UpperCamelCase = splits[-1] if tensor_name not in module._parameters and tensor_name not in module._buffers: raise ValueError(f'{module} does not have a parameter or a buffer named {tensor_name}.' ) UpperCamelCase = tensor_name in module._buffers UpperCamelCase = getattr(_lowercase ,_lowercase ) if old_value.device == torch.device('''meta''' ) and device not in ["meta", torch.device('''meta''' )] and value is None: raise ValueError(f'{tensor_name} is on the meta device, we need a `value` to put in on {device}.' ) UpperCamelCase = False UpperCamelCase = False if is_buffer or not is_bitsandbytes_available(): UpperCamelCase = False UpperCamelCase = False else: UpperCamelCase = hasattr(bnb.nn ,'''Params4bit''' ) and isinstance(module._parameters[tensor_name] ,bnb.nn.Paramsabit ) UpperCamelCase = isinstance(module._parameters[tensor_name] ,bnb.nn.IntaParams ) if is_abit or is_abit: UpperCamelCase = module._parameters[tensor_name] if param.device.type != "cuda": if value is None: UpperCamelCase = old_value.to(_lowercase ) elif isinstance(_lowercase ,torch.Tensor ): UpperCamelCase = value.to('''cpu''' ) if value.dtype == torch.inta: UpperCamelCase = version.parse(importlib.metadata.version('''bitsandbytes''' ) ) > version.parse( '''0.37.2''' ) if not is_abit_serializable: raise ValueError( '''Detected int8 weights but the version of bitsandbytes is not compatible with int8 serialization. ''' '''Make sure to download the latest `bitsandbytes` version. `pip install --upgrade bitsandbytes`.''' ) else: UpperCamelCase = torch.tensor(_lowercase ,device='''cpu''' ) # Support models using `Conv1D` in place of `nn.Linear` (e.g. gpt2) by transposing the weight matrix prior to quantization. # Since weights are saved in the correct "orientation", we skip transposing when loading. if issubclass(module.source_cls ,_lowercase ) and fpaa_statistics is None: UpperCamelCase = new_value.T UpperCamelCase = old_value.__dict__ if is_abit: UpperCamelCase = bnb.nn.IntaParams(_lowercase ,requires_grad=_lowercase ,**_lowercase ).to(_lowercase ) elif is_abit: UpperCamelCase = bnb.nn.Paramsabit(_lowercase ,requires_grad=_lowercase ,**_lowercase ).to(_lowercase ) UpperCamelCase = new_value if fpaa_statistics is not None: setattr(module.weight ,'''SCB''' ,fpaa_statistics.to(_lowercase ) ) else: if value is None: UpperCamelCase = old_value.to(_lowercase ) elif isinstance(_lowercase ,torch.Tensor ): UpperCamelCase = value.to(_lowercase ) else: UpperCamelCase = torch.tensor(_lowercase ,device=_lowercase ) if is_buffer: UpperCamelCase = new_value else: UpperCamelCase = nn.Parameter(_lowercase ,requires_grad=old_value.requires_grad ) UpperCamelCase = new_value def __snake_case ( _lowercase ,_lowercase=None ,_lowercase=None ,_lowercase=None ,_lowercase=False ): """simple docstring""" for name, module in model.named_children(): if current_key_name is None: UpperCamelCase = [] current_key_name.append(_lowercase ) if (isinstance(_lowercase ,nn.Linear ) or isinstance(_lowercase ,_lowercase )) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` if not any(key in '''.'''.join(_lowercase ) for key in modules_to_not_convert ): with init_empty_weights(): if isinstance(_lowercase ,_lowercase ): UpperCamelCase , UpperCamelCase = module.weight.shape else: UpperCamelCase = module.in_features UpperCamelCase = module.out_features if quantization_config.quantization_method() == "llm_int8": UpperCamelCase = bnb.nn.LinearabitLt( _lowercase ,_lowercase ,module.bias is not None ,has_fpaa_weights=quantization_config.llm_inta_has_fpaa_weight ,threshold=quantization_config.llm_inta_threshold ,) UpperCamelCase = True else: if ( quantization_config.llm_inta_skip_modules is not None and name in quantization_config.llm_inta_skip_modules ): pass else: UpperCamelCase = bnb.nn.Linearabit( _lowercase ,_lowercase ,module.bias is not None ,quantization_config.bnb_abit_compute_dtype ,compress_statistics=quantization_config.bnb_abit_use_double_quant ,quant_type=quantization_config.bnb_abit_quant_type ,) UpperCamelCase = True # Store the module class in case we need to transpose the weight later UpperCamelCase = type(_lowercase ) # Force requires grad to False to avoid unexpected errors model._modules[name].requires_grad_(_lowercase ) if len(list(module.children() ) ) > 0: UpperCamelCase , UpperCamelCase = _replace_with_bnb_linear( _lowercase ,_lowercase ,_lowercase ,_lowercase ,has_been_replaced=_lowercase ,) # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def __snake_case ( _lowercase ,_lowercase=None ,_lowercase=None ,_lowercase=None ): """simple docstring""" UpperCamelCase = ['''lm_head'''] if modules_to_not_convert is None else modules_to_not_convert UpperCamelCase , UpperCamelCase = _replace_with_bnb_linear( _lowercase ,_lowercase ,_lowercase ,_lowercase ) if not has_been_replaced: logger.warning( '''You are loading your model in 8bit or 4bit but no linear modules were found in your model.''' ''' Please double check your model architecture, or submit an issue on github if you think this is''' ''' a bug.''' ) return model def __snake_case ( *_lowercase ,**_lowercase ): """simple docstring""" warnings.warn( '''`replace_8bit_linear` will be deprecated in a future version, please use `replace_with_bnb_linear` instead''' ,_lowercase ,) return replace_with_bnb_linear(*_lowercase ,**_lowercase ) def __snake_case ( *_lowercase ,**_lowercase ): """simple docstring""" warnings.warn( '''`set_module_8bit_tensor_to_device` will be deprecated in a future version, please use `set_module_quantized_tensor_to_device` instead''' ,_lowercase ,) return set_module_quantized_tensor_to_device(*_lowercase ,**_lowercase ) def __snake_case ( _lowercase ): """simple docstring""" UpperCamelCase = deepcopy(_lowercase ) # this has 0 cost since it is done inside `init_empty_weights` context manager` tied_model.tie_weights() UpperCamelCase = find_tied_parameters(_lowercase ) # For compatibility with Accelerate < 0.18 if isinstance(_lowercase ,_lowercase ): UpperCamelCase = sum(list(tied_params.values() ) ,[] ) + list(tied_params.keys() ) else: UpperCamelCase = sum(_lowercase ,[] ) UpperCamelCase = len(_lowercase ) > 0 # Check if it is a base model UpperCamelCase = not hasattr(_lowercase ,model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head UpperCamelCase = list(model.named_children() ) UpperCamelCase = [list_modules[-1][0]] # add last module together with tied weights UpperCamelCase = set(_lowercase ) - set(_lowercase ) UpperCamelCase = list(set(_lowercase ) ) + list(_lowercase ) # remove ".weight" from the keys UpperCamelCase = ['''.weight''', '''.bias'''] UpperCamelCase = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: UpperCamelCase = name.replace(_lowercase ,'''''' ) filtered_module_names.append(_lowercase ) return filtered_module_names
34
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = { 'microsoft/swin-tiny-patch4-window7-224': ( 'https://huggingface.co/microsoft/swin-tiny-patch4-window7-224/resolve/main/config.json' ), # See all Swin models at https://huggingface.co/models?filter=swin } class snake_case_ ( lowerCamelCase_ , lowerCamelCase_ ): """simple docstring""" A_ = '''swin''' A_ = { '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers''', } def __init__( self , lowerCamelCase_=2_2_4 , lowerCamelCase_=4 , lowerCamelCase_=3 , lowerCamelCase_=9_6 , lowerCamelCase_=[2, 2, 6, 2] , lowerCamelCase_=[3, 6, 1_2, 2_4] , lowerCamelCase_=7 , lowerCamelCase_=4.0 , lowerCamelCase_=True , lowerCamelCase_=0.0 , lowerCamelCase_=0.0 , lowerCamelCase_=0.1 , lowerCamelCase_="gelu" , lowerCamelCase_=False , lowerCamelCase_=0.02 , lowerCamelCase_=1e-5 , lowerCamelCase_=3_2 , lowerCamelCase_=None , lowerCamelCase_=None , **lowerCamelCase_ , ) -> Any: super().__init__(**lowerCamelCase_) UpperCamelCase = image_size UpperCamelCase = patch_size UpperCamelCase = num_channels UpperCamelCase = embed_dim UpperCamelCase = depths UpperCamelCase = len(lowerCamelCase_) UpperCamelCase = num_heads UpperCamelCase = window_size UpperCamelCase = mlp_ratio UpperCamelCase = qkv_bias UpperCamelCase = hidden_dropout_prob UpperCamelCase = attention_probs_dropout_prob UpperCamelCase = drop_path_rate UpperCamelCase = hidden_act UpperCamelCase = use_absolute_embeddings UpperCamelCase = layer_norm_eps UpperCamelCase = initializer_range UpperCamelCase = encoder_stride # we set the hidden_size attribute in order to make Swin work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model UpperCamelCase = int(embed_dim * 2 ** (len(lowerCamelCase_) - 1)) UpperCamelCase = ['''stem'''] + [F'stage{idx}' for idx in range(1 , len(lowerCamelCase_) + 1)] UpperCamelCase , UpperCamelCase = get_aligned_output_features_output_indices( out_features=lowerCamelCase_ , out_indices=lowerCamelCase_ , stage_names=self.stage_names) class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = version.parse('''1.11''' ) @property def UpperCAmelCase__ ( self) -> Mapping[str, Mapping[int, str]]: return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ]) @property def UpperCAmelCase__ ( self) -> float: return 1e-4
34
"""simple docstring""" from random import randint from tempfile import TemporaryFile import numpy as np def __snake_case ( _lowercase ,_lowercase ,_lowercase ): """simple docstring""" UpperCamelCase = 0 if start < end: UpperCamelCase = randint(_lowercase ,_lowercase ) UpperCamelCase = a[end] UpperCamelCase = a[pivot] UpperCamelCase = temp UpperCamelCase , UpperCamelCase = _in_place_partition(_lowercase ,_lowercase ,_lowercase ) count += _in_place_quick_sort(_lowercase ,_lowercase ,p - 1 ) count += _in_place_quick_sort(_lowercase ,p + 1 ,_lowercase ) return count def __snake_case ( _lowercase ,_lowercase ,_lowercase ): """simple docstring""" UpperCamelCase = 0 UpperCamelCase = randint(_lowercase ,_lowercase ) UpperCamelCase = a[end] UpperCamelCase = a[pivot] UpperCamelCase = temp UpperCamelCase = start - 1 for index in range(_lowercase ,_lowercase ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value UpperCamelCase = new_pivot_index + 1 UpperCamelCase = a[new_pivot_index] UpperCamelCase = a[index] UpperCamelCase = temp UpperCamelCase = a[new_pivot_index + 1] UpperCamelCase = a[end] UpperCamelCase = temp return new_pivot_index + 1, count SCREAMING_SNAKE_CASE_ = TemporaryFile() SCREAMING_SNAKE_CASE_ = 100 # 1000 elements are to be sorted SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = 0, 1 # mean and standard deviation SCREAMING_SNAKE_CASE_ = np.random.normal(mu, sigma, p) np.save(outfile, X) print('The array is') print(X) outfile.seek(0) # using the same array SCREAMING_SNAKE_CASE_ = np.load(outfile) SCREAMING_SNAKE_CASE_ = len(M) - 1 SCREAMING_SNAKE_CASE_ = _in_place_quick_sort(M, 0, r) print( 'No of Comparisons for 100 elements selected from a standard normal distribution' 'is :' ) print(z)
34
1
"""simple docstring""" import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision import transforms from transformers import BitImageProcessor, FocalNetConfig, FocalNetForImageClassification from transformers.image_utils import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, PILImageResampling def __snake_case ( _lowercase ): """simple docstring""" UpperCamelCase = [2, 2, 6, 2] if '''tiny''' in model_name else [2, 2, 18, 2] UpperCamelCase = True if '''large''' in model_name or '''huge''' in model_name else False UpperCamelCase = True if '''large''' in model_name or '''huge''' in model_name else False UpperCamelCase = True if '''large''' in model_name or '''huge''' in model_name else False if "large" in model_name or "xlarge" in model_name or "huge" in model_name: if "fl3" in model_name: UpperCamelCase = [3, 3, 3, 3] UpperCamelCase = [5, 5, 5, 5] elif "fl4" in model_name: UpperCamelCase = [4, 4, 4, 4] UpperCamelCase = [3, 3, 3, 3] if "tiny" in model_name or "small" in model_name or "base" in model_name: UpperCamelCase = [3, 3, 3, 3] if "lrf" in model_name: UpperCamelCase = [3, 3, 3, 3] else: UpperCamelCase = [2, 2, 2, 2] if "tiny" in model_name: UpperCamelCase = 96 elif "small" in model_name: UpperCamelCase = 96 elif "base" in model_name: UpperCamelCase = 128 elif "large" in model_name: UpperCamelCase = 192 elif "xlarge" in model_name: UpperCamelCase = 256 elif "huge" in model_name: UpperCamelCase = 352 # set label information UpperCamelCase = '''huggingface/label-files''' if "large" in model_name or "huge" in model_name: UpperCamelCase = '''imagenet-22k-id2label.json''' else: UpperCamelCase = '''imagenet-1k-id2label.json''' UpperCamelCase = json.load(open(hf_hub_download(_lowercase ,_lowercase ,repo_type='''dataset''' ) ,'''r''' ) ) UpperCamelCase = {int(_lowercase ): v for k, v in idalabel.items()} UpperCamelCase = {v: k for k, v in idalabel.items()} UpperCamelCase = FocalNetConfig( embed_dim=_lowercase ,depths=_lowercase ,focal_levels=_lowercase ,focal_windows=_lowercase ,use_conv_embed=_lowercase ,idalabel=_lowercase ,labelaid=_lowercase ,use_post_layernorm=_lowercase ,use_layerscale=_lowercase ,) return config def __snake_case ( _lowercase ): """simple docstring""" if "patch_embed.proj" in name: UpperCamelCase = name.replace('''patch_embed.proj''' ,'''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: UpperCamelCase = name.replace('''patch_embed.norm''' ,'''embeddings.norm''' ) if "layers" in name: UpperCamelCase = '''encoder.''' + name if "encoder.layers" in name: UpperCamelCase = name.replace('''encoder.layers''' ,'''encoder.stages''' ) if "downsample.proj" in name: UpperCamelCase = name.replace('''downsample.proj''' ,'''downsample.projection''' ) if "blocks" in name: UpperCamelCase = name.replace('''blocks''' ,'''layers''' ) if "modulation.f.weight" in name or "modulation.f.bias" in name: UpperCamelCase = name.replace('''modulation.f''' ,'''modulation.projection_in''' ) if "modulation.h.weight" in name or "modulation.h.bias" in name: UpperCamelCase = name.replace('''modulation.h''' ,'''modulation.projection_context''' ) if "modulation.proj.weight" in name or "modulation.proj.bias" in name: UpperCamelCase = name.replace('''modulation.proj''' ,'''modulation.projection_out''' ) if name == "norm.weight": UpperCamelCase = '''layernorm.weight''' if name == "norm.bias": UpperCamelCase = '''layernorm.bias''' if "head" in name: UpperCamelCase = name.replace('''head''' ,'''classifier''' ) else: UpperCamelCase = '''focalnet.''' + name return name def __snake_case ( _lowercase ,_lowercase ,_lowercase=False ): """simple docstring""" UpperCamelCase = { '''focalnet-tiny''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_srf.pth''', '''focalnet-tiny-lrf''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_tiny_lrf.pth''', '''focalnet-small''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_srf.pth''', '''focalnet-small-lrf''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_small_lrf.pth''', '''focalnet-base''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_srf.pth''', '''focalnet-base-lrf''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_base_lrf.pth''', '''focalnet-large-lrf-fl3''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384.pth''', '''focalnet-large-lrf-fl4''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_large_lrf_384_fl4.pth''', '''focalnet-xlarge-lrf-fl3''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384.pth''', '''focalnet-xlarge-lrf-fl4''': '''https://projects4jw.blob.core.windows.net/focalnet/release/classification/focalnet_xlarge_lrf_384_fl4.pth''', } # fmt: on UpperCamelCase = model_name_to_url[model_name] print('''Checkpoint URL: ''' ,_lowercase ) UpperCamelCase = torch.hub.load_state_dict_from_url(_lowercase ,map_location='''cpu''' )['''model'''] # rename keys for key in state_dict.copy().keys(): UpperCamelCase = state_dict.pop(_lowercase ) UpperCamelCase = val UpperCamelCase = get_focalnet_config(_lowercase ) UpperCamelCase = FocalNetForImageClassification(_lowercase ) model.eval() # load state dict model.load_state_dict(_lowercase ) # verify conversion UpperCamelCase = '''http://images.cocodataset.org/val2017/000000039769.jpg''' UpperCamelCase = BitImageProcessor( do_resize=_lowercase ,size={'''shortest_edge''': 256} ,resample=PILImageResampling.BILINEAR ,do_center_crop=_lowercase ,crop_size=224 ,do_normalize=_lowercase ,image_mean=_lowercase ,image_std=_lowercase ,) UpperCamelCase = Image.open(requests.get(_lowercase ,stream=_lowercase ).raw ) UpperCamelCase = processor(images=_lowercase ,return_tensors='''pt''' ) UpperCamelCase = transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize(mean=[0.485, 0.456, 0.406] ,std=[0.229, 0.224, 0.225] ), ] ) UpperCamelCase = image_transforms(_lowercase ).unsqueeze(0 ) # verify pixel_values assert torch.allclose(inputs.pixel_values ,_lowercase ,atol=1e-4 ) UpperCamelCase = model(**_lowercase ) UpperCamelCase = outputs.logits.argmax(-1 ).item() print('''Predicted class:''' ,model.config.idalabel[predicted_class_idx] ) print('''First values of logits:''' ,outputs.logits[0, :3] ) if model_name == "focalnet-tiny": UpperCamelCase = torch.tensor([0.2166, -0.4368, 0.2191] ) elif model_name == "focalnet-tiny-lrf": UpperCamelCase = torch.tensor([1.1669, 0.0125, -0.1695] ) elif model_name == "focalnet-small": UpperCamelCase = torch.tensor([0.4917, -0.0430, 0.1341] ) elif model_name == "focalnet-small-lrf": UpperCamelCase = torch.tensor([-0.2588, -0.5342, -0.2331] ) elif model_name == "focalnet-base": UpperCamelCase = torch.tensor([-0.1655, -0.4090, -0.1730] ) elif model_name == "focalnet-base-lrf": UpperCamelCase = torch.tensor([0.5306, -0.0483, -0.3928] ) assert torch.allclose(outputs.logits[0, :3] ,_lowercase ,atol=1e-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(f'Saving model and processor of {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(_lowercase ) processor.save_pretrained(_lowercase ) if push_to_hub: print(f'Pushing model and processor of {model_name} to the hub...' ) model.push_to_hub(f'{model_name}' ) processor.push_to_hub(f'{model_name}' ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='focalnet-tiny', type=str, help='Name of the FocalNet model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) parser.add_argument( '--push_to_hub', action='store_true', help='Whether to push the model and processor to the hub.', ) SCREAMING_SNAKE_CASE_ = parser.parse_args() convert_focalnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
34
"""simple docstring""" import os import sys import unittest SCREAMING_SNAKE_CASE_ = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, 'utils')) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, create_dummy_object, find_backend, read_init # noqa: E402 # Align TRANSFORMERS_PATH in check_dummies with the current path SCREAMING_SNAKE_CASE_ = os.path.join(git_repo_path, 'src', 'transformers') SCREAMING_SNAKE_CASE_ = '\n{0} = None\n' SCREAMING_SNAKE_CASE_ = '\nclass {0}(metaclass=DummyObject):\n _backends = {1}\n\n def __init__(self, *args, **kwargs):\n requires_backends(self, {1})\n' SCREAMING_SNAKE_CASE_ = '\ndef {0}(*args, **kwargs):\n requires_backends({0}, {1})\n' class snake_case_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self) -> List[Any]: UpperCamelCase = find_backend(''' _import_structure["models.albert"].append("AlbertTokenizerFast")''') self.assertIsNone(lowerCamelCase_) UpperCamelCase = find_backend(''' if not is_tokenizers_available():''') self.assertEqual(lowerCamelCase_ , '''tokenizers''') UpperCamelCase = find_backend(''' if not is_tensorflow_text_available():''') self.assertEqual(lowerCamelCase_ , '''tensorflow_text''') UpperCamelCase = find_backend(''' if not (is_sentencepiece_available() and is_tokenizers_available()):''') self.assertEqual(lowerCamelCase_ , '''sentencepiece_and_tokenizers''') UpperCamelCase = find_backend( ''' if not (is_sentencepiece_available() and is_tensorflow_text_available()):''') self.assertEqual(lowerCamelCase_ , '''sentencepiece_and_tensorflow_text''') UpperCamelCase = find_backend( ''' if not (is_sentencepiece_available() and is_tokenizers_available() and is_vision_available()):''') self.assertEqual(lowerCamelCase_ , '''sentencepiece_and_tokenizers_and_vision''') def UpperCAmelCase__ ( self) -> int: UpperCamelCase = read_init() # We don't assert on the exact list of keys to allow for smooth grow of backend-specific objects self.assertIn('''torch''' , lowerCamelCase_) self.assertIn('''tensorflow_text''' , lowerCamelCase_) self.assertIn('''sentencepiece_and_tokenizers''' , lowerCamelCase_) # Likewise, we can't assert on the exact content of a key self.assertIn('''BertModel''' , objects['''torch''']) self.assertIn('''TFBertModel''' , objects['''tf''']) self.assertIn('''FlaxBertModel''' , objects['''flax''']) self.assertIn('''BertModel''' , objects['''torch''']) self.assertIn('''TFBertTokenizer''' , objects['''tensorflow_text''']) self.assertIn('''convert_slow_tokenizer''' , objects['''sentencepiece_and_tokenizers''']) def UpperCAmelCase__ ( self) -> Optional[int]: UpperCamelCase = create_dummy_object('''CONSTANT''' , '''\'torch\'''') self.assertEqual(lowerCamelCase_ , '''\nCONSTANT = None\n''') UpperCamelCase = create_dummy_object('''function''' , '''\'torch\'''') self.assertEqual( lowerCamelCase_ , '''\ndef function(*args, **kwargs):\n requires_backends(function, \'torch\')\n''') UpperCamelCase = ''' class FakeClass(metaclass=DummyObject): _backends = \'torch\' def __init__(self, *args, **kwargs): requires_backends(self, \'torch\') ''' UpperCamelCase = create_dummy_object('''FakeClass''' , '''\'torch\'''') self.assertEqual(lowerCamelCase_ , lowerCamelCase_) def UpperCAmelCase__ ( self) -> int: UpperCamelCase = '''# This file is autogenerated by the command `make fix-copies`, do not edit. from ..utils import DummyObject, requires_backends CONSTANT = None def function(*args, **kwargs): requires_backends(function, ["torch"]) class FakeClass(metaclass=DummyObject): _backends = ["torch"] def __init__(self, *args, **kwargs): requires_backends(self, ["torch"]) ''' UpperCamelCase = create_dummy_files({'''torch''': ['''CONSTANT''', '''function''', '''FakeClass''']}) self.assertEqual(dummy_files['''torch'''] , lowerCamelCase_)
34
1
"""simple docstring""" from decimal import Decimal, getcontext from math import ceil, factorial def __snake_case ( _lowercase ): """simple docstring""" if not isinstance(_lowercase ,_lowercase ): raise TypeError('''Undefined for non-integers''' ) elif precision < 1: raise ValueError('''Undefined for non-natural numbers''' ) UpperCamelCase = precision UpperCamelCase = ceil(precision / 14 ) UpperCamelCase = 42_6880 * Decimal(1_0005 ).sqrt() UpperCamelCase = 1 UpperCamelCase = 1359_1409 UpperCamelCase = Decimal(_lowercase ) for k in range(1 ,_lowercase ): UpperCamelCase = factorial(6 * k ) // (factorial(3 * k ) * factorial(_lowercase ) ** 3) linear_term += 5_4514_0134 exponential_term *= -26_2537_4126_4076_8000 partial_sum += Decimal(multinomial_term * linear_term ) / exponential_term return str(constant_term / partial_sum )[:-1] if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = 50 print(f'The first {n} digits of pi is: {pi(n)}')
34
"""simple docstring""" import argparse import requests import torch from PIL import Image from transformers import ViTMAEConfig, ViTMAEForPreTraining, ViTMAEImageProcessor def __snake_case ( _lowercase ): """simple docstring""" if "cls_token" in name: UpperCamelCase = name.replace('''cls_token''' ,'''vit.embeddings.cls_token''' ) if "mask_token" in name: UpperCamelCase = name.replace('''mask_token''' ,'''decoder.mask_token''' ) if "decoder_pos_embed" in name: UpperCamelCase = name.replace('''decoder_pos_embed''' ,'''decoder.decoder_pos_embed''' ) if "pos_embed" in name and "decoder" not in name: UpperCamelCase = name.replace('''pos_embed''' ,'''vit.embeddings.position_embeddings''' ) if "patch_embed.proj" in name: UpperCamelCase = name.replace('''patch_embed.proj''' ,'''vit.embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: UpperCamelCase = name.replace('''patch_embed.norm''' ,'''vit.embeddings.norm''' ) if "decoder_blocks" in name: UpperCamelCase = name.replace('''decoder_blocks''' ,'''decoder.decoder_layers''' ) if "blocks" in name: UpperCamelCase = name.replace('''blocks''' ,'''vit.encoder.layer''' ) if "attn.proj" in name: UpperCamelCase = name.replace('''attn.proj''' ,'''attention.output.dense''' ) if "attn" in name: UpperCamelCase = name.replace('''attn''' ,'''attention.self''' ) if "norm1" in name: UpperCamelCase = name.replace('''norm1''' ,'''layernorm_before''' ) if "norm2" in name: UpperCamelCase = name.replace('''norm2''' ,'''layernorm_after''' ) if "mlp.fc1" in name: UpperCamelCase = name.replace('''mlp.fc1''' ,'''intermediate.dense''' ) if "mlp.fc2" in name: UpperCamelCase = name.replace('''mlp.fc2''' ,'''output.dense''' ) if "decoder_embed" in name: UpperCamelCase = name.replace('''decoder_embed''' ,'''decoder.decoder_embed''' ) if "decoder_norm" in name: UpperCamelCase = name.replace('''decoder_norm''' ,'''decoder.decoder_norm''' ) if "decoder_pred" in name: UpperCamelCase = name.replace('''decoder_pred''' ,'''decoder.decoder_pred''' ) if "norm.weight" in name and "decoder" not in name: UpperCamelCase = name.replace('''norm.weight''' ,'''vit.layernorm.weight''' ) if "norm.bias" in name and "decoder" not in name: UpperCamelCase = name.replace('''norm.bias''' ,'''vit.layernorm.bias''' ) return name def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" for key in orig_state_dict.copy().keys(): UpperCamelCase = orig_state_dict.pop(_lowercase ) if "qkv" in key: UpperCamelCase = key.split('''.''' ) UpperCamelCase = int(key_split[1] ) if "decoder_blocks" in key: UpperCamelCase = config.decoder_hidden_size UpperCamelCase = '''decoder.decoder_layers.''' if "weight" in key: UpperCamelCase = val[:dim, :] UpperCamelCase = val[dim : dim * 2, :] UpperCamelCase = val[-dim:, :] elif "bias" in key: UpperCamelCase = val[:dim] UpperCamelCase = val[dim : dim * 2] UpperCamelCase = val[-dim:] else: UpperCamelCase = config.hidden_size UpperCamelCase = '''vit.encoder.layer.''' if "weight" in key: UpperCamelCase = val[:dim, :] UpperCamelCase = val[dim : dim * 2, :] UpperCamelCase = val[-dim:, :] elif "bias" in key: UpperCamelCase = val[:dim] UpperCamelCase = val[dim : dim * 2] UpperCamelCase = val[-dim:] else: UpperCamelCase = val return orig_state_dict def __snake_case ( _lowercase ,_lowercase ): """simple docstring""" UpperCamelCase = ViTMAEConfig() if "large" in checkpoint_url: UpperCamelCase = 1024 UpperCamelCase = 4096 UpperCamelCase = 24 UpperCamelCase = 16 elif "huge" in checkpoint_url: UpperCamelCase = 14 UpperCamelCase = 1280 UpperCamelCase = 5120 UpperCamelCase = 32 UpperCamelCase = 16 UpperCamelCase = ViTMAEForPreTraining(_lowercase ) UpperCamelCase = torch.hub.load_state_dict_from_url(_lowercase ,map_location='''cpu''' )['''model'''] UpperCamelCase = ViTMAEImageProcessor(size=config.image_size ) UpperCamelCase = convert_state_dict(_lowercase ,_lowercase ) model.load_state_dict(_lowercase ) model.eval() UpperCamelCase = '''https://user-images.githubusercontent.com/11435359/147738734-196fd92f-9260-48d5-ba7e-bf103d29364d.jpg''' UpperCamelCase = Image.open(requests.get(_lowercase ,stream=_lowercase ).raw ) UpperCamelCase = ViTMAEImageProcessor(size=config.image_size ) UpperCamelCase = image_processor(images=_lowercase ,return_tensors='''pt''' ) # forward pass torch.manual_seed(2 ) UpperCamelCase = model(**_lowercase ) UpperCamelCase = outputs.logits if "large" in checkpoint_url: UpperCamelCase = torch.tensor( [[-0.7309, -0.7128, -1.0169], [-1.0161, -0.9058, -1.1878], [-1.0478, -0.9411, -1.1911]] ) elif "huge" in checkpoint_url: UpperCamelCase = torch.tensor( [[-1.1599, -0.9199, -1.2221], [-1.1952, -0.9269, -1.2307], [-1.2143, -0.9337, -1.2262]] ) else: UpperCamelCase = torch.tensor( [[-0.9192, -0.8481, -1.1259], [-1.1349, -1.0034, -1.2599], [-1.1757, -1.0429, -1.2726]] ) # verify logits assert torch.allclose(logits[0, :3, :3] ,_lowercase ,atol=1e-4 ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(_lowercase ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(_lowercase ) if __name__ == "__main__": SCREAMING_SNAKE_CASE_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--checkpoint_url', default='https://dl.fbaipublicfiles.com/mae/visualize/mae_visualize_vit_base.pth', type=str, help='URL of the checkpoint you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) SCREAMING_SNAKE_CASE_ = parser.parse_args() convert_vit_mae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
34
1
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) SCREAMING_SNAKE_CASE_ = '▁' SCREAMING_SNAKE_CASE_ = {'vocab_file': 'sentencepiece.bpe.model', 'monolingual_vocab_file': 'dict.txt'} SCREAMING_SNAKE_CASE_ = { 'vocab_file': { 'vinai/bartpho-syllable': 'https://huggingface.co/vinai/bartpho-syllable/resolve/main/sentencepiece.bpe.model', }, 'monolingual_vocab_file': { 'vinai/bartpho-syllable': 'https://huggingface.co/vinai/bartpho-syllable/resolve/main/dict.txt', }, } SCREAMING_SNAKE_CASE_ = {'vinai/bartpho-syllable': 1024} class snake_case_ ( lowerCamelCase_ ): """simple docstring""" A_ = VOCAB_FILES_NAMES A_ = PRETRAINED_VOCAB_FILES_MAP A_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ = ['''input_ids''', '''attention_mask'''] def __init__( self , lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_="<s>" , lowerCamelCase_="</s>" , lowerCamelCase_="</s>" , lowerCamelCase_="<s>" , lowerCamelCase_="<unk>" , lowerCamelCase_="<pad>" , lowerCamelCase_="<mask>" , lowerCamelCase_ = None , **lowerCamelCase_ , ) -> None: # Mask token behave like a normal word, i.e. include the space before it UpperCamelCase = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_) if isinstance(lowerCamelCase_ , lowerCamelCase_) else mask_token UpperCamelCase = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=lowerCamelCase_ , eos_token=lowerCamelCase_ , unk_token=lowerCamelCase_ , sep_token=lowerCamelCase_ , cls_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , mask_token=lowerCamelCase_ , sp_model_kwargs=self.sp_model_kwargs , **lowerCamelCase_ , ) UpperCamelCase = vocab_file UpperCamelCase = monolingual_vocab_file UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.Load(str(lowerCamelCase_)) # Load the reduced vocab # Keep order of special tokens for backward compatibility UpperCamelCase = {} UpperCamelCase = 0 for token in [bos_token, pad_token, eos_token, unk_token, sep_token, cls_token]: if str(lowerCamelCase_) not in self.fairseq_tokens_to_ids: UpperCamelCase = cnt cnt += 1 with open(lowerCamelCase_ , '''r''' , encoding='''utf-8''') as f: for line in f.readlines(): UpperCamelCase = line.strip().split()[0] UpperCamelCase = len(self.fairseq_tokens_to_ids) if str(lowerCamelCase_) not in self.fairseq_tokens_to_ids: UpperCamelCase = len(self.fairseq_tokens_to_ids) UpperCamelCase = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def __getstate__( self) -> str: UpperCamelCase = self.__dict__.copy() UpperCamelCase = None UpperCamelCase = self.sp_model.serialized_model_proto() return state def __setstate__( self , lowerCamelCase_) -> Optional[int]: UpperCamelCase = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs'''): UpperCamelCase = {} UpperCamelCase = spm.SentencePieceProcessor(**self.sp_model_kwargs) self.sp_model.LoadFromSerializedProto(self.sp_model_proto) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCamelCase = [self.cls_token_id] UpperCamelCase = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None , lowerCamelCase_ = False) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ , token_ids_a=lowerCamelCase_ , already_has_special_tokens=lowerCamelCase_) if token_ids_a is None: return [1] + ([0] * len(lowerCamelCase_)) + [1] return [1] + ([0] * len(lowerCamelCase_)) + [1, 1] + ([0] * len(lowerCamelCase_)) + [1] def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> List[int]: UpperCamelCase = [self.sep_token_id] UpperCamelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep) * [0] @property def UpperCAmelCase__ ( self) -> int: return len(self.fairseq_ids_to_tokens) def UpperCAmelCase__ ( self) -> List[Any]: UpperCamelCase = {self.convert_ids_to_tokens(lowerCamelCase_): i for i in range(self.vocab_size)} vocab.update(self.added_tokens_encoder) return vocab def UpperCAmelCase__ ( self , lowerCamelCase_) -> List[str]: return self.sp_model.encode(lowerCamelCase_ , out_type=lowerCamelCase_) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Union[str, Any]: if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] else: return self.unk_token_id def UpperCAmelCase__ ( self , lowerCamelCase_) -> List[str]: return self.fairseq_ids_to_tokens[index] def UpperCAmelCase__ ( self , lowerCamelCase_) -> Union[str, Any]: UpperCamelCase = ''''''.join(lowerCamelCase_).replace(lowerCamelCase_ , ''' ''').strip() return out_string def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_ = None) -> Tuple[str]: if not os.path.isdir(lowerCamelCase_): logger.error(F'Vocabulary path ({save_directory}) should be a directory') return UpperCamelCase = os.path.join( lowerCamelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file''']) UpperCamelCase = os.path.join( lowerCamelCase_ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''monolingual_vocab_file'''] , ) if os.path.abspath(self.vocab_file) != os.path.abspath(lowerCamelCase_) and os.path.isfile(self.vocab_file): copyfile(self.vocab_file , lowerCamelCase_) elif not os.path.isfile(self.vocab_file): with open(lowerCamelCase_ , '''wb''') as fi: UpperCamelCase = self.sp_model.serialized_model_proto() fi.write(lowerCamelCase_) if os.path.abspath(self.monolingual_vocab_file) != os.path.abspath( lowerCamelCase_) and os.path.isfile(self.monolingual_vocab_file): copyfile(self.monolingual_vocab_file , lowerCamelCase_) elif not os.path.isfile(self.monolingual_vocab_file): with open(lowerCamelCase_ , '''w''' , encoding='''utf-8''') as fp: for token in self.fairseq_tokens_to_ids: if token not in self.all_special_tokens: fp.write(F'{str(lowerCamelCase_)} \n') return out_vocab_file, out_monolingual_vocab_file
34
"""simple docstring""" import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def __snake_case ( ): """simple docstring""" raise RuntimeError('''CUDA out of memory.''' ) class snake_case_ ( nn.Module ): """simple docstring""" def __init__( self) -> Any: super().__init__() UpperCamelCase = nn.Linear(3 , 4) UpperCamelCase = nn.BatchNormad(4) UpperCamelCase = nn.Linear(4 , 5) def UpperCAmelCase__ ( self , lowerCamelCase_) -> Union[str, Any]: return self.lineara(self.batchnorm(self.lineara(lowerCamelCase_))) class snake_case_ ( unittest.TestCase ): """simple docstring""" def UpperCAmelCase__ ( self) -> List[Any]: UpperCamelCase = [] @find_executable_batch_size(starting_batch_size=1_2_8) def mock_training_loop_function(lowerCamelCase_): nonlocal batch_sizes batch_sizes.append(lowerCamelCase_) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(lowerCamelCase_ , [1_2_8, 6_4, 3_2, 1_6, 8]) def UpperCAmelCase__ ( self) -> Optional[Any]: UpperCamelCase = [] @find_executable_batch_size(starting_batch_size=1_2_8) def mock_training_loop_function(lowerCamelCase_ , lowerCamelCase_): nonlocal batch_sizes batch_sizes.append(lowerCamelCase_) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga UpperCamelCase , UpperCamelCase = mock_training_loop_function('''hello''') self.assertListEqual(lowerCamelCase_ , [1_2_8, 6_4, 3_2, 1_6, 8]) self.assertListEqual([bs, arga] , [8, '''hello''']) def UpperCAmelCase__ ( self) -> Tuple: @find_executable_batch_size(starting_batch_size=0) def mock_training_loop_function(lowerCamelCase_): pass with self.assertRaises(lowerCamelCase_) as cm: mock_training_loop_function() self.assertIn('''No executable batch size found, reached zero.''' , cm.exception.args[0]) def UpperCAmelCase__ ( self) -> List[Any]: @find_executable_batch_size(starting_batch_size=1_6) def mock_training_loop_function(lowerCamelCase_): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(lowerCamelCase_) as cm: mock_training_loop_function() self.assertIn('''No executable batch size found, reached zero.''' , cm.exception.args[0]) def UpperCAmelCase__ ( self) -> Union[str, Any]: @find_executable_batch_size(starting_batch_size=1_2_8) def mock_training_loop_function(lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(lowerCamelCase_) as cm: mock_training_loop_function(1_2_8 , '''hello''' , '''world''') self.assertIn('''Batch size was passed into `f`''' , cm.exception.args[0]) self.assertIn('''`f(arg1=\'hello\', arg2=\'world\')''' , cm.exception.args[0]) def UpperCAmelCase__ ( self) -> Dict: @find_executable_batch_size(starting_batch_size=1_6) def mock_training_loop_function(lowerCamelCase_): raise ValueError('''Oops, we had an error!''') with self.assertRaises(lowerCamelCase_) as cm: mock_training_loop_function() self.assertIn('''Oops, we had an error!''' , cm.exception.args[0]) @require_cuda def UpperCAmelCase__ ( self) -> Optional[int]: UpperCamelCase = torch.cuda.memory_allocated() UpperCamelCase = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated() , lowerCamelCase_) UpperCamelCase = release_memory(lowerCamelCase_) self.assertEqual(torch.cuda.memory_allocated() , lowerCamelCase_)
34
1
"""simple docstring""" import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () SCREAMING_SNAKE_CASE_ = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). SCREAMING_SNAKE_CASE_ = [0, 25, 50] SCREAMING_SNAKE_CASE_ = [25, 50, 75] SCREAMING_SNAKE_CASE_ = fuzz.membership.trimf(X, abca) SCREAMING_SNAKE_CASE_ = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. SCREAMING_SNAKE_CASE_ = np.ones(75) SCREAMING_SNAKE_CASE_ = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) SCREAMING_SNAKE_CASE_ = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) SCREAMING_SNAKE_CASE_ = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) SCREAMING_SNAKE_CASE_ = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) SCREAMING_SNAKE_CASE_ = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] SCREAMING_SNAKE_CASE_ = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) SCREAMING_SNAKE_CASE_ = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] SCREAMING_SNAKE_CASE_ = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] SCREAMING_SNAKE_CASE_ = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
34
"""simple docstring""" from typing import Dict from transformers import EvalPrediction, HfArgumentParser, TrainingArguments, is_torch_available from transformers.testing_utils import ( TestCasePlus, execute_subprocess_async, get_torch_dist_unique_port, require_torch_multi_gpu, require_torch_neuroncore, ) from transformers.training_args import ParallelMode from transformers.utils import logging SCREAMING_SNAKE_CASE_ = logging.get_logger(__name__) if is_torch_available(): import torch from torch import nn from torch.utils.data import Dataset from transformers import Trainer class snake_case_ ( lowerCamelCase_ ): """simple docstring""" def __init__( self , lowerCamelCase_ = 1_0_1) -> Tuple: UpperCamelCase = length def __len__( self) -> List[str]: return self.length def __getitem__( self , lowerCamelCase_) -> int: return i class snake_case_ : """simple docstring""" def __call__( self , lowerCamelCase_) -> str: return {"input_ids": torch.tensor(lowerCamelCase_), "labels": torch.tensor(lowerCamelCase_)} class snake_case_ ( nn.Module ): """simple docstring""" def __init__( self) -> List[Any]: super().__init__() # Add some (unused) params otherwise DDP will complain. UpperCamelCase = nn.Linear(1_2_0 , 8_0) def UpperCAmelCase__ ( self , lowerCamelCase_ , lowerCamelCase_=None) -> Any: if labels is not None: return torch.tensor(0.0 , device=input_ids.device), input_ids else: return input_ids class snake_case_ ( lowerCamelCase_ ): """simple docstring""" @require_torch_neuroncore def UpperCAmelCase__ ( self) -> Tuple: UpperCamelCase = F'--nproc_per_node=2\n --master_port={get_torch_dist_unique_port()}\n {self.test_file_dir}/test_trainer_distributed.py\n '.split() UpperCamelCase = self.get_auto_remove_tmp_dir() UpperCamelCase = F'--output_dir {output_dir}'.split() UpperCamelCase = ['''torchrun'''] + distributed_args + args execute_subprocess_async(lowerCamelCase_ , env=self.get_env()) # successful return here == success - any errors would have caused an error in the sub-call class snake_case_ ( lowerCamelCase_ ): """simple docstring""" @require_torch_multi_gpu def UpperCAmelCase__ ( self) -> Union[str, Any]: UpperCamelCase = F'--nproc_per_node={torch.cuda.device_count()}\n --master_port={get_torch_dist_unique_port()}\n {self.test_file_dir}/test_trainer_distributed.py\n '.split() UpperCamelCase = self.get_auto_remove_tmp_dir() UpperCamelCase = F'--output_dir {output_dir}'.split() UpperCamelCase = ['''torchrun'''] + distributed_args + args execute_subprocess_async(lowerCamelCase_ , env=self.get_env()) # successful return here == success - any errors would have caused an error in the sub-call if __name__ == "__main__": # The script below is meant to be run under torch.distributed, on a machine with multiple GPUs: # # PYTHONPATH="src" python -m torch.distributed.run --nproc_per_node 2 --output_dir output_dir ./tests/test_trainer_distributed.py SCREAMING_SNAKE_CASE_ = HfArgumentParser((TrainingArguments,)) SCREAMING_SNAKE_CASE_ = parser.parse_args_into_dataclasses()[0] logger.warning( f'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}, ' f'distributed training: {training_args.parallel_mode != ParallelMode.NOT_DISTRIBUTED}' ) # Essentially, what we want to verify in the distributed case is that we get all samples back, # in the right order. (this is crucial for prediction for instance) for dataset_length in [101, 40, 7]: SCREAMING_SNAKE_CASE_ = DummyDataset(dataset_length) def __snake_case ( _lowercase ): """simple docstring""" UpperCamelCase = list(range(len(_lowercase ) ) ) UpperCamelCase = p.predictions.tolist() == sequential and p.label_ids.tolist() == sequential if not success and training_args.local_rank == 0: logger.warning( '''Predictions and/or labels do not match expected results:\n - predictions: ''' f'{p.predictions.tolist()}\n - labels: {p.label_ids.tolist()}\n - expected: {sequential}' ) return {"success": success} SCREAMING_SNAKE_CASE_ = Trainer( model=DummyModel(), args=training_args, data_collator=DummyDataCollator(), eval_dataset=dataset, compute_metrics=compute_metrics, ) SCREAMING_SNAKE_CASE_ = trainer.evaluate() logger.info(metrics) if metrics["eval_success"] is not True: logger.error(metrics) exit(1) SCREAMING_SNAKE_CASE_ = trainer.predict(dataset) logger.info(p.metrics) if p.metrics["test_success"] is not True: logger.error(p.metrics) exit(1) SCREAMING_SNAKE_CASE_ = 2 SCREAMING_SNAKE_CASE_ = trainer.evaluate() logger.info(metrics) if metrics["eval_success"] is not True: logger.error(metrics) exit(1) SCREAMING_SNAKE_CASE_ = trainer.predict(dataset) logger.info(p.metrics) if p.metrics["test_success"] is not True: logger.error(p.metrics) exit(1) SCREAMING_SNAKE_CASE_ = None
34
1