code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' import json import os import shutil import warnings from argparse import ArgumentParser, Namespace from pathlib import Path from typing import List from ..utils import logging from . import BaseTransformersCLICommand try: from cookiecutter.main import cookiecutter ...
350
'''simple docstring''' import argparse import json import os import re import torch from transformers import BloomConfig, BloomModel from transformers.file_utils import CONFIG_NAME, WEIGHTS_NAME from transformers.utils import logging logging.set_verbosity_info() a = [ 'wor...
350
1
'''simple docstring''' import unittest from transformers import XLMConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_co...
720
'''simple docstring''' def a__ ( lowerCAmelCase__ ) -> float: if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) UpperCAmelCase__ : Tuple = sum(lowerCAmelCase__ ) / len(lowerCAmelCase__ ) # Calculate the average...
312
0
"""simple docstring""" def a ( __UpperCAmelCase : str ) -> str: return " ".join(input_str.split()[::-1] ) if __name__ == "__main__": import doctest doctest.testmod()
96
"""simple docstring""" import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def a ( __UpperCAmelCase : bytes , __UpperCAmelCase : int ) -> np.array: __magic_name__: Optional[i...
96
1
import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint _...
84
import string import numpy def a_ ( __magic_name__ , __magic_name__ ) -> int: """simple docstring""" return b if a == 0 else greatest_common_divisor(b % a , __magic_name__ ) class a_ : A__ : List[Any] = string.asc...
84
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available UpperCAmelCase__ : List[Any] = { 'configuration_data2vec_audio': ['DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Dat...
223
'''simple docstring''' import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging impo...
541
0
'''simple docstring''' import argparse import copy def lowercase_ ( lowercase__ ) ->str: _snake_case: Union[str, Any] = {} with open(lowercase__ ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: ...
273
'''simple docstring''' import inspect import unittest from transformers import ViTConfig from transformers.testing_utils import ( require_accelerate, require_torch, require_torch_gpu, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_avai...
273
1
"""simple docstring""" from typing import Optional from torch import nn from .transformer_ad import TransformeraDModel, TransformeraDModelOutput class UpperCamelCase__ ( nn.Module): """simple docstring""" def __init__( self : Tuple , Upp...
545
"""simple docstring""" import json import logging import math import os import sys from dataclasses import dataclass, field from typing import Optional from datasets import Dataset, load_dataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_FOR_MASKED_LM_MAPP...
545
1
import argparse import os import torch from transformers.utils import WEIGHTS_NAME __snake_case : int =['small', 'medium', 'large'] __snake_case : List[Any] ='lm_head.decoder.weight' __snake_case : str ='lm_head.weight' def lowerCAmelCase__ ( lowe...
90
import gc import unittest import torch from parameterized import parameterized from diffusers import AutoencoderKL from diffusers.utils import floats_tensor, load_hf_numpy, require_torch_gpu, slow, torch_all_close, torch_device from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.te...
90
1
'''simple docstring''' from math import asin, atan, cos, radians, sin, sqrt, tan SCREAMING_SNAKE_CASE_ = 6_37_81_37.0 SCREAMING_SNAKE_CASE_ = 6_35_67_52.31_42_45 SCREAMING_SNAKE_CASE_ = 6_37_81_37 def UpperCamelCase__ ( _lowercase : str , _lowerca...
523
import warnings from ...utils import logging from .image_processing_yolos import YolosImageProcessor _lowercase : int =logging.get_logger(__name__) class UpperCamelCase_ ( snake_case__ ): def __init__( self : Tuple , *lowerCamelC...
364
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligne...
716
'''simple docstring''' def a_ ( lowerCamelCase : Tuple , lowerCamelCase : Tuple ): # "extended trapezoidal rule" # int(f) = dx/2 * (f1 + 2f2 + ... + fn) lowerCAmelCase = (boundary[1] - boundary[0]) / steps lowerCAmelCase = boundary[0] lowerCAmelCase...
513
0
'''simple docstring''' import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumen...
126
'''simple docstring''' from typing import Any import numpy as np def _lowerCAmelCase ( _lowerCAmelCase )-> bool: return np.array_equal(_lowerCAmelCase , matrix.conjugate().T ) def _lowerCAmelCase ( _lowerCAmelCase , _lowerCAmelCase )-> Any: __UpperCAmel...
126
1
"""simple docstring""" from __future__ import annotations from typing import TypedDict class UpperCamelCase ( _UpperCamelCase ): UpperCAmelCase : str UpperCAmelCase : int def __UpperCAmelCase ( UpperCAmelCase_ : Any ) ...
707
"""simple docstring""" from typing import Any class UpperCamelCase : def __init__(self : List[str] , _A : Any) -> int: __snake_case : Any = data __snake_case : Dict = None def __repr__(self : ...
192
0
def __a ( __UpperCAmelCase ): a__ = len(__UpperCAmelCase ) a__ = len(matrix[0] ) a__ = min(__UpperCAmelCase , __UpperCAmelCase ) for row in range(__UpperCAmelCase ): # Check if diagonal element is not zero if matrix[row][row] !=...
194
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _int...
194
1
import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ....
199
from typing import TYPE_CHECKING from ...utils import _LazyModule __a : Optional[int] = {"tokenization_bertweet": ["BertweetTokenizer"]} if TYPE_CHECKING: from .tokenization_bertweet import BertweetTokenizer else: import sys __a : int = _LazyModule(__name__, globa...
199
1
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTConfig, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTImageProcessor, ) from transformer...
192
from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from...
192
1
'''simple docstring''' import numpy as np from transformers import BatchFeature from transformers.testing_utils import require_tf, require_torch from .test_feature_extraction_common import FeatureExtractionSavingTestMixin class lowerCamelCase ( __UpperCAmelCase ): # to overwrite a...
273
'''simple docstring''' A : List[str] = '\n# Transformers 설치 방법\n! pip install transformers datasets\n# 마지막 릴리스 대신 소스에서 설치하려면, 위 명령을 주석으로 바꾸고 아래 명령을 해제하세요.\n# ! pip install git+https://github.com/huggingface/transformers.git\n' A : List[str] = [{'type': 'code', 'content': INS...
273
1
import math import flax.linen as nn import jax.numpy as jnp def __lowerCAmelCase ( A_ : jnp.ndarray , A_ : int , A_ : float = 1 , A_ : float = 1 , A_ : float = 1.0e4 , A_ : bool = False , A_ : float = 1.0 , ) -> jnp.ndarray: assert...
221
import os import tempfile import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from torch import nn from transformers import ( Adafactor, AdamW, get_constant_schedu...
221
1
from __future__ import annotations from random import choice def __a ( __lowerCAmelCase ) -> int: return choice(__lowerCAmelCase ) def __a ( __lowerCAmelCase , __lowerCAmelCase ) -> int: SCREAMING_SNAKE_CASE : str = rando...
308
import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class lowercase : '''simple docstring''' def __init__( self : Tuple , snake_case : ...
308
1
"""simple docstring""" import json import os from typing import Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase : int = logging.get_logger(__name__) UpperCAmelCase : Union[str, Any] = { "voc...
567
"""simple docstring""" from __future__ import annotations from collections import namedtuple from dataclasses import dataclass @dataclass class SCREAMING_SNAKE_CASE__ : lowercase__ = 42 lowercase__ = None lowercase__ = None UpperCAmelCase : Dict =...
567
1
'''simple docstring''' def _a ( _lowercase : int , _lowercase : int ): '''simple docstring''' return number | (1 << position) def _a ( _lowercase : int , _lowercase : int ): '''simpl...
710
'''simple docstring''' from typing import Optional, Tuple, Union import tensorflow as tf from ...activations_tf import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_tf_outputs import ( TFBaseModelOutput...
266
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices A = logging.get_logger(__name__) A = { '''facebook/convnextv...
52
"""simple docstring""" from __future__ import annotations class __lowercase : '''simple docstring''' def __init__( self , _UpperCAmelCase , _UpperCAmelCase ): __a , __a : List[Any] =...
52
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class lowerCAmelCase__ ( metaclass=UpperCAmelCase_ ): lowercase__ : Union[str, Any] = ["""torch""", """transformers""", """onnx"""] def __init__( self , *UpperCamelCase__ , **UpperCamelC...
261
"""simple docstring""" import logging import os import sys from dataclasses import dataclass, field from importlib import import_module from typing import Dict, List, Optional, Tuple import numpy as np from seqeval.metrics import accuracy_score, fa_score, precision_score, recall_score from torch import nn from ...
261
1
import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, ca...
70
from __future__ import annotations def _snake_case( SCREAMING_SNAKE_CASE__ ) -> list[int]: lowercase : int = [True] * limit lowercase : Tuple = False lowercase : List[Any] = False lowercase : Union[str, Any] = True ...
336
0
'''simple docstring''' from __future__ import annotations from cmath import sqrt def __UpperCamelCase ( _lowercase, _lowercase, _lowercase ) -> tuple[complex, complex]: if a == 0: raise ValueError('Coefficient \'a\' must not be zero.' ) _lowercase : Tuple ...
4
'''simple docstring''' import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFe...
4
1
'''simple docstring''' import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def ...
109
'''simple docstring''' import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict impo...
374
0
"""simple docstring""" def __UpperCamelCase ( SCREAMING_SNAKE_CASE = 10_00 ) -> int: """simple docstring""" __snake_case , __snake_case = 1, 1 __snake_case = [] for i in range(1 , n + 1 ): __snake_case ...
614
"""simple docstring""" import argparse from transformers import ( TapasConfig, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasTokenizer, load_tf_weights_in_tapas, ) from transformers.utils import logging logging.set_verbosity_info() d...
614
1
from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_...
64
import os import time import warnings from dataclasses import dataclass, field from enum import Enum from typing import List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import logging f...
278
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ......
20
"""simple docstring""" # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org...
20
1
def lowerCAmelCase_ ( __lowerCamelCase = 1_0_0_0_0_0_0 ): __snake_case : Any = 1 __snake_case : Any = 1 __snake_case : Union[str, Any] = {1: 1} for inputa in range(2 , __lowerCamelCase ): __sna...
81
from __future__ import annotations import sys from collections import deque from typing import Generic, TypeVar snake_case = TypeVar("T") class __A ( Generic[T] ): '''simple docstring''' a_ = 42 # Cache store of keys a_ = 42 # References of the keys in...
424
0
from __future__ import annotations import copy import inspect import unittest import numpy as np from transformers import is_tf_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from transformers.utils import cached_property from ...
376
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = { "nvidia/segforme...
376
1
from math import pi def lowerCAmelCase__ ( UpperCamelCase_ : int , UpperCamelCase_ : int )-> float: return 2 * pi * radius * (angle / 3_6_0) if __name__ == "__main__": print(arc_length(90, 10))
632
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _lowercase = { "configuration_upernet": ["UperNetConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: ...
632
1
import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers clas...
297
import contextlib import copy import random from typing import Any, Dict, Iterable, Optional, Union import numpy as np import torch from .utils import deprecate, is_transformers_available if is_transformers_available(): import transformers def _lowerCamelCase ( _a ): """simple docstring""" ...
297
1
"""simple docstring""" from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torc...
571
"""simple docstring""" def a_ ( __a ): assert ( isinstance(__a , __a ) and number_of_steps > 0 ), f'''number_of_steps needs to be positive integer, your input {number_of_steps}''' if number_of_steps == 1: return 1 A__ , A__ ...
571
1
"""simple docstring""" from math import ceil def _UpperCamelCase ( UpperCamelCase = 1001 ) -> str: """simple docstring""" __UpperCAmelCase : Any = 1 for i in range(1 , int(ceil(n / 2.0 ) ) ): __UpperCAmelCase : int = ...
720
"""simple docstring""" import inspect import unittest from transformers import SegformerConfig, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTest...
487
0
import math def _lowerCAmelCase ( A__ ): if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All ...
622
"""simple docstring""" import numpy as np from numpy import ndarray from scipy.optimize import Bounds, LinearConstraint, minimize def _lowerCAmelCase ( UpperCamelCase_ ): return np.dot(UpperCamelCase_ , UpperCamelCase_ ) class SCREAMING_SNAKE_CASE_ : """simple d...
155
0
import torch from transformers import AutoModel class lowerCAmelCase__ ( torch.nn.Module ): def __init__( self : Union[str, Any] , __UpperCamelCase : List[Any]="sayef/fsner-bert-base-uncased" ) -> Dict: super(__UpperCamelCase , self ).__init__()...
224
import argparse import collections import os import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_table.py __snake_case :Tuple ='src/transformers' __snake_case ...
224
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { '''facebook/data2vec-...
40
'''simple docstring''' from __future__ import annotations # This is the precision for this function which can be altered. # It is recommended for users to keep this number greater than or equal to 10. __A : Tuple = 10 def UpperCamelCase_ ( A__ : ...
275
0
import os import unittest from transformers.models.bartpho.tokenization_bartpho import VOCAB_FILES_NAMES, BartphoTokenizer from transformers.testing_utils import get_tests_dir from ...test_tokenization_common import TokenizerTesterMixin _snake_case = get_tests_dir("fixtures/test_sentencepiece_...
720
import datasets from .nmt_bleu import compute_bleu # From: https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py _snake_case = "\\n@INPROCEEDINGS{Papineni02bleu:a,\n author = {Kishore Papineni and Salim Roukos and Todd Ward and Wei-jing Zhu},\n title = {BLEU: a Method for Automatic Evalua...
658
0
'''simple docstring''' # NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from ...utils import deprecate from ..controlnet.pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline #...
42
'''simple docstring''' from __future__ import annotations def _UpperCamelCase ( __UpperCamelCase ) -> bool: lowerCamelCase_ = str(__UpperCamelCase ) return len(__UpperCamelCase ) == 9 and set(__UpperCamelCase ) == set('123456789' ) def _UpperCamelCase ( ) -> ...
42
1
"""simple docstring""" def A ( snake_case__ , snake_case__ ): '''simple docstring''' SCREAMING_SNAKE_CASE__ = len(snake_case__ ) SCREAMING_SNAKE_CASE__ = [[False] * (required_sum + 1) for _ in range(arr_len + 1 )] # for each arr ...
616
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging A_ : List[Any] = logging.get_logger(__name__) A_ : Dict = { "microsoft/unispeech-large-1500h-cv": ( "https:...
616
1
from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.m...
412
import numpy as np from numpy import ndarray from scipy.optimize import Bounds, LinearConstraint, minimize def UpperCAmelCase_ ( UpperCAmelCase__ ): return np.dot(UpperCAmelCase__ , UpperCAmelCase__ ) class UpperCamelCase__ : def __init__( self : Any , *, ...
412
1
import collections import os import re from pathlib import Path __magic_name__ = '''src/transformers''' # Matches is_xxx_available() __magic_name__ = re.compile(r'''is\_([a-z_]*)_available()''') # Catches a one-line _import_struct = {xxx} __magic_name__ = re.compile(r'''^_import...
700
from __future__ import annotations def SCREAMING_SNAKE_CASE__ ( __lowerCAmelCase , __lowerCAmelCase ): if len(__lowerCAmelCase ) == 0: return False snake_case__ = len(__lowerCAmelCase ) // 2 if a_list[midpoint] == item: return True if item < a...
530
0
from ....configuration_utils import PretrainedConfig from ....utils import logging _a: int = logging.get_logger(__name__) _a: int = { """CarlCochet/trajectory-transformer-halfcheetah-medium-v2""": ( """https://huggingface.co/CarlCochet/trajectory-transformer-halfcheetah-me...
162
'''simple docstring''' from abc import ABC, abstractmethod from argparse import ArgumentParser class A ( SCREAMING_SNAKE_CASE__ ): """simple docstring""" @staticmethod @abstractmethod def _UpperCAmelCase ( __lowerCAmelCase ): raise NotImplementedError() ...
208
0
"""simple docstring""" import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester fro...
706
"""simple docstring""" import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common imp...
663
0
'''simple docstring''' import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime...
119
'''simple docstring''' def __UpperCamelCase ( lowercase__ : List[str], lowercase__ : Tuple ): '''simple docstring''' __lowercase =[0 for i in range(r + 1 )] # nc0 = 1 __lowercase =1 for i in range(1, n + 1 ): # to comput...
119
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available snake_case = { "configuration_nezha": ["NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP", "NezhaConfig"], } try: if not is_torch_available(): raise Optional...
587
import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_u...
587
1
'''simple docstring''' import argparse import os import numpy as np import tensorflow as tf import torch from transformers import BertModel def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : Any = ('dense.weight', 'attention.self....
195
'''simple docstring''' from pathlib import Path import fire def A__ ( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ): _UpperCamelCase : int = Path(UpperCAmelCase_ ) _UpperCamelCase : str = Path(UpperCAmelCase_ ) dest_di...
195
1
import json import os from collections import Counter import torch import torchvision import torchvision.transforms as transforms from PIL import Image from torch import nn from torch.utils.data import Dataset __lowerCamelCase : List[str] = {1: (1, 1), 2: (2, 1), 3: (3, 1), 4: (2, 2), 5: (5, 1),...
714
import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, PerceiverTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common import TokenizerTesterMixin...
316
0
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from ...utils import logging, randn_tensor from ..pipeline_utils import AudioPipelineOutput, DiffusionPipeline _a = logging.get_logger(__name__) # pylint: disable=invalid-name class ...
19
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __A : Optional[int] = logging.get_logger(__name__) __A : O...
231
0
"""simple docstring""" import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py _lowercase ...
705
"""simple docstring""" import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_tran...
22
0
import inspect import os import torch from transformers import AutoModel from transformers.testing_utils import mockenv_context from transformers.trainer_utils import set_seed import accelerate from accelerate.accelerator import Accelerator from accelerate.state import AcceleratorState from accelerate....
14
import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers cla...
14
1
import os import sys import transformers __lowercase : str = """3""" print("""Python version:""", sys.version) print("""transformers version:""", transformers.__version__) try: import torch print("""Torch version:""", torch.__version__) print("""Cuda available:""", torch.cuda.is_available()) pri...
706
"""simple docstring""" import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require...
66
0
'''simple docstring''' import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_att...
42
'''simple docstring''' import os import unittest from transformers import LxmertTokenizer, LxmertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tok...
189
0
"""simple docstring""" import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels lowerCAmelCase__ = object() # For specifying empty leaf dict `{}` lowerCAmelCase__ = o...
681
"""simple docstring""" def a__ ( SCREAMING_SNAKE_CASE : int = 1_0 , SCREAMING_SNAKE_CASE : int = 2_2 ): '''simple docstring''' lowerCAmelCase : Dict = range(1 , SCREAMING_SNAKE_CASE ) lowerCAmelCase : List[str] = ran...
681
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase = logging.get_logger(__name__) __lowerCAmelCase = { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t...
466
'''simple docstring''' # Note: if you intend to run this script make sure you look under scripts/fsmt/ # to locate the appropriate script to do the work correctly. There is a set of scripts to: # - download and prepare data and run the conversion script # - perform eval to get the best h...
466
1
from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import TensorType, logging if TYPE_CHECKING: from ...onnx.config import PatchingSpec from ...tokeniz...
26
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple reposit...
26
1
'''simple docstring''' import inspect import re from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCAmelCase_ = 'src/transformers' # This is to...
173
'''simple docstring''' # DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from dataclasses import dataclass from typing import Optional, Tuple, Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import Base...
173
1
"""simple docstring""" from unittest.mock import Mock, patch from file_transfer.send_file import send_file @patch('socket.socket' ) @patch('builtins.open' ) def SCREAMING_SNAKE_CASE_ ( snake_case : Optional[Any] , snake_case : Tuple )-> List[Any]: # ===== initialization ...
701
"""simple docstring""" import random import unittest import torch from diffusers import IFInpaintingPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ...
222
0
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] ) -> List[str]: if not head: return True # split the list to two parts __snake_case , __snake_case = head.next, head while fast and fast.next: __snake_case = fast.nex...
69
'''simple docstring''' def lowerCamelCase__ ( a ): if number < 0: raise ValueError('number must not be negative' ) return number & (number - 1) == 0 if __name__ == "__main__": import doctest doctest.testmod()
356
0
'''simple docstring''' import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class __A (__magic_name__ ): snake_case :Dict = (PNDMScheduler,) snake_case :List[Any] = (("...
10
'''simple docstring''' def _lowercase ( lowerCamelCase__ = 100 ) -> int: """simple docstring""" __UpperCAmelCase : Optional[Any] = (n * (n + 1) // 2) ** 2 __UpperCAmelCase : Any = n * (n + 1) * (2 * n + 1) // 6 retu...
10
1
import tensorflow as tf from ...tf_utils import shape_list class A__ ( tf.keras.layers.Layer ): def __init__( self , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase , lowerCamelCase=1 , lowerCamelCase=False...
154
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unles...
154
1
import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _lowercase = get_tests_dir('''fixtures/spiece.model''')...
96
from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def _A (UpperCamelCase : str ) ->None: '''simple docstring''' lowerCamelCase__ ,lowerCamelCase__ : List[str] = analyze_text(UpperCamelCase ) l...
96
1
from __future__ import annotations class SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self : Any , UpperCAmelCase_ : int ): SCREAMING_SNAKE_CASE : int = data SCREAMING_SNAKE_CASE : Node | None ...
62
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available a_ = { 'configuration_bridgetower': [ 'BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BridgeTowerConfi...
296
0
from __future__ import annotations from collections import deque class UpperCAmelCase_ : '''simple docstring''' def __init__( self : str , a : Tuple ) -> List[Any]: SCREAMING_SNAKE_CASE = [] self.adlist.append( ...
703
from __future__ import annotations import typing from collections.abc import Iterable import numpy as np __A : Dict = typing.Union[Iterable[float], Iterable[int], np.ndarray] # noqa: UP007 __A : Tuple = typing.Union[np.floataa, int, float] # noqa: UP007 def lowerCamelC...
450
0
"""simple docstring""" import itertools import math def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all ev...
389
"""simple docstring""" import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger _snake_case = get_logger(__name__) class UpperCamelCase ( enum.Enum ): UpperCamelCase : str ...
389
1
import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Acc...
353
import tempfile import torch from diffusers import PNDMScheduler from .test_schedulers import SchedulerCommonTest class _A ( __magic_name__): SCREAMING_SNAKE_CASE : List[str] = (PNDMScheduler,) SCREAMING_SNAKE_CASE : Dict = (('''num_inference_steps''', 50),) def UpperCAmelC...
353
1
'''simple docstring''' from timeit import timeit _lowerCAmelCase = { "MALAYALAM": True, "String": False, "rotor": True, "level": True, "A": True, "BB": True, "ABC": False, "amanaplanacanalpanama": True, # "a man a plan a canal panama" } # Ensure our t...
161
'''simple docstring''' import logging import math from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Sequence, Tuple, Union import torch from .tensor_utils import tensor_tree_map, tree_map def _lowerCAmelCase ( lowercase :...
161
1
'''simple docstring''' from typing import Dict, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, flip_channel_order, get_resize_output_image_size, ...
704
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_ava...
68
0
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import numpy as np import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor import transformers from tr...
637
from ....configuration_utils import PretrainedConfig from ....utils import logging __a : List[Any] = logging.get_logger(__name__) # TODO: upload to AWS __a : Union[str, Any] = { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/r...
637
1
"""simple docstring""" from __future__ import annotations def snake_case (A_ :list[int] ): '''simple docstring''' if len(A_ ) == 0: return array a, a : Any = min(A_ ), max(A_ ) # Compute the variables a : int = _max - _min + 1 a, a ...
118
"""simple docstring""" from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from...
118
1
def lowercase__ ( _UpperCamelCase , _UpperCamelCase) -> Optional[Any]: """simple docstring""" UpperCamelCase = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def lowe...
280
from collections.abc import Callable from math import pi, sqrt from random import uniform from statistics import mean def _A ( SCREAMING_SNAKE_CASE ): # A local function to see if a dot lands in the circle. def is_in_circle(SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE ) -> bool: UpperCAmelCase...
113
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) snake_case = { """configuration_blip""": [ """BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", ...
535
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_availabl...
535
1
'''simple docstring''' import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor...
561
'''simple docstring''' from typing import List import datasets from datasets.tasks import AudioClassification from ..folder_based_builder import folder_based_builder lowerCAmelCase :Dict = datasets.utils.logging.get_logger(__name__) class _lowerCamelCase ( folder_based_builder.FolderBasedBuil...
561
1
'''simple docstring''' from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) _lowerCamelCase = { 'google/umt5-small': ...
709
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowerCamelCase = {'configuration_glpn': ['GLPN_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GLPNConfig']} try: if not is_vision_available():...
572
0
"""simple docstring""" import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from ...
572
"""simple docstring""" def _lowerCAmelCase ( lowerCamelCase__ : float ) -> float: if edge <= 0 or not isinstance(lowerCamelCase__, lowerCamelCase__ ): raise ValueError("Length must be a positive." ) return 3 * ((2_5 + 1_0 * (5 ** (1 / 2))) ** (1 / 2)) * (edge**2) ...
572
1
from maths.is_square_free import is_square_free from maths.prime_factors import prime_factors def _UpperCAmelCase ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' lowerCAmelCase : List[Any] = prime_factors(__A ) if is_square...
701
lowerCAmelCase : str ={ 'Pillow': 'Pillow<10.0.0', 'accelerate': 'accelerate>=0.20.3', 'av': 'av==9.2.0', 'beautifulsoup4': 'beautifulsoup4', 'black': 'black~=23.1', 'codecarbon': 'codecarbon==1.2.0', 'cookiecutter': 'cookiecutter==1.7.3', 'dataclasses': '...
693
0
import socket def lowerCamelCase( ): _SCREAMING_SNAKE_CASE =socket.socket(socket.AF_INET ,socket.SOCK_STREAM) _SCREAMING_SNAKE_CASE =socket.gethostname() _SCREAMING_SNAKE_CASE =1_2312 sock.connect((host, port)) sock.send(b'''Hello server!''') with open('''R...
691
from typing import TYPE_CHECKING from ....utils import _LazyModule snake_case_ : Dict = {'''tokenization_tapex''': ['''TapexTokenizer''']} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys snake_case_ : Union[str, Any] = ...
691
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCAmelCase : int = logging.get_logger(__name__) _lowerCAmelCase : Optional[Any] = { "microsoft/trocr-base-handwritten": ( "https://huggingface.co/...
719
'''simple docstring''' from math import isqrt def _A ( snake_case__ : int ): return all(number % divisor != 0 for divisor in range(2 , isqrt(snake_case__ ) + 1 ) ) def _A ( snake_case__ : int = 10**6 ): snake_case__ : str = 0 snake_case__ : List...
694
0
'''simple docstring''' from typing import Optional, Tuple, Union import torch from einops import rearrange, reduce from diffusers import DDIMScheduler, DDPMScheduler, DiffusionPipeline, ImagePipelineOutput, UNetaDConditionModel from diffusers.schedulers.scheduling_ddim import DDIMSche...
342
'''simple docstring''' from typing import List, Optional, Union import numpy as np import tensorflow as tf from .utils import logging _lowercase = logging.get_logger(__name__) def __UpperCamelCase ( a : Union[tf.Tensor, np.ndarray] ) ->List[in...
342
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _snake_case = logging.get_logger(__name__) _snake_case = { "facebook/dpr-ctx_encoder-single-nq-base": ( "https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolv...
718
def _a ( __lowercase , __lowercase = 0 ) -> list: """simple docstring""" __UpperCamelCase = length or len(__lowercase ) __UpperCamelCase = False for i in range(length - 1 ): if list_data[i] > list_data[i + 1]: ...
567
0
'''simple docstring''' from random import shuffle import tensorflow as tf from numpy import array def _A ( A__ , A__ ): """simple docstring""" __lowercase = int(A__ ) assert noofclusters < len(A__ ) # Find out the dimensionality __lowercase = len(vectors...
41
"""simple docstring""" import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () A_ : int =np.linspace(start=0, stop=7_5, num=7_5, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership f...
650
0
import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor lowercase_ = logging.get_logger(__name__) class __a ( SCREAMING_SNAKE_CASE ): def __init__( self : Optional[int] , *snake_case_ : str , **snake_case_ ...
712
import warnings from ...utils import logging from .image_processing_poolformer import PoolFormerImageProcessor lowercase_ = logging.get_logger(__name__) class __a ( SCREAMING_SNAKE_CASE ): def __init__( self : Optional[Any] , *snake_case_ : List[str] ,...
456
0
def lowerCAmelCase_ ( _snake_case : Dict ) -> Any: '''simple docstring''' return [ { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], }, ...
124
from __future__ import annotations from typing import Any class _snake_case ( snake_case ): pass class _snake_case : def __init__( self , _a ): __magic_name__ : Any = data __magic_name__ : Node | None = None def __it...
124
1
import os import random import sys from . import cryptomath_module as cryptomath from . import rabin_miller A__: Union[str, Any] = 3 def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : int ) -> int: print("""Generating primitive root of p""" ) ...
708
'''simple docstring''' def SCREAMING_SNAKE_CASE_ ( _UpperCAmelCase : str ,_UpperCAmelCase : str = " " ) -> list: _a : int =[] _a : Tuple =0 for index, char in enumerate(_UpperCAmelCase ): ...
506
0
SCREAMING_SNAKE_CASE : Optional[Any] = tuple[float, float, float] SCREAMING_SNAKE_CASE : int = tuple[float, float, float] def UpperCamelCase_( lowerCamelCase_ , lowerCamelCase_ ) -> Vectorad: _lowercase : int = end_pointa[0] - end_pointa[0] ...
89
class __UpperCAmelCase : """simple docstring""" def __init__( self , __A ): __a = set_counts __a = max(__A ) __a = len(__A ) __a = [1] * num_sets __a = list(range(__A ) ) def snake_...
99
0
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinvaConfig, SwinvaForImageClassification def UpperCamelCase ( _lowerCAmelCase : ...
718
"""simple docstring""" import importlib import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Union import torch from ..utils import BaseOutput __A = """scheduler_config.json""" class a ( A_ ): A_ : ...
173
0
'''simple docstring''' import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_ch...
536
"""simple docstring""" from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional from packaging import version if TYPE_CHECKING: from ... import PreTrainedTokenizer, TensorType from ...configuration_utils import PretrainedConfig from ...on...
449
0
from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def __lowerCamelCase ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ = 1 / sqrt(2 ) ): """simple docstring""" lowercase__ : str = tau * frequency / samplerate lowerca...
706
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appli...
81
0
'''simple docstring''' import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def A_ ( snake_case , snake_case , snake_case ): SCREAMING_SNA...
143
'''simple docstring''' import argparse import os import re import packaging.version A_ = "examples/" A_ = { "examples": (re.compile(R"^check_min_version\(\"[^\"]+\"\)\s*$", re.MULTILINE), "check_min_version(\"VERSION\")\n"), "init": (re.compile(R"^__version__\s+=\s+...
143
1
'''simple docstring''' import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, XLMRobertaTokenizer from diffusers import AltDiffusionPipeline, AutoencoderKL, DDIMScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.a...
718
from datetime import datetime import matplotlib.pyplot as plt import torch def __lowerCAmelCase ( _UpperCamelCase ) -> int: '''simple docstring''' for param in module.parameters(): lowerCamelCase__: Optional[int] = False def...
242
0
"""simple docstring""" import json import os import shutil import tempfile import unittest import numpy as np from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES, BertTokenizer from transformers.testing_utils import require_tokenizers, require_vision f...
182
"""simple docstring""" import cmath import math def a__ ( lowerCAmelCase , lowerCAmelCase , lowerCAmelCase , lowerCAmelCase ) -> complex: UpperCAmelCase__ : str = math.radians(lowerCAmelCase ) UpperCAmelCase__ : Optional[int] = math.radians(...
182
1
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( ...
719
def a__ (__lowercase :str , __lowercase :str ) -> bool: _A : Dict = len(__lowercase ) + 1 _A : Optional[int] = len(__lowercase ) + 1 # dp is a 2d matrix where dp[i][j] denotes whether prefix string of # length ...
332
0
import math from typing import Any, Callable, List, Optional, Tuple, Union import numpy as np import torch from ...models import TaFilmDecoder from ...schedulers import DDPMScheduler from ...utils import is_onnx_available, logging, randn_tensor if is_onnx_available(): from ..onnx_utils import OnnxRuntimeMode...
6
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase = logging.get_logger(__name__) class UpperCamelCase_ ( UpperCamelCase__ ): lowerCamelCase_ = "encoder-decoder" lowerCamelCase_ = ...
6
1
import os from shutil import copyfile from typing import List, Optional, Tuple from tokenizers import processors from ...tokenization_utils import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_...
335
import json from typing import TYPE_CHECKING, List, Optional, Tuple from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_gpta import GPTaTokenizer if ...
335
1
import copy import random from transformers import CLIPTokenizer class UpperCamelCase ( _UpperCAmelCase ): def __init__( self , *UpperCAmelCase__ , **UpperCAmelCase__ ): super().__init__(*UpperCAmelCase__ , **UpperCAmelCase__ ) A__ = {} ...
491
import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging UpperCAmelCase_ : Dict = logging.get_logger(__name__) # pylint: disable=invalid-name class UpperCa...
491
1
"""simple docstring""" import json import multiprocessing as mp import re from collections import defaultdict from functools import partial from typing import Dict, List, Optional, Set, Tuple, Type from datasets import Dataset from datasketch import MinHash, MinHashLSH from dpu_utils.util...
714
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) lowerCamelCase : Tuple = {"""configuration_deit""": ["""DEIT_PRE...
168
0
'''simple docstring''' from sklearn.metrics import fa_score import datasets SCREAMING_SNAKE_CASE = '\nThe F1 score is the harmonic mean of the precision and recall. It can be computed with the equation:\nF1 = 2 * (precision * recall) / (precision + recall)\n' SCREAMING_SNAKE_CASE = ...
94
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): f...
94
1
"""simple docstring""" import torch from transformers import PreTrainedModel, XLMRobertaConfig, XLMRobertaModel class lowercase__ ( snake_case__ ): _UpperCAmelCase :str = "M-CLIP" def __init__( self : List[Any] , snake_case__ : Tuple=1024 , sna...
244
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mobilebert import MobileBertTokenizer A__ : List[str] = logging....
244
1
from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record _UpperCamelCase = """\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and P...
243
"""simple docstring""" import warnings from ...utils import logging from .image_processing_donut import DonutImageProcessor SCREAMING_SNAKE_CASE__:Dict = logging.get_logger(__name__) class snake_case__ ( snake_case_ ): def __init__( self , *lowerCamelCase , **lowerCame...
528
0
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) SCREA...
715
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger SCREAMING_SNAKE_CASE__ = get_logger(__name__) class __lowerCamelCase ( enum.Enum ): """simple docstring""" lowerCAmelCase__ =...
601
0