code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ...
5
import os from datetime import datetime as dt from github import Github A : Union[str, Any] = [ "good first issue", "good second issue", "good difficult issue", "enhancement", "new pipeline/model", "new scheduler", "wip", ] def lowercase_ ( ): ...
5
1
def lowercase_ ( _A : int = 2000000 ): """simple docstring""" lowerCamelCase__ : Optional[Any] = [0 for i in range(n + 1 )] lowerCamelCase__ : Union[str, Any] = 1 lowerCamelCase__ : Dict = 1 for i in range(2 ...
5
from __future__ import annotations def lowercase_ ( _A : str , _A : list[str] | None = None , _A : dict[str, float] | None = None , _A : bool = False , ): """simple docstring""" lowerCamelCase__ : Tuple = ...
5
1
import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class _lowercase ( unittest.TestCase): """simple docstring""" A__ = JukeboxTokenizer A__ = { "artist": ...
5
def lowercase_ ( _A : int ): """simple docstring""" if not isinstance(_A , _A ): lowerCamelCase__ : List[str] = F"Input value of [number={number}] must be an integer" raise TypeError(_A ) if number < 0: retu...
5
1
from typing import Dict, List, Optional from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging A : str = logging.get_logger(__name__) A : int = { "nielsr/canine-s": 2048, } # Unicode defines 1,114,112 total “codepoints” A : ...
5
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) A : Optional[int] = { "configuration_speecht5": [ "SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP", "SPEECHT...
5
1
import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( WavaVecaConfig, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaForCTC, WavaVecaForPreTraining, WavaVecaProcessor, logging, ) from...
5
from __future__ import annotations import time import numpy as np A : Dict = [8, 5, 9, 7] A : Optional[Any] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] A : Any = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1...
5
1
A : Optional[int] = "\n# Transformers installation\n! pip install transformers datasets\n# To install from source instead of the last release, comment the command above and uncomment the following one.\n# ! pip install git+https://github.com/huggingface/transformers.git\n" A : int = [{"ty...
5
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sent...
5
1
def lowercase_ ( _A : int = 3 , _A : int = 7 , _A : int = 1000000 ): """simple docstring""" lowerCamelCase__ : List[str] = 0 lowerCamelCase__ : List[str] = 1 for current_denominator in range(1 , ...
5
import cva import numpy as np class _lowercase : """simple docstring""" def __init__( self : Union[str, Any] , __lowerCamelCase : float , __lowerCamelCase : int ): '''simple docstring''' ...
5
1
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import center_crop, normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_M...
5
import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, id...
5
1
import math import random def lowercase_ ( _A : float , _A : bool = False ): """simple docstring""" if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value A : List[Any] = 0.0_2 def l...
5
import os def lowercase_ ( _A : str = "input.txt" ): """simple docstring""" with open(os.path.join(os.path.dirname(_A ) , _A ) ) as input_file: lowerCamelCase__ : List[Any] = [ [int(_A ) for element in line.split("," ...
5
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : Union[str, Any] = logging.get_logger(__name__) A : Dict = { "kssteven/ibert-roberta-base": "ht...
5
import datasets from .nmt_bleu import compute_bleu # From: https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py A : Tuple = "\\n@INPROCEEDINGS{Papineni02bleu:a,\n author = {Kishore Papineni and Salim Roukos and Todd Ward and Wei-jing Zhu},\n title = {BLEU: a Method for Autom...
5
1
from typing import List import datasets from datasets.tasks import AudioClassification from ..folder_based_builder import folder_based_builder A : Optional[int] = datasets.utils.logging.get_logger(__name__) class _lowercase ( folder_based_builder.FolderBasedBuild...
5
import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print("Googling.....") A : str = "https://www.google.com/search?q=" + " ".join(sys.argv[1:]) A : Optional[int] = requests.get(url...
5
1
from typing import Dict from .base import GenericTensor, Pipeline class _lowercase ( lowercase__): """simple docstring""" def lowerCAmelCase ( self : Tuple , __lowerCamelCase : str=None , __lowerCamelCase : ...
5
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCa...
5
1
import json from typing import Dict, List, Optional, Tuple, Union from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding, EncodedInput from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import PaddingStrategy, logging ...
5
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...
5
1
from manim import * class _lowercase ( lowercase__): """simple docstring""" def lowerCAmelCase ( self : List[str] ): '''simple docstring''' lowerCamelCase__ : str = Rectangle(height=0.5 ...
5
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : int = logging.get_logger(__name__) A : Optional[int] = { "facebook/xmod-base": "https://huggin...
5
1
from abc import ABC, abstractmethod from argparse import ArgumentParser class _lowercase ( lowercase__): """simple docstring""" @staticmethod @abstractmethod def lowerCAmelCase ( __lowerCamelCase : ArgumentParser ): ...
5
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_t...
5
1
from .testing import ( are_the_same_tensors, execute_subprocess_async, require_bnb, require_cpu, require_cuda, require_huggingface_suite, require_mps, require_multi_gpu, require_multi_xpu, require_safetensors, require_single_gpu, require_single_xpu, ...
5
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : Union[str, Any] = logging.get_logger(__name__) A : Dict = { "kssteven/ibert-roberta-base": "ht...
5
1
from typing import List, Optional, Tuple, Union import PIL import torch from torchvision import transforms from diffusers.pipeline_utils import DiffusionPipeline, ImagePipelineOutput from diffusers.schedulers import DDIMScheduler from diffusers.utils import randn_tensor A : Tuple = trans...
5
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : Dict = logging.get_logger(__name__) A : Union[str, Any] = { "roberta-base": "https://huggingfa...
5
1
from collections import defaultdict def lowercase_ ( _A : str , _A : str ): """simple docstring""" lowerCamelCase__ : Optional[int] = first_str.lower().strip() lowerCamelCase__ : Optional[Any] = second_str.lower().strip() ...
5
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tok...
5
1
import unittest from transformers import EsmConfig, is_torch_available from transformers.testing_utils import TestCasePlus, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...
5
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, loggi...
5
1
from __future__ import annotations def lowercase_ ( _A : Optional[int] , _A : int ): print(F"Vertex\tShortest Distance from vertex {src}" ) for i, d in enumerate(SCREAMING_SNAKE_CASE_ ): print(F"{i}\t\t{d}" ) def lowercase_ ( _A : Union[str,...
700
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, ...
5
0
import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint A : ...
701
def lowercase_ ( _A : int , _A : int ): """simple docstring""" if a < 0 or b < 0: raise ValueError("the value of both inputs must be positive" ) lowerCamelCase__ : List[str] = str(bin(_A ) )[2:] # remove the leading "0b" ...
5
0
import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) A : Dict = pytest.mark.integration @pytest.mark.parametrize("path" , [...
702
import os from pathlib import Path def lowercase_ ( ): """simple docstring""" from torch.utils.cpp_extension import load lowerCamelCase__ : Any = Path(_A ).resolve().parent.parent.parent / "kernels" / "deformable_detr" lowerCamelCase__ : Optiona...
5
0
from ...configuration_utils import PretrainedConfig from ...utils import logging A : Any = logging.get_logger(__name__) A : Optional[int] = { "microsoft/swinv2-tiny-patch4-window8-256": ( "https://huggingface.co/microsoft/swinv2-tiny-patch4-window8-256/resolve/main/con...
703
import os from datetime import datetime as dt from github import Github A : Union[str, Any] = [ "good first issue", "good second issue", "good difficult issue", "enhancement", "new pipeline/model", "new scheduler", "wip", ] def lowercase_ ( ): ...
5
0
from __future__ import annotations import math def lowercase_ ( _A : Tuple , _A : List[str] ): """simple docstring""" lowerCamelCase__ : str = u for i in range(1 , _lowercase ): lowerCamelCase__ : s...
704
from __future__ import annotations def lowercase_ ( _A : str , _A : list[str] | None = None , _A : dict[str, float] | None = None , _A : bool = False , ): """simple docstring""" lowerCamelCase__ : Tuple = ...
5
0
from __future__ import annotations class _lowercase : """simple docstring""" def __init__( self : str , __lowerCamelCase : int = 0 ): '''simple docstring''' lowerCamelCase__ : Optional[int] = key ...
705
def lowercase_ ( _A : int ): """simple docstring""" if not isinstance(_A , _A ): lowerCamelCase__ : List[str] = F"Input value of [number={number}] must be an integer" raise TypeError(_A ) if number < 0: retu...
5
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available A : List[str] = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "PoolFormerO...
706
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) A : Optional[int] = { "configuration_speecht5": [ "SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP", "SPEECHT...
5
0
import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class _lowercase ( __A): """simple docstring""" ...
707
from __future__ import annotations import time import numpy as np A : Dict = [8, 5, 9, 7] A : Optional[Any] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] A : Any = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1...
5
0
def lowercase_ ( _A : int ): """simple docstring""" if number > 0: raise ValueError("input must be a negative integer" ) lowerCamelCase__ : List[Any] = len(bin(__lowerCAmelCase )[3:] ) lowerCamelCase__ : Tuple = bin(abs(__lowerCAmel...
708
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sent...
5
0
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IM...
709
import cva import numpy as np class _lowercase : """simple docstring""" def __init__( self : Union[str, Any] , __lowerCamelCase : float , __lowerCamelCase : int ): '''simple docstring''' ...
5
0
import os import time import warnings from dataclasses import dataclass, field from enum import Enum from typing import List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils i...
710
import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, id...
5
0
import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import Tokenize...
711
import os def lowercase_ ( _A : str = "input.txt" ): """simple docstring""" with open(os.path.join(os.path.dirname(_A ) , _A ) ) as input_file: lowerCamelCase__ : List[Any] = [ [int(_A ) for element in line.split("," ...
5
0
from unittest.mock import Mock, patch from file_transfer.send_file import send_file @patch("socket.socket" ) @patch("builtins.open" ) def lowercase_ ( _A : Optional[int] , _A : Any ): lowerCamelCase__ : Dict = Mock() lowerCamelCase__ : str ...
712
import datasets from .nmt_bleu import compute_bleu # From: https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py A : Tuple = "\\n@INPROCEEDINGS{Papineni02bleu:a,\n author = {Kishore Papineni and Salim Roukos and Todd Ward and Wei-jing Zhu},\n title = {BLEU: a Method for Autom...
5
0
from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar A : int = TypeVar("KEY") A : Any = TypeVar("VAL") @dataclass(frozen=lowercase__ , slots=lowercase__) class _lowercase ( Generic[KEY, VAL]): ...
713
import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print("Googling.....") A : str = "https://www.google.com/search?q=" + " ".join(sys.argv[1:]) A : Optional[int] = requests.get(url...
5
0
import argparse import json import os import tensorstore as ts import torch from flax import serialization from flax.traverse_util import flatten_dict, unflatten_dict from tensorflow.io import gfile from transformers.modeling_utils import dtype_byte_size from transformers.models.switch_transformers.c...
714
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCa...
5
0
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_OBJECT_DETECTION_MAPPING, AutoFeatureExtractor, AutoModelForObjectDetection, ObjectDetectionPipeline, is_vision_available, pipeline, ) from transformers.testing_utils import ( is_pipeline...
715
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...
5
0
from __future__ import annotations import typing from collections import Counter def lowercase_ ( _A : List[Any] ): """simple docstring""" lowerCamelCase__ : typing.Counter[int] = Counter() for base in range(1 , max_perimeter + 1 ): ...
716
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : int = logging.get_logger(__name__) A : Optional[int] = { "facebook/xmod-base": "https://huggin...
5
0
from __future__ import annotations import numpy as np from numpy import floataa from numpy.typing import NDArray def lowercase_ ( _A : NDArray[floataa] , _A : NDArray[floataa] , _A : list[int] , _A : int , ): """simple docstring""" ...
717
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_t...
5
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): ...
718
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : Union[str, Any] = logging.get_logger(__name__) A : Dict = { "kssteven/ibert-roberta-base": "ht...
5
0
'''simple docstring''' import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtract...
719
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : Dict = logging.get_logger(__name__) A : Union[str, Any] = { "roberta-base": "https://huggingfa...
5
0
from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def lowercase_ ( _A : str ): """simple docstring""" lowerCamelCase__ : str = analyze_text(__UpperCamelCase ) lowerCamelCase__ : Any = li...
720
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tok...
5
0
import torch from transformers import CamembertForMaskedLM, CamembertTokenizer def lowercase_ ( _A : Tuple , _A : int , _A : List[str] , _A : Any=5 ): """simple docstring""" assert masked_input.count("<mask>" ) == 1 l...
721
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, loggi...
5
0
from ....configuration_utils import PretrainedConfig from ....utils import logging A : str = logging.get_logger(__name__) # TODO: upload to AWS A : Optional[int] = { "yjernite/retribert-base-uncased": ( "https://huggingface.co/yjernite/retribert-base-uncased/resolve/m...
700
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, ...
5
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): ...
701
def lowercase_ ( _A : int , _A : int ): """simple docstring""" if a < 0 or b < 0: raise ValueError("the value of both inputs must be positive" ) lowerCamelCase__ : List[str] = str(bin(_A ) )[2:] # remove the leading "0b" ...
5
0
# NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from ...utils import deprecate from ..controlnet.pipeline_flax_controlnet import FlaxStableDiffusionControlNetPipeline # noqa: F401 d...
702
import os from pathlib import Path def lowercase_ ( ): """simple docstring""" from torch.utils.cpp_extension import load lowerCamelCase__ : Any = Path(_A ).resolve().parent.parent.parent / "kernels" / "deformable_detr" lowerCamelCase__ : Optiona...
5
0
from __future__ import annotations from typing import Any class _lowercase : """simple docstring""" def __init__( self : Tuple , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : fl...
703
import os from datetime import datetime as dt from github import Github A : Union[str, Any] = [ "good first issue", "good second issue", "good difficult issue", "enhancement", "new pipeline/model", "new scheduler", "wip", ] def lowercase_ ( ): ...
5
0
from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attenti...
704
from __future__ import annotations def lowercase_ ( _A : str , _A : list[str] | None = None , _A : dict[str, float] | None = None , _A : bool = False , ): """simple docstring""" lowerCamelCase__ : Tuple = ...
5
0
from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPipeline, UNe...
705
def lowercase_ ( _A : int ): """simple docstring""" if not isinstance(_A , _A ): lowerCamelCase__ : List[str] = F"Input value of [number={number}] must be an integer" raise TypeError(_A ) if number < 0: retu...
5
0
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING A : int = logging.get_logger(__name__) A : i...
706
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) A : Optional[int] = { "configuration_speecht5": [ "SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP", "SPEECHT...
5
0
from ..utils import DummyObject, requires_backends class _lowercase ( metaclass=_UpperCAmelCase): """simple docstring""" A__ = ["flax", "transformers"] def __init__( self : Optional[Any] , *__lowerCamelCase ...
707
from __future__ import annotations import time import numpy as np A : Dict = [8, 5, 9, 7] A : Optional[Any] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] A : Any = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1...
5
0
import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import BatchEncoding, MarianTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import is_sentencepiece_available, is_tf_available, is_torch_availabl...
708
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sent...
5
0
import argparse import glob import logging import os import sys import time from collections import defaultdict from pathlib import Path from typing import Dict, List, Tuple import numpy as np import pytorch_lightning as pl import torch from callbacks import SeqaSeqLoggingCallback, get_checkpoint_ca...
709
import cva import numpy as np class _lowercase : """simple docstring""" def __init__( self : Union[str, Any] , __lowerCamelCase : float , __lowerCamelCase : int ): '''simple docstring''' ...
5
0
from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _lowercase ( __UpperCAmelCase): ...
710
import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, id...
5
0
import operator def lowercase_ ( _A : List[str] , _A : Union[str, Any] = False , _A : int = None ): """simple docstring""" lowerCamelCase__ : int = operator.lt if reverse else operator.gt lowerCamelCase__ : Tuple ...
711
import os def lowercase_ ( _A : str = "input.txt" ): """simple docstring""" with open(os.path.join(os.path.dirname(_A ) , _A ) ) as input_file: lowerCamelCase__ : List[Any] = [ [int(_A ) for element in line.split("," ...
5
0
import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def lowercase_ ( ): with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest....
712
import datasets from .nmt_bleu import compute_bleu # From: https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py A : Tuple = "\\n@INPROCEEDINGS{Papineni02bleu:a,\n author = {Kishore Papineni and Salim Roukos and Todd Ward and Wei-jing Zhu},\n title = {BLEU: a Method for Autom...
5
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A : Optional[int] = { "configuration_instructblip": [ "INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP", "InstructBlipConfig", "InstructBlipQFormerConfig", ...
713
import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print("Googling.....") A : str = "https://www.google.com/search?q=" + " ".join(sys.argv[1:]) A : Optional[int] = requests.get(url...
5
0
import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print("Googling.....") A : Union[str, Any] = "https://www.google.com/search?q=" + " ".join(sys.argv[1:]) A : str = requests.get(u...
714
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCa...
5
0
'''simple docstring''' import math import random from typing import Any from .hill_climbing import SearchProblem def lowercase_ ( _A : Optional[int] , _A : List[Any] = True , _A : List[str] = math.inf , _A : List[str] = -math.inf , ...
715
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...
5
0
import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, AutoMode...
716
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : int = logging.get_logger(__name__) A : Optional[int] = { "facebook/xmod-base": "https://huggin...
5
0
from ...utils import deprecate from ..controlnet.multicontrolnet import MultiControlNetModel # noqa: F401 from ..controlnet.pipeline_controlnet import StableDiffusionControlNetPipeline # noqa: F401 deprecate( "stable diffusion controlnet", "0.22.0", "Importing `StableDiffusionControlNetPipeline` or `...
717
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_t...
5
0
import unittest from diffusers import FlaxAutoencoderKL from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax from .test_modeling_common_flax import FlaxModelTesterMixin if is_flax_available(): import jax @require_flax class _low...
718
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : Union[str, Any] = logging.get_logger(__name__) A : Dict = { "kssteven/ibert-roberta-base": "ht...
5
0
'''simple docstring''' import argparse import json import os import pickle import shutil import numpy as np import torch from distiller import Distiller from lm_seqs_dataset import LmSeqsDataset from transformers import ( BertConfig, BertForMaskedLM, BertTokenizer, Dist...
719
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : Dict = logging.get_logger(__name__) A : Union[str, Any] = { "roberta-base": "https://huggingfa...
5
0
def lowercase_ ( _A : int , _A : int ): """simple docstring""" return 1 if input_a == input_a else 0 def lowercase_ ( ): """simple docstring""" assert xnor_gate(0 , 0 ) == 1 assert xnor_gate(0 , 1 ) == 0 assert xnor_gate...
720
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tok...
5
0
import argparse import os from accelerate.utils import ComputeEnvironment from .cluster import get_cluster_input from .config_args import cache_dir, default_config_file, default_yaml_config_file, load_config_from_file # noqa: F401 from .config_utils import _ask_field, _ask_options, _convert_compute_envi...
721
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, loggi...
5
0
import io import json import unittest from parameterized import parameterized from transformers import FSMTForConditionalGeneration, FSMTTokenizer from transformers.testing_utils import get_tests_dir, require_torch, slow, torch_device from utils import calculate_bleu A : int = get_tests_...
700
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, ...
5
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
701
def lowercase_ ( _A : int , _A : int ): """simple docstring""" if a < 0 or b < 0: raise ValueError("the value of both inputs must be positive" ) lowerCamelCase__ : List[str] = str(bin(_A ) )[2:] # remove the leading "0b" ...
5
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) A : Optional[Any] = { 'configuration_speecht5': [ 'SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SPEEC...
702
import os from pathlib import Path def lowercase_ ( ): """simple docstring""" from torch.utils.cpp_extension import load lowerCamelCase__ : Any = Path(_A ).resolve().parent.parent.parent / "kernels" / "deformable_detr" lowerCamelCase__ : Optiona...
5
0
import logging import os import sys from dataclasses import dataclass, field from typing import Optional import numpy as np import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor import transforme...
703
import os from datetime import datetime as dt from github import Github A : Union[str, Any] = [ "good first issue", "good second issue", "good difficult issue", "enhancement", "new pipeline/model", "new scheduler", "wip", ] def lowercase_ ( ): ...
5
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_availab...
704
from __future__ import annotations def lowercase_ ( _A : str , _A : list[str] | None = None , _A : dict[str, float] | None = None , _A : bool = False , ): """simple docstring""" lowerCamelCase__ : Tuple = ...
5
0
import math from collections.abc import Callable def lowercase_ ( _A : Callable[[float], float] , _A : float , _A : float ): """simple docstring""" lowerCamelCase__ : Dict = xa lowerCamelCase__ : int = xa while Tru...
705
def lowercase_ ( _A : int ): """simple docstring""" if not isinstance(_A , _A ): lowerCamelCase__ : List[str] = F"Input value of [number={number}] must be an integer" raise TypeError(_A ) if number < 0: retu...
5
0
from ...configuration_utils import PretrainedConfig from ...utils import logging A : Optional[Any] = logging.get_logger(__name__) A : Optional[Any] = { "facebook/vit-mae-base": "https://huggingface.co/facebook/vit-mae-base/resolve/main/config.json", # See all ViT MAE model...
706
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) A : Optional[int] = { "configuration_speecht5": [ "SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP", "SPEECHT...
5
0
import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging A : Union[str, Any] = "\\n\n" A : Dict = "\nPerplexity (PPL) is one of the most common metrics for ev...
707
from __future__ import annotations import time import numpy as np A : Dict = [8, 5, 9, 7] A : Optional[Any] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] A : Any = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1...
5
0
from __future__ import annotations def lowercase_ ( _A : Optional[Any] , _A : Optional[Any] , _A : int , _A : Union[str, Any] ): """simple docstring""" if (direction == 1 and array[indexa] > array[indexa]) or ( direction == 0 and ...
708
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sent...
5
0
import math def lowercase_ ( _A : int , _A : Dict ): """simple docstring""" if ( not isinstance(snake_case_ , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError("power_fact...
709
import cva import numpy as np class _lowercase : """simple docstring""" def __init__( self : Union[str, Any] , __lowerCamelCase : float , __lowerCamelCase : int ): '''simple docstring''' ...
5
0
import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging A : Dict = logging.get_logger(__name__) A : ...
710
import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, id...
5
0
import logging import math from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Sequence, Tuple, Union import torch from .tensor_utils import tensor_tree_map, tree_map def lowercase_ ( _A : Union[dict, list, tuple, torch.Tensor] ): """simp...
711
import os def lowercase_ ( _A : str = "input.txt" ): """simple docstring""" with open(os.path.join(os.path.dirname(_A ) , _A ) ) as input_file: lowerCamelCase__ : List[Any] = [ [int(_A ) for element in line.split("," ...
5
0
from __future__ import annotations def lowercase_ ( _A : int | float | str , _A : int | float | str ): if nth_term == "": return [""] lowerCamelCase__ : Any = int(__snake_case ) lowerCamelCase__ : Union[str, Any] = ...
712
import datasets from .nmt_bleu import compute_bleu # From: https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py A : Tuple = "\\n@INPROCEEDINGS{Papineni02bleu:a,\n author = {Kishore Papineni and Salim Roukos and Todd Ward and Wei-jing Zhu},\n title = {BLEU: a Method for Autom...
5
0
import math import sys def lowercase_ ( _A : Tuple ): """simple docstring""" if number != int(a__ ): raise ValueError("the value of input must be a natural number" ) if number < 0: raise ValueError("the value of input must not be a negative number" ) ...
713
import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print("Googling.....") A : str = "https://www.google.com/search?q=" + " ".join(sys.argv[1:]) A : Optional[int] = requests.get(url...
5
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) A : Tuple = {"configuration_deit": ["DEIT_PRETRAINED_CONFIG_ARCHIVE_MAP", "DeiTConfig", "DeiTOnnxConfig"]} ...
714
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCa...
5
0
'''simple docstring''' def lowercase_ ( _A : Union[str, Any] , _A : List[str] , _A : Dict=False ): """simple docstring""" if isinstance(_UpperCAmelCase , _UpperCAmelCase ) and isinstance(_UpperCAmelCase , _UpperCAmelC...
715
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer from ...configuration_utils import PretrainedConfig from ...file_utils import TensorType, is_torch_available from ...onnx import OnnxConfig, OnnxConfigWithPast, OnnxSeqaSeqConfigWithPast from ...
5
0
import numpy as np import qiskit def lowercase_ ( _A : int = 8 , _A : int | None = None ): """simple docstring""" lowerCamelCase__ : str = np.random.default_rng(seed=_A ) # Roughly 25% of the qubits will contribute to the key. ...
716
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : int = logging.get_logger(__name__) A : Optional[int] = { "facebook/xmod-base": "https://huggin...
5
0
import comet # From: unbabel-comet import torch import datasets A : List[str] = datasets.logging.get_logger(__name__) A : Union[str, Any] = "\\n@inproceedings{rei-EtAl:2020:WMT,\n author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon},\n title = {Unb...
717
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_t...
5
0
import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import BatchEncoding, MarianTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import is_sentencepiece_available, is_tf_available, is_torc...
718
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : Union[str, Any] = logging.get_logger(__name__) A : Dict = { "kssteven/ibert-roberta-base": "ht...
5
0
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_i...
719
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A : Dict = logging.get_logger(__name__) A : Union[str, Any] = { "roberta-base": "https://huggingfa...
5
0
import json import os import unittest from typing import Tuple from transformers import WavaVecaPhonemeCTCTokenizer from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.models.wavaveca_phoneme.tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizerOutput from t...
720
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tok...
5
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available A : str = { """configuration_poolformer""": [ """POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """PoolFormerConfig""", """Poo...
721
import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, loggi...
5
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig A : List[Any] = { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/config.json""", """albert-large-v1""": """h...
700
import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_image, ...
5
0
import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, ...
701
def lowercase_ ( _A : int , _A : int ): """simple docstring""" if a < 0 or b < 0: raise ValueError("the value of both inputs must be positive" ) lowerCamelCase__ : List[str] = str(bin(_A ) )[2:] # remove the leading "0b" ...
5
0
import os from glob import glob import imageio import torch import torchvision import wandb from img_processing import custom_to_pil, loop_post_process, preprocess, preprocess_vqgan from loaders import load_vqgan from PIL import Image from torch import nn from transformers import CLIPModel, CLIPTo...
702
import os from pathlib import Path def lowercase_ ( ): """simple docstring""" from torch.utils.cpp_extension import load lowerCamelCase__ : Any = Path(_A ).resolve().parent.parent.parent / "kernels" / "deformable_detr" lowerCamelCase__ : Optiona...
5
0
import numpy as np A : Union[str, Any] = [ ["a", "b", "c", "d", "e"], ["f", "g", "h", "i", "k"], ["l", "m", "n", "o", "p"], ["q", "r", "s", "t", "u"], ["v", "w", "x", "y", "z"], ] class _lowercase : """simple docstring""" de...
703
import os from datetime import datetime as dt from github import Github A : Union[str, Any] = [ "good first issue", "good second issue", "good difficult issue", "enhancement", "new pipeline/model", "new scheduler", "wip", ] def lowercase_ ( ): ...
5
0
from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable A : List[Any] = { "configuration_gpt_neox_japanese": ["GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoXJapaneseConfig"], "t...
704
from __future__ import annotations def lowercase_ ( _A : str , _A : list[str] | None = None , _A : dict[str, float] | None = None , _A : bool = False , ): """simple docstring""" lowerCamelCase__ : Tuple = ...
5
0
import unittest from pathlib import Path from shutil import copyfile from transformers import SPIECE_UNDERLINE, is_sentencepiece_available from transformers.models.speech_to_text import SpeechaTextTokenizer from transformers.models.speech_to_text.tokenization_speech_to_text import VOCAB_FILES_NAMES, save_json from ...
705
def lowercase_ ( _A : int ): """simple docstring""" if not isinstance(_A , _A ): lowerCamelCase__ : List[str] = F"Input value of [number={number}] must be an integer" raise TypeError(_A ) if number < 0: retu...
5
0
import os from distutils.util import strtobool def lowercase_ ( _A : Any , _A : List[str] ): """simple docstring""" for e in env_keys: lowerCamelCase__ : Any = int(os.environ.get(lowerCAmelCase_ , -1 ) ) if val >= 0: return val ret...
706
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_torch_available, ) A : Optional[int] = { "configuration_speecht5": [ "SPEECHT5_PRETRAINED_CONFIG_ARCHIVE_MAP", "SPEECHT...
5
0
import pickle import numpy as np from matplotlib import pyplot as plt class _lowercase : """simple docstring""" def __init__( self : Optional[Any] , __lowerCamelCase : List[Any] , __lowerCamelCase : List[Any] ,...
707
from __future__ import annotations import time import numpy as np A : Dict = [8, 5, 9, 7] A : Optional[Any] = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] A : Any = [ [3, 2, 1, 4], [0, 2, 5, 2], [5, 1...
5
0
# Imports import numpy as np class _lowercase : """simple docstring""" def __init__( self : Tuple , __lowerCamelCase : Tuple=None , __lowerCamelCase : Tuple=None , __lowerCamelCase : int=None , __lowerCamelCase ...
708
import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sent...
5
0
from argparse import ArgumentParser, Namespace from typing import Any, List, Optional from ..pipelines import Pipeline, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand try: from fastapi import Body, FastAPI, HTTPException from fastapi.routing...
709
import cva import numpy as np class _lowercase : """simple docstring""" def __init__( self : Union[str, Any] , __lowerCamelCase : float , __lowerCamelCase : int ): '''simple docstring''' ...
5
0
A : Any = '''Input must be a string of 8 numbers plus letter''' A : Optional[Any] = '''TRWAGMYFPDXBNJZSQVHLCKE''' def lowercase_ ( _A : Tuple ): """simple docstring""" if not isinstance(_A , _A ): lowerCamelCase__ : List[str...
710
import unittest from transformers import AlbertConfig, is_torch_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, id...
5
0
import builtins import sys from ...utils.imports import _is_package_available from . import cursor, input from .helpers import Direction, clear_line, forceWrite, linebreak, move_cursor, reset_cursor, writeColor from .keymap import KEYMAP A : List[str] = False try: A : List[Any] ...
711
import os def lowercase_ ( _A : str = "input.txt" ): """simple docstring""" with open(os.path.join(os.path.dirname(_A ) , _A ) ) as input_file: lowerCamelCase__ : List[Any] = [ [int(_A ) for element in line.split("," ...
5
0
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import CLIPTokenizer, CLIPTokenizerFast from transformers.models.clip.tokenization_clip import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transform...
712
import datasets from .nmt_bleu import compute_bleu # From: https://github.com/tensorflow/nmt/blob/master/nmt/scripts/bleu.py A : Tuple = "\\n@INPROCEEDINGS{Papineni02bleu:a,\n author = {Kishore Papineni and Salim Roukos and Todd Ward and Wei-jing Zhu},\n title = {BLEU: a Method for Autom...
5
0
import inspect import unittest import numpy as np from tests.test_modeling_common import floats_tensor from transformers import DetrConfig, MaskFormerConfig, SwinConfig, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_torch_multi_gpu, require_vision, slow, torc...
713
import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print("Googling.....") A : str = "https://www.google.com/search?q=" + " ".join(sys.argv[1:]) A : Optional[int] = requests.get(url...
5
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available A : Any = { "configuration_m2m_100": ["M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP", "M2M100Config", "M2M100OnnxConfig"], "tokenization_m2m_100": ["...
714
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFCa...
5
0