code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _A : List[Any] ={} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvail...
721
'''simple docstring''' import os from collections.abc import Iterator def __UpperCamelCase ( _lowercase = "." ) -> Iterator[str]: for dir_path, dir_names, filenames in os.walk(_lowercase ): _lowercase : Optional[int] = [d for d in dir_names if d != 'scripts' and ...
4
0
'''simple docstring''' import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotConfig, is_flax_available from transformers.testing_utils import jax_device, require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_...
700
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _A : Union[str, Any] ={'''configuration_reformer''': ['''REFORMER_PRETRAINED_...
4
0
'''simple docstring''' import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer _A : List[str] =logging.getLogger(__name__) def __UpperCamelCase ( ) -> List[str]: '''simple docstring''' _lowercase ...
701
'''simple docstring''' import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.n...
4
0
import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to h...
702
'''simple docstring''' import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFe...
4
0
'''simple docstring''' from unittest import TestCase from datasets import Sequence, Value from datasets.arrow_dataset import Dataset class lowerCamelCase__ ( lowercase__ ): '''simple docstring''' def __UpperCAmelCase ( self : str ) -> Any: ''...
703
'''simple docstring''' from __future__ import annotations import requests def __UpperCamelCase ( _lowercase ) -> dict: _lowercase : Optional[int] = f'''https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty''' return requests.get(_lowercase ).jso...
4
0
'''simple docstring''' import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def __UpperCam...
704
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : Dict =logging.get_logger(__name__) _A : Dict ={ # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class lowerCamelCase__ ( A...
4
0
'''simple docstring''' from __future__ import annotations def __UpperCamelCase ( _lowercase ) -> int: # This function is recursive _lowercase : Optional[int] = len(snake_case__ ) # If the array contains only one element, we return it (it's the stop condition of ...
705
'''simple docstring''' import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def __UpperCamelCase ( _lowercase ) -> List[Any]: _lowercase : Tuple = args.pruning_method _lowercase : ...
4
0
'''simple docstring''' from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxT...
706
'''simple docstring''' _A : Optional[Any] ='''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def __UpperCamelCase ( _lowercase ) -> bytes: # Make sure the supplied data is a bytes-like object if not isinstance(_lowercase, _lowercase ): _lo...
4
0
'''simple docstring''' from __future__ import annotations import math _A : Union[str, Any] ="2020.9.26" _A : Any ="xcodz-dot, cclaus, dhruvmanila" def __UpperCamelCase ( _lowercase, _lowercase, _lowercase, _lowercase, _lowercase ) -> tuple[f...
707
'''simple docstring''' def __UpperCamelCase ( _lowercase ) -> bool: return str(_lowercase ) == str(_lowercase )[::-1] def __UpperCamelCase ( _lowercase ) -> int: return int(_lowercase ) + int(str(_lowercase )[::-1] ) def __UpperCam...
4
0
'''simple docstring''' import argparse import os from io import BytesIO from pathlib import Path import requests from clip_retrieval.clip_client import ClipClient from PIL import Image from tqdm import tqdm def __UpperCamelCase ( _lowercase, _lowercase, _lowercase ) -> str: _low...
708
'''simple docstring''' import argparse from collections import defaultdict def __UpperCamelCase ( _lowercase, _lowercase, _lowercase, _lowercase, _lowercase ) -> int: _lowercase : Optional[int] = f'''{file}_{class_name}_{test_name}''' done_test[_id] += 1 ...
4
0
'''simple docstring''' import argparse import json from typing import List from ltp import LTP from transformers.models.bert.tokenization_bert import BertTokenizer def __UpperCamelCase ( _lowercase ) -> str: if ( (cp >= 0x4E_00 and cp <= 0x9F_FF) or (cp >= 0x34_00 and cp <...
709
'''simple docstring''' from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available()...
4
0
'''simple docstring''' from __future__ import annotations def __UpperCamelCase ( _lowercase ) -> Optional[int]: _lowercase : Union[str, Any] = str(_snake_case ) return n == n[::-1] def __UpperCamelCase ( _lowercase = 100_0000 ) -> An...
710
'''simple docstring''' from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def __UpperCamelCase ( _lowercase ) -> None: _lowercase , _lowercase : List[Any] = analyze_text(_lowercase ) _lowercase ...
4
0
'''simple docstring''' def __UpperCamelCase ( _lowercase ) -> Any: # noqa: E741 _lowercase : str = len(_lowercase ) _lowercase : Union[str, Any] = 0 _lowercase : Any = [0] * n _lowercase : Dict = [False] * n _lowerca...
711
'''simple docstring''' import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sent...
4
0
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_OBJECT_DETECTION_MAPPING, AutoFeatureExtractor, AutoModelForObjectDetection, ObjectDetectionPipeline, is_vision_available, pipeline, ) from transformers.testing_utils import ( is_pipeline_test, nested_simpl...
712
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING _A : int =logging.get_logger(__name_...
4
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) _A : str ={'''configuration_vit_mae''': ['''VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMA...
713
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common impo...
4
0
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is...
714
'''simple docstring''' _A : Dict =''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' _A : ...
4
0
'''simple docstring''' import math def __UpperCamelCase ( _lowercase, _lowercase ) -> float: return math.pow(_lowerCAmelCase, 2 ) - a def __UpperCamelCase ( _lowercase ) -> float: return 2 * x def __UpperCamelCase ( _lowerca...
715
'''simple docstring''' import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, Bert...
4
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import ( Diffusio...
716
'''simple docstring''' import unittest from knapsack import greedy_knapsack as kp class lowerCamelCase__ ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : int ) -> Any: '''simple docstring''' _lowercase ...
4
0
'''simple docstring''' import warnings from .generation import TFGenerationMixin class lowerCamelCase__ ( UpperCAmelCase__ ): '''simple docstring''' warnings.warn( """Importing `TFGenerationMixin` from `src/transformers/generation_tf_utils.py` is deprecated and will ""...
717
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _A : Optional[Any] ={'''configuration_xlnet''': ['''XLNE...
4
0
'''simple docstring''' def __UpperCamelCase ( _lowercase, _lowercase, _lowercase ) -> Union[str, Any]: if len(lowercase__ ) != len(lowercase__ ): raise ValueError('The length of profit and weight must be same.' ) if max_weight <= 0: raise ValueError('max_weight m...
718
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : Optional[Any] =logging.get_logger(__name__) _A : Optional[int] ={ '''microsoft/markuplm-base''': '''https://huggingface.co/microsoft/markuplm-base/resolve/main/config....
4
0
from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment,...
719
'''simple docstring''' import argparse import requests import torch from PIL import Image from torchvision.transforms import Compose, Normalize, Resize, ToTensor from transformers import SwinaSRConfig, SwinaSRForImageSuperResolution, SwinaSRImageProcessor def __UpperCamelCase ( _lowercase ...
4
0
'''simple docstring''' def __UpperCamelCase ( _lowercase ) -> Union[str, Any]: _lowercase : Dict = len(_UpperCamelCase ) for i in range(length - 1 ): _lowercase : Any = i for k in range(i + 1, _UpperCamelCase ): if collection[k] < colle...
720
'''simple docstring''' def __UpperCamelCase ( _lowercase, _lowercase ) -> list: _lowercase : List[str] = word.split() def justify(_lowercase, _lowercase, _lowercase ) -> str: _lowercase : Dict = max_width - width _lowercase : Tupl...
4
0
'''simple docstring''' from __future__ import annotations from scipy.special import comb # type: ignore class lowerCamelCase__ : '''simple docstring''' def __init__( self : Optional[Any] , UpperCamelCase_ : Tuple ) -> List[Any]: '''simple docstring''' ...
721
'''simple docstring''' import os from collections.abc import Iterator def __UpperCamelCase ( _lowercase = "." ) -> Iterator[str]: for dir_path, dir_names, filenames in os.walk(_lowercase ): _lowercase : Optional[int] = [d for d in dir_names if d != 'scripts' and ...
4
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_...
700
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _A : Union[str, Any] ={'''configuration_reformer''': ['''REFORMER_PRETRAINED_...
4
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _A : int ={ '''configuration_roberta''': ['''ROBERTA_PRETRAIN...
701
'''simple docstring''' import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.n...
4
0
from __future__ import annotations import typing from collections import Counter def __UpperCamelCase ( _lowercase ) -> Optional[Any]: _lowercase : typing.Counter[int] = Counter() for base in range(1, max_perimeter + 1 ): for perpendicular in range(lowerCAmel...
702
'''simple docstring''' import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFe...
4
0
'''simple docstring''' from copy import deepcopy class lowerCamelCase__ : '''simple docstring''' def __init__( self : Optional[Any] , UpperCamelCase_ : Any = None , UpperCamelCase_ : List[str] = None ) -> Union[str, Any]: '''simple docstring''' ...
703
'''simple docstring''' from __future__ import annotations import requests def __UpperCamelCase ( _lowercase ) -> dict: _lowercase : Optional[int] = f'''https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty''' return requests.get(_lowercase ).jso...
4
0
'''simple docstring''' from math import ceil def __UpperCamelCase ( _lowercase = 1001 ) -> int: _lowercase : Union[str, Any] = 1 for i in range(1, int(ceil(n / 2.0 ) ) ): _lowercase : Any = 2 * i + 1 _lowercase : str ...
704
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : Dict =logging.get_logger(__name__) _A : Dict ={ # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class lowerCamelCase__ ( A...
4
0
'''simple docstring''' import os def __UpperCamelCase ( _lowercase = "input.txt" ) -> Optional[Any]: with open(os.path.join(os.path.dirname(_lowercase ), _lowercase ) ) as input_file: _lowercase : List[str] = [ [int(_lowercase ) for ele...
705
'''simple docstring''' import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def __UpperCamelCase ( _lowercase ) -> List[Any]: _lowercase : Tuple = args.pruning_method _lowercase : ...
4
0
'''simple docstring''' import gc import unittest import numpy as np import torch from torch.backends.cuda import sdp_kernel from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) from diffusers.utils import randn_tensor, slow, torch_device from diffusers.utils.te...
706
'''simple docstring''' _A : Optional[Any] ='''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def __UpperCamelCase ( _lowercase ) -> bytes: # Make sure the supplied data is a bytes-like object if not isinstance(_lowercase, _lowercase ): _lo...
4
0
'''simple docstring''' def __UpperCamelCase ( _lowercase = 6008_5147_5143 ) -> int: try: _lowercase : Optional[int] = int(__UpperCamelCase ) except (TypeError, ValueError): raise TypeError('Parameter n must be int or castable to int.' ) if n <= 0: raise Value...
707
'''simple docstring''' def __UpperCamelCase ( _lowercase ) -> bool: return str(_lowercase ) == str(_lowercase )[::-1] def __UpperCamelCase ( _lowercase ) -> int: return int(_lowercase ) + int(str(_lowercase )[::-1] ) def __UpperCam...
4
0
'''simple docstring''' import argparse from typing import List import evaluate import numpy as np import torch from datasets import DatasetDict, load_dataset # New Code # # We'll be using StratifiedKFold for this example from sklearn.model_selection import StratifiedKFold from torch.optim import AdamW from torch...
708
'''simple docstring''' import argparse from collections import defaultdict def __UpperCamelCase ( _lowercase, _lowercase, _lowercase, _lowercase, _lowercase ) -> int: _lowercase : Optional[int] = f'''{file}_{class_name}_{test_name}''' done_test[_id] += 1 ...
4
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_...
709
'''simple docstring''' from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available()...
4
0
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import BeitConfig, BeitForImageClassification, BeitForMaskedImageModeling, BeitImageProcessor from transformers.image_uti...
710
'''simple docstring''' from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def __UpperCamelCase ( _lowercase ) -> None: _lowercase , _lowercase : List[Any] = analyze_text(_lowercase ) _lowercase ...
4
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _A : Optional[int] = logging.get_logger(__name__) _A : List[str] = { ''...
711
'''simple docstring''' import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sent...
4
0
'''simple docstring''' import argparse import os import pickle import sys import torch from transformers import TransfoXLConfig, TransfoXLLMHeadModel, load_tf_weights_in_transfo_xl from transformers.models.transfo_xl import tokenization_transfo_xl as data_utils from transformers.models.transfo_xl.tokenization_tr...
712
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING _A : int =logging.get_logger(__name_...
4
0
'''simple docstring''' import csv from collections import defaultdict from dataclasses import dataclass, field from typing import List, Optional import matplotlib.pyplot as plt import numpy as np from matplotlib.ticker import ScalarFormatter from transformers import HfArgumentParser def __UpperCamel...
713
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common impo...
4
0
'''simple docstring''' import argparse import gc import json import os import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from ac...
714
'''simple docstring''' _A : Dict =''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' _A : ...
4
0
'''simple docstring''' def __UpperCamelCase ( _lowercase, _lowercase, _lowercase ) -> str: return round(float(moles / volume ) * nfactor ) def __UpperCamelCase ( _lowercase, _lowercase, _lowercase ) -> List[str]: return round(float((moles * 0....
715
'''simple docstring''' import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, Bert...
4
0
'''simple docstring''' from __future__ import annotations _A : Dict ='''#''' class lowerCamelCase__ : '''simple docstring''' def __init__( self : List[Any] ) -> None: '''simple docstring''' _lowercase : dict = {} ...
716
'''simple docstring''' import unittest from knapsack import greedy_knapsack as kp class lowerCamelCase__ ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : int ) -> Any: '''simple docstring''' _lowercase ...
4
0
'''simple docstring''' import pyarrow.parquet as pq import pytest from datasets import Audio, Dataset, DatasetDict, Features, NamedSplit, Sequence, Value, config from datasets.features.image import Image from datasets.io.parquet import ParquetDatasetReader, ParquetDatasetWriter, get_writer_batch_size from ..util...
717
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _A : Optional[Any] ={'''configuration_xlnet''': ['''XLNE...
4
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_poolformer import PoolFormerImageProcessor _A : str =logging.get_logger(__name__) class lowerCamelCase__ ( __snake_case ): '''simple docstring''' def __init__( self ...
718
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : Optional[Any] =logging.get_logger(__name__) _A : Optional[int] ={ '''microsoft/markuplm-base''': '''https://huggingface.co/microsoft/markuplm-base/resolve/main/config....
4
0
_A : int =tuple[float, float, float] _A : Optional[Any] =tuple[float, float, float] def __UpperCamelCase ( _lowercase, _lowercase ) -> Vectorad: _lowercase : List[Any] = end_pointa[0] - end_pointa[0] _lowercase : Union[str, ...
719
'''simple docstring''' import argparse import requests import torch from PIL import Image from torchvision.transforms import Compose, Normalize, Resize, ToTensor from transformers import SwinaSRConfig, SwinaSRForImageSuperResolution, SwinaSRImageProcessor def __UpperCamelCase ( _lowercase ...
4
0
'''simple docstring''' import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def __UpperCamelCase ( ) -> Tuple: raise RuntimeError('CUDA out of memory.' ) clas...
720
'''simple docstring''' def __UpperCamelCase ( _lowercase, _lowercase ) -> list: _lowercase : List[str] = word.split() def justify(_lowercase, _lowercase, _lowercase ) -> str: _lowercase : Dict = max_width - width _lowercase : Tupl...
4
0
'''simple docstring''' import json import os import re import unittest from transformers import CodeGenTokenizer, CodeGenTokenizerFast from transformers.models.codegen.tokenization_codegen import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common im...
721
'''simple docstring''' import os from collections.abc import Iterator def __UpperCamelCase ( _lowercase = "." ) -> Iterator[str]: for dir_path, dir_names, filenames in os.walk(_lowercase ): _lowercase : Optional[int] = [d for d in dir_names if d != 'scripts' and ...
4
0
'''simple docstring''' import argparse import pathlib import fairseq import torch from fairseq.models.roberta import RobertaModel as FairseqRobertaModel from fairseq.modules import TransformerSentenceEncoderLayer from packaging import version from transformers import XLMRobertaConfig, XLMRobertaXLForMaskedLM, XL...
700
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _A : Union[str, Any] ={'''configuration_reformer''': ['''REFORMER_PRETRAINED_...
4
0
'''simple docstring''' from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class lowerCamelCase__ ( _UpperCAmelCase ): '''simple docst...
701
'''simple docstring''' import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.n...
4
0
from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Imag...
702
'''simple docstring''' import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFe...
4
0
'''simple docstring''' import os # Precomputes a list of the 100 first triangular numbers _A : List[Any] =[int(0.5 * n * (n + 1)) for n in range(1, 1_0_1)] def __UpperCamelCase ( ) -> Union[str, Any]: _lowercase : str = os.path.dirname(os.path.realpath(_...
703
'''simple docstring''' from __future__ import annotations import requests def __UpperCamelCase ( _lowercase ) -> dict: _lowercase : Optional[int] = f'''https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty''' return requests.get(_lowercase ).jso...
4
0
'''simple docstring''' import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from tr...
704
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : Dict =logging.get_logger(__name__) _A : Dict ={ # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class lowerCamelCase__ ( A...
4
0
'''simple docstring''' # Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # ...
705
'''simple docstring''' import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def __UpperCamelCase ( _lowercase ) -> List[Any]: _lowercase : Tuple = args.pruning_method _lowercase : ...
4
0
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # #...
706
'''simple docstring''' _A : Optional[Any] ='''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def __UpperCamelCase ( _lowercase ) -> bytes: # Make sure the supplied data is a bytes-like object if not isinstance(_lowercase, _lowercase ): _lo...
4
0
'''simple docstring''' _A : List[str] ='''Input must be a string of 8 numbers plus letter''' _A : str ='''TRWAGMYFPDXBNJZSQVHLCKE''' def __UpperCamelCase ( _lowercase ) -> int: if not isinstance(_snake_case, _snake_case ): _lowercase :...
707
'''simple docstring''' def __UpperCamelCase ( _lowercase ) -> bool: return str(_lowercase ) == str(_lowercase )[::-1] def __UpperCamelCase ( _lowercase ) -> int: return int(_lowercase ) + int(str(_lowercase )[::-1] ) def __UpperCam...
4
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _A : List[str] ={ 'configuration_gpt_neo': ['GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GPTNeoConfig', 'GPTNeoOnnxConfig'], } try: ...
708
'''simple docstring''' import argparse from collections import defaultdict def __UpperCamelCase ( _lowercase, _lowercase, _lowercase, _lowercase, _lowercase ) -> int: _lowercase : Optional[int] = f'''{file}_{class_name}_{test_name}''' done_test[_id] += 1 ...
4
0
'''simple docstring''' # # This a `torch.distributed` diagnostics script that checks that all GPUs in the cluster (one or # many nodes) can talk to each other via nccl and allocate gpu memory. # # To run first adjust the number of processes and nodes: # # python -m torch.distributed.run --nproc_per_node 2 --nnodes...
709
'''simple docstring''' from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available()...
4
0
'''simple docstring''' # This code is adapted from OpenAI's release # https://github.com/openai/human-eval/blob/master/human_eval/execution.py import contextlib import faulthandler import io import multiprocessing import os import platform import signal import tempfile def __UpperCamelCase ( ...
710
'''simple docstring''' from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def __UpperCamelCase ( _lowercase ) -> None: _lowercase , _lowercase : List[Any] = analyze_text(_lowercase ) _lowercase ...
4
0
'''simple docstring''' import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask _A : Any = logging.getLogger(__name__) class lowerCamelCase__ ( A ): '''simple do...
711
'''simple docstring''' import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sent...
4
0
'''simple docstring''' from ...utils import logging from ..ta.modeling_tf_ta import TFTaEncoderModel, TFTaForConditionalGeneration, TFTaModel from .configuration_mta import MTaConfig _A : Any =logging.get_logger(__name__) _A : Dict ='''T5Config''' class lowerCamelCase__ ...
712
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING _A : int =logging.get_logger(__name_...
4
0
'''simple docstring''' from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class lowerCamelCase__ ( __lowerCamelCase ): '''simple doc...
713
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common impo...
4
0
'''simple docstring''' import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class lowerCamelCase__ ( pl.LightningModule ): '''simple docstring''' def __init__( self : List[...
714
'''simple docstring''' _A : Dict =''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' _A : ...
4
0
'''simple docstring''' from collections import OrderedDict from typing import TYPE_CHECKING, Any, List, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import TensorType, logging if TYPE_CHECKING: from ...onnx.config import PatchingSpe...
715
'''simple docstring''' import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, Bert...
4
0
'''simple docstring''' import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTe...
716
'''simple docstring''' import unittest from knapsack import greedy_knapsack as kp class lowerCamelCase__ ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : int ) -> Any: '''simple docstring''' _lowercase ...
4
0
'''simple docstring''' import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class lowerCamelCase__ : '''simple docstring''' @pr...
717
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _A : Optional[Any] ={'''configuration_xlnet''': ['''XLNE...
4
0
'''simple docstring''' from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_ima...
718
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : Optional[Any] =logging.get_logger(__name__) _A : Optional[int] ={ '''microsoft/markuplm-base''': '''https://huggingface.co/microsoft/markuplm-base/resolve/main/config....
4
0
def __UpperCamelCase ( _lowercase ) -> Union[str, Any]: if a < 0: raise ValueError('Input value must be a positive integer' ) elif isinstance(__a, __a ): raise TypeError('Input value must be a \'int\' type' ) return bin(__a ).count('1' ) if __name__ == "__mai...
719
'''simple docstring''' import argparse import requests import torch from PIL import Image from torchvision.transforms import Compose, Normalize, Resize, ToTensor from transformers import SwinaSRConfig, SwinaSRForImageSuperResolution, SwinaSRImageProcessor def __UpperCamelCase ( _lowercase ...
4
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _A : Union[str, Any] ={ """configuration_nezha""": ["""NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NezhaConfig"""], } try: i...
720
'''simple docstring''' def __UpperCamelCase ( _lowercase, _lowercase ) -> list: _lowercase : List[str] = word.split() def justify(_lowercase, _lowercase, _lowercase ) -> str: _lowercase : Dict = max_width - width _lowercase : Tupl...
4
0
'''simple docstring''' def __UpperCamelCase ( _lowercase ) -> list[int]: if num <= 0: raise ValueError('Input must be a positive integer' ) _lowercase : List[Any] = [True] * (num + 1) _lowercase : Optional[int] = 2 while p * p <= num: if primes[...
721
'''simple docstring''' import os from collections.abc import Iterator def __UpperCamelCase ( _lowercase = "." ) -> Iterator[str]: for dir_path, dir_names, filenames in os.walk(_lowercase ): _lowercase : Optional[int] = [d for d in dir_names if d != 'scripts' and ...
4
0
'''simple docstring''' import tempfile import unittest import numpy as np from diffusers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionPipeline, PNDMScheduler, ) from diffusers.uti...
700
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _A : Union[str, Any] ={'''configuration_reformer''': ['''REFORMER_PRETRAINED_...
4
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaInpaintPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffusers.utils import floats_tensor, load_...
701
'''simple docstring''' import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.n...
4
0
import sys from typing import Tuple import numpy as np import torch from PIL import Image from torch import nn from transformers.image_utils import PILImageResampling from utils import img_tensorize class lowerCamelCase__ : '''simple docstring''' def __init__( self : str , Uppe...
702
'''simple docstring''' import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFe...
4
0
'''simple docstring''' import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": _A : Union[str, Any] =pd.read_csv('''sample_data.csv''', header...
703
'''simple docstring''' from __future__ import annotations import requests def __UpperCamelCase ( _lowercase ) -> dict: _lowercase : Optional[int] = f'''https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty''' return requests.get(_lowercase ).jso...
4
0
'''simple docstring''' import darl # noqa import gym import tqdm from diffusers.experimental import ValueGuidedRLPipeline _A : str ={ "n_samples": 6_4, "horizon": 3_2, "num_inference_steps": 2_0, "n_guide_steps": 2, # can set to 0 for faster sampling, does not use value netwo...
704
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : Dict =logging.get_logger(__name__) _A : Dict ={ # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class lowerCamelCase__ ( A...
4
0
'''simple docstring''' def __UpperCamelCase ( _lowercase, _lowercase ) -> int: return 1 if input_a == input_a else 0 def __UpperCamelCase ( ) -> None: assert xnor_gate(0, 0 ) == 1 assert xnor_gate(0, 1 ) == 0 assert xnor_gate(1, 0 ) == 0 ass...
705
'''simple docstring''' import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def __UpperCamelCase ( _lowercase ) -> List[Any]: _lowercase : Tuple = args.pruning_method _lowercase : ...
4
0
'''simple docstring''' import json from typing import Iterator, List, Union from tokenizers import AddedToken, Regex, Tokenizer, decoders, normalizers, pre_tokenizers, trainers from tokenizers.implementations.base_tokenizer import BaseTokenizer from tokenizers.models import Unigram from tokenizers.processors import T...
706
'''simple docstring''' _A : Optional[Any] ='''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def __UpperCamelCase ( _lowercase ) -> bytes: # Make sure the supplied data is a bytes-like object if not isinstance(_lowercase, _lowercase ): _lo...
4
0
'''simple docstring''' import gzip import hashlib import json import multiprocessing import os import re import shutil import time from pathlib import Path import numpy as np from arguments import PreprocessingArguments from datasets import load_dataset from minhash_deduplication import deduplicate_dataset from ...
707
'''simple docstring''' def __UpperCamelCase ( _lowercase ) -> bool: return str(_lowercase ) == str(_lowercase )[::-1] def __UpperCamelCase ( _lowercase ) -> int: return int(_lowercase ) + int(str(_lowercase )[::-1] ) def __UpperCam...
4
0
'''simple docstring''' import math def __UpperCamelCase ( _lowercase ) -> Optional[Any]: assert isinstance(lowerCAmelCase__, lowerCAmelCase__ ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif num...
708
'''simple docstring''' import argparse from collections import defaultdict def __UpperCamelCase ( _lowercase, _lowercase, _lowercase, _lowercase, _lowercase ) -> int: _lowercase : Optional[int] = f'''{file}_{class_name}_{test_name}''' done_test[_id] += 1 ...
4
0
'''simple docstring''' import math import flax.linen as nn import jax.numpy as jnp def __UpperCamelCase ( _lowercase, _lowercase, _lowercase = 1, _lowercase = 1, _lowercase = 1.0E4, _lowercase = False, _lowercase = 1.0, ) -> jnp.ndarray: assert timesteps.ndim == 1, "T...
709
'''simple docstring''' from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available()...
4
0
'''simple docstring''' def __UpperCamelCase ( _lowercase ) -> int: assert column_title.isupper() _lowercase : int = 0 _lowercase : Optional[int] = len(_lowercase ) - 1 _lowercase : Optional[Any] = 0 while index >= 0: _lowercase ...
710
'''simple docstring''' from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def __UpperCamelCase ( _lowercase ) -> None: _lowercase , _lowercase : List[Any] = analyze_text(_lowercase ) _lowercase ...
4
0
'''simple docstring''' def __UpperCamelCase ( _lowercase ) -> list[int]: if num <= 0: raise ValueError('Input must be a positive integer' ) _lowercase : int = [True] * (num + 1) _lowercase : int = 2 while p * p <= num: if primes[p]: for i in...
711
'''simple docstring''' import unittest from transformers import AutoTokenizer, is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, slow if is_flax_available(): import jax.numpy as jnp from transformers import FlaxXLMRobertaModel @require_sent...
4
0
'''simple docstring''' from math import isqrt, loga def __UpperCamelCase ( _lowercase ) -> list[int]: _lowercase : int = [True] * max_number for i in range(2, isqrt(max_number - 1 ) + 1 ): if is_prime[i]: for j in range(i**2, a_, a_ ): _l...
712
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING _A : int =logging.get_logger(__name_...
4
0
'''simple docstring''' def __UpperCamelCase ( _lowercase, _lowercase ) -> Dict: if mass < 0: raise ValueError('The mass of a body cannot be negative' ) return 0.5 * mass * abs(UpperCAmelCase__ ) * abs(UpperCAmelCase__ ) if __name__ == "__main__": import doctest ...
713
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common impo...
4
0
'''simple docstring''' _A : str =[0, 2, 4, 6, 8] _A : List[Any] =[1, 3, 5, 7, 9] def __UpperCamelCase ( _lowercase, _lowercase, _lowercase, _lowercase ) -> int: if remaining_length == 0: if digits[0] == 0 or digits[-1] == 0: return 0 for...
714
'''simple docstring''' _A : Dict =''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' _A : ...
4
0
'''simple docstring''' import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(A ) , ""...
715
'''simple docstring''' import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, Bert...
4
0
'''simple docstring''' # DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .schedulin...
716
'''simple docstring''' import unittest from knapsack import greedy_knapsack as kp class lowerCamelCase__ ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self : int ) -> Any: '''simple docstring''' _lowercase ...
4
0
'''simple docstring''' import argparse import os import shutil from pathlib import Path import onnx import torch from packaging import version from torch.onnx import export from diffusers import OnnxRuntimeModel, OnnxStableDiffusionPipeline, StableDiffusionPipeline _A : Optional[Any] =version.p...
717
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _A : Optional[Any] ={'''configuration_xlnet''': ['''XLNE...
4
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : List[str] =logging.get_logger(__name__) _A : Optional[int] ={ 'facebook/nllb-moe-54B': 'https://huggingface.co/facebook/nllb-moe-54b/resolve/main/config.json', } cl...
718
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : Optional[Any] =logging.get_logger(__name__) _A : Optional[int] ={ '''microsoft/markuplm-base''': '''https://huggingface.co/microsoft/markuplm-base/resolve/main/config....
4
0
import os import sys _A : List[Any] =os.path.join(os.path.dirname(__file__), '''src''') sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification,...
719
'''simple docstring''' import argparse import requests import torch from PIL import Image from torchvision.transforms import Compose, Normalize, Resize, ToTensor from transformers import SwinaSRConfig, SwinaSRForImageSuperResolution, SwinaSRImageProcessor def __UpperCamelCase ( _lowercase ...
4
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import _LazyModule _A : Optional[int] ={'''tokenization_wav2vec2_phoneme''': ['''Wav2Vec2PhonemeCTCTokenizer''']} if TYPE_CHECKING: from .tokenization_wavaveca_phoneme import WavaVecaPhonemeCTCTokenizer else: import sys ...
720
'''simple docstring''' def __UpperCamelCase ( _lowercase, _lowercase ) -> list: _lowercase : List[str] = word.split() def justify(_lowercase, _lowercase, _lowercase ) -> str: _lowercase : Dict = max_width - width _lowercase : Tupl...
4
0
'''simple docstring''' import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis_float32 (there's also the fix_lavis branch) # also note: to convert Vicuna checkpoints, we ...
721
'''simple docstring''' import os from collections.abc import Iterator def __UpperCamelCase ( _lowercase = "." ) -> Iterator[str]: for dir_path, dir_names, filenames in os.walk(_lowercase ): _lowercase : Optional[int] = [d for d in dir_names if d != 'scripts' and ...
4
0
'''simple docstring''' import argparse import logging import os from pathlib import Path from typing import Any, Dict import pytorch_lightning as pl from pytorch_lightning.utilities import rank_zero_info from transformers import ( AdamW, AutoConfig, AutoModel, AutoModelForPreTraining, AutoMod...
700
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) _A : Union[str, Any] ={'''configuration_reformer''': ['''REFORMER_PRETRAINED_...
4
0
'''simple docstring''' import gc import random import tempfile import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, ControlNetModel, DDIMScheduler, StableDiffusionControlN...
701
'''simple docstring''' import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.n...
4
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _A : Optional[Any] =logging.get_logger(__name__) _A : Optional[int] ={ '''bigcode/gpt_bigcode-santacoder''': '''https://huggingface.co/bigcode/gpt_bigcode-santacoder/resolve/main/config.json''', }...
702
'''simple docstring''' import copy from typing import Any, Dict, List, Optional, Union import numpy as np import torch from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFe...
4
0
'''simple docstring''' from __future__ import annotations import os from typing import Any import requests _A : int ='''https://api.github.com''' # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user _A : int =BASE_URL + '''/user''' ...
703
'''simple docstring''' from __future__ import annotations import requests def __UpperCamelCase ( _lowercase ) -> dict: _lowercase : Optional[int] = f'''https://hacker-news.firebaseio.com/v0/item/{story_id}.json?print=pretty''' return requests.get(_lowercase ).jso...
4
0
'''simple docstring''' import os def __UpperCamelCase ( _lowercase = "input.txt" ) -> int: with open(os.path.join(os.path.dirname(_lowercase ), _lowercase ) ) as input_file: _lowercase : Optional[Any] = [ [int(_lowercase ) for element in ...
704
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _A : Dict =logging.get_logger(__name__) _A : Dict ={ # See all MEGATRON_BERT models at https://huggingface.co/models?filter=bert } class lowerCamelCase__ ( A...
4
0
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .modeling_utils import ModelMixin from .vae import Decode...
705
'''simple docstring''' import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def __UpperCamelCase ( _lowercase ) -> List[Any]: _lowercase : Tuple = args.pruning_method _lowercase : ...
4
0
'''simple docstring''' from __future__ import annotations def __UpperCamelCase ( _lowercase ) -> int: if not nums: return 0 _lowercase : Tuple = nums[0] _lowercase : Optional[int] = 0 for num in nums[1:]: _lowercase : int = ( ...
706
'''simple docstring''' _A : Optional[Any] ='''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/''' def __UpperCamelCase ( _lowercase ) -> bytes: # Make sure the supplied data is a bytes-like object if not isinstance(_lowercase, _lowercase ): _lo...
4
0
'''simple docstring''' import argparse from collections import defaultdict import yaml _A : List[Any] ='''docs/source/en/_toctree.yml''' def __UpperCamelCase ( _lowercase ) -> Optional[int]: _lowercase : Union[str, Any] = defaultdict(_lowercase )...
707
'''simple docstring''' def __UpperCamelCase ( _lowercase ) -> bool: return str(_lowercase ) == str(_lowercase )[::-1] def __UpperCamelCase ( _lowercase ) -> int: return int(_lowercase ) + int(str(_lowercase )[::-1] ) def __UpperCam...
4
0
'''simple docstring''' from __future__ import annotations def __UpperCamelCase ( _lowercase ) -> bool: if len(_lowercase ) < 2: raise ValueError('Monogons and Digons are not polygons in the Euclidean space' ) if any(i <= 0 for i in nums ): raise ValueError('All values ...
708
'''simple docstring''' import argparse from collections import defaultdict def __UpperCamelCase ( _lowercase, _lowercase, _lowercase, _lowercase, _lowercase ) -> int: _lowercase : Optional[int] = f'''{file}_{class_name}_{test_name}''' done_test[_id] += 1 ...
4
0
'''simple docstring''' from __future__ import annotations import os from collections.abc import Mapping _A : str =tuple[int, int] class lowerCamelCase__ : '''simple docstring''' def __init__( self : List[Any] , UpperCamelCase_ : set[int] , UpperCamelCase_ ...
709
'''simple docstring''' from collections import UserDict from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available()...
4
0
'''simple docstring''' from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def __UpperCamelCase ( _lowercase ) -> None: _lowercase : List[Any] = analyze_text(_lowercase ) _lowercase : Any ...
710
'''simple docstring''' from __future__ import annotations import math from collections import Counter from string import ascii_lowercase def __UpperCamelCase ( _lowercase ) -> None: _lowercase , _lowercase : List[Any] = analyze_text(_lowercase ) _lowercase ...
4
0