code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
import os import random import sys from . import cryptomath_module as cryptoMath # noqa: N812 from . import rabin_miller as rabinMiller # noqa: N812 def __magic_name__ ( ): '''simple docstring''' print("Making key files..." ) make_key_files("rsa", 1024 ) print("...
712
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer __lowerCAmelCase : List[Any] = logging.getLogger(__name__) def __magic_name__ ( ): '''simple docstring''' a = argparse.ArgumentParser( ...
662
0
__lowerCAmelCase : List[Any] = { 'km/h': 1.0, 'm/s': 3.6, 'mph': 1.6_0_9_3_4_4, 'knot': 1.8_5_2, } __lowerCAmelCase : Union[str, Any] = { 'km/h': 1.0, 'm/s': 0.2_7_7_7_7_7_7_7_8, 'mph': 0.6_2_1_3_7_1_1_9_2, 'knot': 0.5_3_9_9_5_6_8_0_3, } def __magic_name...
713
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def __magic_name__ ( A : List[str] ): '''simple docstring''' a = {} a = tokeni...
662
0
from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake __lowerCAmelCase : Dict = numpy.array([0, 0]) __lowerCAmelCase : Optional[int] = numpy.array([0.5, 0.8_6_6_0_2_5_4]) __lowerCAmelCase : Optional[int] ...
714
import json import multiprocessing import os import re from collections import defaultdict import torch from accelerate import Accelerator from accelerate.utils import set_seed from arguments import HumanEvalArguments from datasets import load_dataset, load_metric from torch.utils.data import IterableDataset from...
662
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase : str = logging.get_logger(__name__) __lowerCAmelCase : Dict = { "microsoft/cvt-13": "https://huggingface.co/microsoft/cvt-13/resolve/main/config.json", # See...
715
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCAmelCase : Any = { 'configuration_roc_bert': ['ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RoCBertConfig'], 'tokenization_roc_bert': ['RoCBertTok...
662
0
from __future__ import annotations def __magic_name__ ( A : List[Any] ): '''simple docstring''' create_state_space_tree(_lowerCamelCase, [], 0, [0 for i in range(len(_lowerCamelCase ) )] ) def __magic_name__ ( A : Optiona...
716
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_c...
662
0
import time import warnings from abc import ABC from copy import deepcopy from typing import Optional import torch from ..utils import add_start_docstrings, logging __lowerCAmelCase : List[Any] = logging.get_logger(__name__) __lowerCAmelCase : str = r'\n Args:\n input_ids (`...
717
from typing import TYPE_CHECKING from ....utils import _LazyModule __lowerCAmelCase : int = {'tokenization_tapex': ['TapexTokenizer']} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys __lowerCAmelCase : Tuple = _LazyModule(__name__, globals()['__fi...
662
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) __lowerCAmelCase : List[str] = { "naver-clova-ix/donut-base": "https://huggingface.co/naver-clova-ix/donut-base/resolve/main/config.json", ...
718
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) __lowerCAmelCase : Dict = { 'configuration_blip': [ 'BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BlipConfig...
662
0
__lowerCAmelCase : List[Any] = ''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' __lowerCAmelCase : ...
719
import math import flax.linen as nn import jax.numpy as jnp def __magic_name__ ( A : jnp.ndarray, A : int, A : float = 1, A : float = 1, A : float = 1.0E4, A : bool = False, A : float = 1.0, ): '''simple docstring''' ...
662
0
from typing import Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): from ..models.auto.modeling_auto ...
720
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils imp...
662
0
import argparse import os import re __lowerCAmelCase : Any = '''src/transformers''' # Pattern that looks at the indentation in a line. __lowerCAmelCase : Optional[Any] = re.compile(r'^(\s*)\S') # Pattern that matches `"key":" and puts `key` in group 0. __lowerCAmelCase : str = re.compil...
721
import argparse import os import re import numpy as np import PIL import torch from timm import create_model from torch.optim.lr_scheduler import OneCycleLR from torch.utils.data import DataLoader, Dataset from torchvision.transforms import Compose, RandomResizedCrop, Resize, ToTensor from accelerate import Acce...
662
0
import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex __lowerCAmelCase : List[str] = logging.getLogger(__name__) class snake_case__ : """simple docstring""" def _...
700
# this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since...
662
0
from __future__ import annotations import math import random from typing import Any class snake_case__ : """simple docstring""" def __init__( self : List[str] ) -> Union[str, Any]: a = [] a = 0 a = 0 def __Upp...
701
def __magic_name__ ( A : int, A : int, A : int ): '''simple docstring''' if exponent == 1: return base if exponent % 2 == 0: a = _modexpt(A, exponent // 2, A ) % modulo_value return (x * x) % modulo_value else...
662
0
'''simple docstring''' import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def __magic_name_...
702
def __magic_name__ ( A : str, A : str ): '''simple docstring''' def get_matched_characters(A : str, A : str ) -> str: a = [] a = min(len(_stra ), len(_stra ) ) // 2 for i, l in enumerate(_stra ...
662
0
import math def __magic_name__ ( A : int = 100 ): '''simple docstring''' a = sum(i * i for i in range(1, n + 1 ) ) a = int(math.pow(sum(range(1, n + 1 ) ), 2 ) ) return square_of_sum - sum_of_squares if __name...
703
__lowerCAmelCase : List[Any] = {str(digit): digit**5 for digit in range(10)} def __magic_name__ ( A : int ): '''simple docstring''' return sum(DIGITS_FIFTH_POWER[digit] for digit in str(A ) ) def __magic_name__ ( ): '''simple ...
662
0
import numpy class snake_case__ : """simple docstring""" def __init__( self : Optional[int] , __lowerCamelCase : numpy.ndarray , __lowerCamelCase : numpy.ndarray ) -> None: a = input_array # Random in...
704
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is...
662
0
'''simple docstring''' def __magic_name__ ( A : list[int] ): '''simple docstring''' if not numbers: return 0 if not isinstance(A, (list, tuple) ) or not all( isinstance(A, A ) for number in numbers ): raise ValueError("numbe...
705
def __magic_name__ ( A : list ): '''simple docstring''' for i in range(len(A ) - 1, 0, -1 ): a = False for j in range(A, 0, -1 ): if unsorted[j] < unsorted[j - 1]: a , a = unsorted[j - 1], u...
662
0
from __future__ import annotations def __magic_name__ ( A : list[int] ): # This function is recursive a = len(snake_case__ ) # If the array contains only one element, we return it (it's the stop condition of # recursion) if array_length <= 1: return array ...
706
from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo __lowerCAmelCase : Optional[Any] = '\\n@misc{wu2016googles,\n title={Google\'s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation},\n au...
662
0
'''simple docstring''' import torch from torch import nn class snake_case__ (nn.Module ): """simple docstring""" def __init__( self : Tuple , __lowerCamelCase : str , __lowerCamelCase : Dict , __lowerCamelCase : str , __lo...
707
import argparse import os import re __lowerCAmelCase : Union[str, Any] = 'src/transformers/models/auto' # re pattern that matches mapping introductions: # SUPER_MODEL_MAPPING_NAMES = OrderedDict or SUPER_MODEL_MAPPING = OrderedDict __lowerCAmelCase : Dict = re.compile(r'[A-Z_]+_MAPPING(\...
662
0
import argparse import logging import os import re import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, DataCollatorForLanguageModeling, PushToHubCallback, TFAutoModelForMaskedLM, create_optimizer, ) __lowerCAmelCase : Dict = logging.getLogger(__name__...
708
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __lowerCAmelCase : int = logging.get_logger(__name__) __lowerCAmelCase : Optional[int] =...
662
0
import qiskit def __magic_name__ ( A : int, A : int ): '''simple docstring''' a = qiskit.Aer.get_backend("aer_simulator" ) # Create a Quantum Circuit acting on the q register a = qiskit.QuantumCircuit(_lowercase, _lowercase ) ...
709
from __future__ import annotations import time import numpy as np __lowerCAmelCase : List[str] = [8, 5, 9, 7] __lowerCAmelCase : str = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] __lowerCAmelCase : Optional[Any] = [ [3, 2, 1, 4]...
662
0
from collections.abc import Iterable from typing import Any class snake_case__ : """simple docstring""" def __init__( self : Optional[int] , __lowerCamelCase : int | None = None ) -> Optional[Any]: a = value a = None # Added ...
710
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _in...
662
0
from copy import deepcopy class snake_case__ : """simple docstring""" def __init__( self : Union[str, Any] , __lowerCamelCase : Dict = None , __lowerCamelCase : List[str] = None ) -> Dict: if arr is None and size is not None: ...
711
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird impor...
662
0
import sys __lowerCAmelCase : Optional[Any] = ( '73167176531330624919225119674426574742355349194934' '96983520312774506326239578318016984801869478851843' '85861560789112949495459501737958331952853208805511' '12540698747158523863050715693290963295227443043557' '6689664895044524452316173185...
712
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer __lowerCAmelCase : List[Any] = logging.getLogger(__name__) def __magic_name__ ( ): '''simple docstring''' a = argparse.ArgumentParser( ...
662
0
import argparse import math import traceback import dateutil.parser as date_parser import requests def __magic_name__ ( A : Optional[int] ): '''simple docstring''' a = {} a = job["started_at"] a = job["completed_at"] a = date_pars...
713
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def __magic_name__ ( A : List[str] ): '''simple docstring''' a = {} a = tokeni...
662
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCAmelCase : int = { 'configuration_instructblip': [ 'INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'InstructBlipConfig', 'InstructBlipQFormerConfig', ...
714
import json import multiprocessing import os import re from collections import defaultdict import torch from accelerate import Accelerator from accelerate.utils import set_seed from arguments import HumanEvalArguments from datasets import load_dataset, load_metric from torch.utils.data import IterableDataset from...
662
0
'''simple docstring''' from math import pi def __magic_name__ ( A : List[str], A : Optional[Any] ): '''simple docstring''' return 2 * pi * radius * (angle / 360) if __name__ == "__main__": print(arc_length(90, 10))
715
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCAmelCase : Any = { 'configuration_roc_bert': ['ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RoCBertConfig'], 'tokenization_roc_bert': ['RoCBertTok...
662
0
from random import randint from tempfile import TemporaryFile import numpy as np def __magic_name__ ( A : List[Any], A : Optional[Any], A : Dict ): '''simple docstring''' a = 0 if start < end: a = randint(_lowerCAmelCase, ...
716
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_c...
662
0
import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast __lowerCAmelCase : List[str] = datasets.utils.logging.get_logger(__name__) @dataclass class snake_cas...
717
from typing import TYPE_CHECKING from ....utils import _LazyModule __lowerCAmelCase : int = {'tokenization_tapex': ['TapexTokenizer']} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys __lowerCAmelCase : Tuple = _LazyModule(__name__, globals()['__fi...
662
0
from __future__ import annotations def __magic_name__ ( A : Any ): # This function is recursive '''simple docstring''' a = len(__lowerCAmelCase ) # If the array contains only one element, we return it (it's the stop condition of # recursion) if arra...
718
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) __lowerCAmelCase : Dict = { 'configuration_blip': [ 'BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BlipConfig...
662
0
from ...processing_utils import ProcessorMixin class snake_case__ (__a ): """simple docstring""" SCREAMING_SNAKE_CASE_ : Union[str, Any] = ["image_processor", "feature_extractor"] SCREAMING_SNAKE_CASE_ : List[Any] = "TvltImageProcessor" ...
719
import math import flax.linen as nn import jax.numpy as jnp def __magic_name__ ( A : jnp.ndarray, A : int, A : float = 1, A : float = 1, A : float = 1.0E4, A : bool = False, A : float = 1.0, ): '''simple docstring''' ...
662
0
import bza import gzip import lzma import os import shutil import struct import tarfile import warnings import zipfile from abc import ABC, abstractmethod from pathlib import Path from typing import Dict, List, Optional, Type, Union from .. import config from .filelock import FileLock from .logging import get_log...
720
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils imp...
662
0
from typing import List import jiwer import jiwer.transforms as tr from packaging import version import datasets from datasets.config import PY_VERSION if PY_VERSION < version.parse('3.8'): import importlib_metadata else: import importlib.metadata as importlib_metadata __lowerCAmelCase : str = "" ...
721
import argparse import os import re import numpy as np import PIL import torch from timm import create_model from torch.optim.lr_scheduler import OneCycleLR from torch.utils.data import DataLoader, Dataset from torchvision.transforms import Compose, RandomResizedCrop, Resize, ToTensor from accelerate import Acce...
662
0
import inspect import os import unittest from pathlib import Path import torch import accelerate from accelerate.test_utils import execute_subprocess_async from accelerate.test_utils.testing import run_command class snake_case__ (unittest.TestCase ): """simple docstring""" SCREAMING...
700
# this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since...
662
0
import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .tokenization_wavaveca import WavaVecaCTCTokenizer class snake_case__ (_UpperCamelCase ): """simple docstring""" ...
701
def __magic_name__ ( A : int, A : int, A : int ): '''simple docstring''' if exponent == 1: return base if exponent % 2 == 0: a = _modexpt(A, exponent // 2, A ) % modulo_value return (x * x) % modulo_value else...
662
0
'''simple docstring''' import operator as op def __magic_name__ ( A : Any ): '''simple docstring''' a = [] a = lambda A, A : int(x / y ) # noqa: E731 integer division operation a = { "^": op.pow, "*": op.mul, ...
702
def __magic_name__ ( A : str, A : str ): '''simple docstring''' def get_matched_characters(A : str, A : str ) -> str: a = [] a = min(len(_stra ), len(_stra ) ) // 2 for i, l in enumerate(_stra ...
662
0
from __future__ import annotations from collections.abc import Callable from typing import Generic, TypeVar __magic_name__ : Any = TypeVar('T') __magic_name__ : str = TypeVar('U') class snake_case__ (Generic[T, U] ): """simple docstring""" def __init__( self ...
703
__lowerCAmelCase : List[Any] = {str(digit): digit**5 for digit in range(10)} def __magic_name__ ( A : int ): '''simple docstring''' return sum(DIGITS_FIFTH_POWER[digit] for digit in str(A ) ) def __magic_name__ ( ): '''simple ...
662
0
from copy import deepcopy class snake_case__ : """simple docstring""" def __init__( self : Optional[Any] , __lowerCamelCase : Any = None , __lowerCamelCase : str = None ) -> List[str]: if arr is None and size is not...
704
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is...
662
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase : Tuple = logging.get_logger(__name__) __lowerCAmelCase : Union[str, Any] = { 'sayakpaul/vit-msn-base': 'https://huggingface.co/sayakpaul/vit-msn-base/resolve/mai...
705
def __magic_name__ ( A : list ): '''simple docstring''' for i in range(len(A ) - 1, 0, -1 ): a = False for j in range(A, 0, -1 ): if unsorted[j] < unsorted[j - 1]: a , a = unsorted[j - 1], u...
662
0
import argparse import importlib from pathlib import Path # Test all the extensions added in the setup __lowerCAmelCase : Dict = [ """kernels/rwkv/wkv_cuda.cu""", """kernels/rwkv/wkv_op.cpp""", """kernels/deformable_detr/ms_deform_attn.h""", """kernels/deformable_detr/cuda/ms_deform_im2...
706
from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo __lowerCAmelCase : Optional[Any] = '\\n@misc{wu2016googles,\n title={Google\'s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation},\n au...
662
0
'''simple docstring''' import argparse import torch from datasets import load_dataset from donut import DonutModel from transformers import ( DonutImageProcessor, DonutProcessor, DonutSwinConfig, DonutSwinModel, MBartConfig, MBartForCausalLM, VisionEncoderDecoderModel, XLMRob...
707
import argparse import os import re __lowerCAmelCase : Union[str, Any] = 'src/transformers/models/auto' # re pattern that matches mapping introductions: # SUPER_MODEL_MAPPING_NAMES = OrderedDict or SUPER_MODEL_MAPPING = OrderedDict __lowerCAmelCase : Dict = re.compile(r'[A-Z_]+_MAPPING(\...
662
0
def __magic_name__ ( A : str ): '''simple docstring''' if isinstance(__UpperCamelCase, __UpperCamelCase ): raise TypeError("\'float\' object cannot be interpreted as an integer" ) if isinstance(__UpperCamelCase, __UpperCamelCase ): raise...
708
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __lowerCAmelCase : int = logging.get_logger(__name__) __lowerCAmelCase : Optional[int] =...
662
0
from sklearn.metrics import recall_score import datasets __lowerCAmelCase : Optional[Any] = """ Recall is the fraction of the positive examples that were correctly labeled by the model as positive. It can be computed with the equation: Recall = TP / (TP + FN) Where TP is the true positives and FN is the...
709
from __future__ import annotations import time import numpy as np __lowerCAmelCase : List[str] = [8, 5, 9, 7] __lowerCAmelCase : str = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] __lowerCAmelCase : Optional[Any] = [ [3, 2, 1, 4]...
662
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by...
710
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _in...
662
0
import math def __magic_name__ ( A : int ): '''simple docstring''' assert isinstance(__UpperCamelCase, __UpperCamelCase ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True...
711
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird impor...
662
0
from math import cos, sin, sqrt, tau from audio_filters.iir_filter import IIRFilter def __magic_name__ ( A : int, A : int, A : float = 1 / sqrt(2 ) ): '''simple docstring''' a = tau * frequency / samplerate a = sin(lowerCamelCase_ ...
712
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer __lowerCAmelCase : List[Any] = logging.getLogger(__name__) def __magic_name__ ( ): '''simple docstring''' a = argparse.ArgumentParser( ...
662
0
from dataclasses import dataclass from typing import Dict, Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .attention_processor import AttentionProcessor, AttnProcessor from .modeli...
713
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def __magic_name__ ( A : List[str] ): '''simple docstring''' a = {} a = tokeni...
662
0
import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def _UpperCamelCase ...
714
import json import multiprocessing import os import re from collections import defaultdict import torch from accelerate import Accelerator from accelerate.utils import set_seed from arguments import HumanEvalArguments from datasets import load_dataset, load_metric from torch.utils.data import IterableDataset from...
662
0
'''simple docstring''' from math import pow, sqrt def __magic_name__ ( *A : float ): '''simple docstring''' a = len(UpperCamelCase__ ) > 0 and all(value > 0.0 for value in values ) return result def __magic_name__ ( A : float...
715
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCAmelCase : Any = { 'configuration_roc_bert': ['ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RoCBertConfig'], 'tokenization_roc_bert': ['RoCBertTok...
662
0
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase : Optional[Any] = logging.get_logger(__name__) __lowerCAmelCase : Any = { '''microsoft/wavlm-base''': '''https://huggingface.co/microsoft/wavlm-base/resolve...
716
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_c...
662
0
from __future__ import annotations def __magic_name__ ( A : int ): '''simple docstring''' a = 2 a = [] while i * i <= n: if n % i: i += 1 else: n //= i factors.append(A ) if n > 1: factors.append(A ...
717
from typing import TYPE_CHECKING from ....utils import _LazyModule __lowerCAmelCase : int = {'tokenization_tapex': ['TapexTokenizer']} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys __lowerCAmelCase : Tuple = _LazyModule(__name__, globals()['__fi...
662
0
import argparse import torch from diffusers.pipelines.stable_diffusion.convert_from_ckpt import download_from_original_stable_diffusion_ckpt if __name__ == "__main__": __lowerCAmelCase : List[Any] = argparse.ArgumentParser() parser.add_argument( '--checkpoint_path', default=None, type=str,...
718
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) __lowerCAmelCase : Dict = { 'configuration_blip': [ 'BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BlipConfig...
662
0
import pytest from datasets.parallel import ParallelBackendConfig, parallel_backend from datasets.utils.py_utils import map_nested from .utils import require_dill_gt_0_3_2, require_joblibspark, require_not_windows def __magic_name__ ( A : Union[str, Any] ): # picklable for multiprocess...
719
import math import flax.linen as nn import jax.numpy as jnp def __magic_name__ ( A : jnp.ndarray, A : int, A : float = 1, A : float = 1, A : float = 1.0E4, A : bool = False, A : float = 1.0, ): '''simple docstring''' ...
662
0
from PIL import Image def __magic_name__ ( A : Image ): '''simple docstring''' a = image.size a = 0 a = image.load() for i in range(lowercase_ ): for j in range(lowercase_ ): a = pixels[j, i] mean += p...
720
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils imp...
662
0
from collections import namedtuple import requests from lxml import html # type: ignore __lowerCAmelCase : int = namedtuple('covid_data', 'cases deaths recovered') def __magic_name__ ( A : str = "https://www.worldometers.info/coronavirus/" ): '''simple docstring''' ...
721
import argparse import os import re import numpy as np import PIL import torch from timm import create_model from torch.optim.lr_scheduler import OneCycleLR from torch.utils.data import DataLoader, Dataset from torchvision.transforms import Compose, RandomResizedCrop, Resize, ToTensor from accelerate import Acce...
662
0
import warnings from ...utils import logging from .image_processing_beit import BeitImageProcessor __lowerCAmelCase : int = logging.get_logger(__name__) class snake_case__ (_UpperCamelCase ): """simple docstring""" def __init__( self : Tuple , *__lowerCame...
700
# this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since...
662
0
import json import os import shutil import tempfile import unittest from transformers import BatchEncoding, CanineTokenizer from transformers.testing_utils import require_tokenizers, require_torch from transformers.tokenization_utils import AddedToken from transformers.utils import cached_property from ...test_t...
701
def __magic_name__ ( A : int, A : int, A : int ): '''simple docstring''' if exponent == 1: return base if exponent % 2 == 0: a = _modexpt(A, exponent // 2, A ) % modulo_value return (x * x) % modulo_value else...
662
0
'''simple docstring''' def __magic_name__ ( A : str ): '''simple docstring''' return [ txt[:a] + txt[a].upper() + txt[a + 1 :] for a in range(len(SCREAMING_SNAKE_CASE_ ) ) if txt[a].isalpha() ] if __name__ == "__main__": __import__('...
702
def __magic_name__ ( A : str, A : str ): '''simple docstring''' def get_matched_characters(A : str, A : str ) -> str: a = [] a = min(len(_stra ), len(_stra ) ) // 2 for i, l in enumerate(_stra ...
662
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __magic_name__ : Dict = {'configuration_mbart': ['MBART_PRETRA...
703
__lowerCAmelCase : List[Any] = {str(digit): digit**5 for digit in range(10)} def __magic_name__ ( A : int ): '''simple docstring''' return sum(DIGITS_FIFTH_POWER[digit] for digit in str(A ) ) def __magic_name__ ( ): '''simple ...
662
0
from __future__ import annotations import pandas as pd def __magic_name__ ( A : list[int], A : list[int], A : int ): '''simple docstring''' a = [0] * no_of_processes a = [0] * no_of_processes # Copy t...
704
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is...
662
0
'''simple docstring''' from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable __lowerCAmelCase : Tuple = { 'configuration_gpt_neox_japanese': ['GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP', 'GPTNeoXJap...
705
def __magic_name__ ( A : list ): '''simple docstring''' for i in range(len(A ) - 1, 0, -1 ): a = False for j in range(A, 0, -1 ): if unsorted[j] < unsorted[j - 1]: a , a = unsorted[j - 1], u...
662
0
def __magic_name__ ( A : Optional[Any], A : int, A : Dict ): if len(lowerCAmelCase__ ) != len(lowerCAmelCase__ ): raise ValueError("The length of profit and weight must be same." ) if max_weight <= 0: raise ValueError("max_weight must greater ...
706
from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo __lowerCAmelCase : Optional[Any] = '\\n@misc{wu2016googles,\n title={Google\'s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation},\n au...
662
0
'''simple docstring''' def __magic_name__ ( A : int, A : Tuple ): '''simple docstring''' a = len(lowerCAmelCase_ ) a = [] for i in range(len(lowerCAmelCase_ ) - pat_len + 1 ): a = True for j in range(lowerCAm...
707
import argparse import os import re __lowerCAmelCase : Union[str, Any] = 'src/transformers/models/auto' # re pattern that matches mapping introductions: # SUPER_MODEL_MAPPING_NAMES = OrderedDict or SUPER_MODEL_MAPPING = OrderedDict __lowerCAmelCase : Dict = re.compile(r'[A-Z_]+_MAPPING(\...
662
0
from typing import Optional from urllib.parse import quote import huggingface_hub as hfh from packaging import version def __magic_name__ ( A : List[str], A : Any, A : Union[str, Any] = None ): '''simple docstring''' if version.parse(hfh.__version__...
708
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __lowerCAmelCase : int = logging.get_logger(__name__) __lowerCAmelCase : Optional[int] =...
662
0
from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_...
709
from __future__ import annotations import time import numpy as np __lowerCAmelCase : List[str] = [8, 5, 9, 7] __lowerCAmelCase : str = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] __lowerCAmelCase : Optional[Any] = [ [3, 2, 1, 4]...
662
0
from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar __lowerCAmelCase : List[str] = TypeVar('KEY') __lowerCAmelCase : str = TypeVar('VAL') @dataclass(frozen=_A , slots=_A ) class snake_case__ (Generic[...
710
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _in...
662
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCAmelCase : List[Any] = logging.get_logger(__name__) __lowerCAmelCase : Dict ...
711
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird impor...
662
0
from __future__ import annotations import unittest from transformers import BlenderbotConfig, BlenderbotTokenizer, is_tf_available from transformers.testing_utils import require_tf, require_tokenizers, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...t...
712
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer __lowerCAmelCase : List[Any] = logging.getLogger(__name__) def __magic_name__ ( ): '''simple docstring''' a = argparse.ArgumentParser( ...
662
0
from operator import delitem, getitem, setitem import pytest from data_structures.hashing.hash_map import HashMap def __magic_name__ ( A : Union[str, Any] ): '''simple docstring''' return getitem, k def __magic_name__ ( A : Any, A : Dic...
713
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def __magic_name__ ( A : List[str] ): '''simple docstring''' a = {} a = tokeni...
662
0
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple reposito...
714
import json import multiprocessing import os import re from collections import defaultdict import torch from accelerate import Accelerator from accelerate.utils import set_seed from arguments import HumanEvalArguments from datasets import load_dataset, load_metric from torch.utils.data import IterableDataset from...
662
0
'''simple docstring''' def __magic_name__ ( A : Union[str, Any], A : int ): '''simple docstring''' while second != 0: a = first & second first ^= second a = c << 1 return first if __name__ == "__main__": import doctest ...
715
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCAmelCase : Any = { 'configuration_roc_bert': ['ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RoCBertConfig'], 'tokenization_roc_bert': ['RoCBertTok...
662
0
import os from distutils.util import strtobool def __magic_name__ ( A : List[str], A : Union[str, Any] ): '''simple docstring''' for e in env_keys: a = int(os.environ.get(a__, -1 ) ) if val >= 0: return val return de...
716
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_c...
662
0
import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTest...
717
from typing import TYPE_CHECKING from ....utils import _LazyModule __lowerCAmelCase : int = {'tokenization_tapex': ['TapexTokenizer']} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys __lowerCAmelCase : Tuple = _LazyModule(__name__, globals()['__fi...
662
0
import torch from diffusers import EulerDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class snake_case__ (_UpperCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE_ : List[Any] = (EulerDiscreteSc...
718
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) __lowerCAmelCase : Dict = { 'configuration_blip': [ 'BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BlipConfig...
662
0
import faiss # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import requests # noqa: F401 # Here to have a nice missing dependency error message early on import sklearn # noqa: F401 # Here to ...
719
import math import flax.linen as nn import jax.numpy as jnp def __magic_name__ ( A : jnp.ndarray, A : int, A : float = 1, A : float = 1, A : float = 1.0E4, A : bool = False, A : float = 1.0, ): '''simple docstring''' ...
662
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __lowerCAmelCase : Dict = {'configuration_xglm': ['XGLM_PRETRA...
720
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils imp...
662
0
__lowerCAmelCase : Optional[int] = """ # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git """ __lowerCAmelCase : ...
721
import argparse import os import re import numpy as np import PIL import torch from timm import create_model from torch.optim.lr_scheduler import OneCycleLR from torch.utils.data import DataLoader, Dataset from torchvision.transforms import Compose, RandomResizedCrop, Resize, ToTensor from accelerate import Acce...
662
0
import torch from transformers import AutoModel class snake_case__ (torch.nn.Module ): """simple docstring""" def __init__( self : int , __lowerCamelCase : Tuple="sayef/fsner-bert-base-uncased" ) -> Dict: super(UpperCAmelCase__ , self ...
700
# this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since...
662
0
import importlib import inspect import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py __lowerCAmelCase : Tuple = 'src/transformers' # This is to make sure the transformers module imported...
701
def __magic_name__ ( A : int, A : int, A : int ): '''simple docstring''' if exponent == 1: return base if exponent % 2 == 0: a = _modexpt(A, exponent // 2, A ) % modulo_value return (x * x) % modulo_value else...
662
0
'''simple docstring''' import copy import tempfile import unittest from huggingface_hub import HfFolder, delete_repo from parameterized import parameterized from requests.exceptions import HTTPError from transformers import AutoConfig, GenerationConfig from transformers.testing_utils import TOKEN, USER, is_...
702
def __magic_name__ ( A : str, A : str ): '''simple docstring''' def get_matched_characters(A : str, A : str ) -> str: a = [] a = min(len(_stra ), len(_stra ) ) // 2 for i, l in enumerate(_stra ...
662
0
from __future__ import annotations def __magic_name__ ( A : List[Any], A : Dict, A : Dict, A : Dict ): # noqa: E741 '''simple docstring''' while r - l > 1: a = (l + r) // 2 if v[m] >= key: a = m els...
703
__lowerCAmelCase : List[Any] = {str(digit): digit**5 for digit in range(10)} def __magic_name__ ( A : int ): '''simple docstring''' return sum(DIGITS_FIFTH_POWER[digit] for digit in str(A ) ) def __magic_name__ ( ): '''simple ...
662
0
import logging import os from dataclasses import dataclass, field from typing import Dict, Optional import numpy as np from utils_multiple_choice import MultipleChoiceDataset, Split, processors import transformers from transformers import ( AutoConfig, AutoModelForMultipleChoice, Aut...
704
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is...
662
0
'''simple docstring''' import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transform...
705
def __magic_name__ ( A : list ): '''simple docstring''' for i in range(len(A ) - 1, 0, -1 ): a = False for j in range(A, 0, -1 ): if unsorted[j] < unsorted[j - 1]: a , a = unsorted[j - 1], u...
662
0
import argparse import re from pathlib import Path import requests import torch from PIL import Image from torchvision.transforms import CenterCrop, Compose, Normalize, Resize, ToTensor from transformers import ( EfficientFormerConfig, EfficientFormerForImageClassificationWithTeacher, EfficientFormer...
706
from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo __lowerCAmelCase : Optional[Any] = '\\n@misc{wu2016googles,\n title={Google\'s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation},\n au...
662
0
'''simple docstring''' import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import BatchEncoding, MarianTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import is_sentencepiece_available, is_t...
707
import argparse import os import re __lowerCAmelCase : Union[str, Any] = 'src/transformers/models/auto' # re pattern that matches mapping introductions: # SUPER_MODEL_MAPPING_NAMES = OrderedDict or SUPER_MODEL_MAPPING = OrderedDict __lowerCAmelCase : Dict = re.compile(r'[A-Z_]+_MAPPING(\...
662
0
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __lowerCAmelCase : List[str] = logging.get_logger(__name__) __lowerCAmelCase : Dict = { 'microsoft/wavlm-base': 'https://huggingface.co/microsoft/wavlm-base/resolve/main/con...
708
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __lowerCAmelCase : int = logging.get_logger(__name__) __lowerCAmelCase : Optional[int] =...
662
0
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface impo...
709
from __future__ import annotations import time import numpy as np __lowerCAmelCase : List[str] = [8, 5, 9, 7] __lowerCAmelCase : str = [ [2, 0, 1, 1], [0, 1, 2, 1], [4, 0, 0, 3], [0, 2, 1, 0], [1, 0, 3, 0], ] __lowerCAmelCase : Optional[Any] = [ [3, 2, 1, 4]...
662
0
import unittest import numpy as np import timeout_decorator # noqa from transformers import BlenderbotConfig, is_flax_available from transformers.testing_utils import jax_device, require_flax, slow from ...generation.test_flax_utils import FlaxGenerationTesterMixin from ...test_modeling_flax_common import Flax...
710
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _in...
662
0
import unittest from transformers.models.xlm_prophetnet.tokenization_xlm_prophetnet import SPIECE_UNDERLINE, XLMProphetNetTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTes...
711
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_big_bird impor...
662
0
def __magic_name__ ( A : Union[str, Any] ): '''simple docstring''' if not grid or not grid[0]: raise TypeError("The grid does not contain the appropriate information" ) for cell_n in range(1, len(grid[0] ) ): grid[0][cell_n] += grid[0][cell_n - 1...
712
import argparse import logging import os import datasets import tensorflow as tf from transformers import AutoTokenizer __lowerCAmelCase : List[Any] = logging.getLogger(__name__) def __magic_name__ ( ): '''simple docstring''' a = argparse.ArgumentParser( ...
662
0
def __magic_name__ ( A : List[Any] ): '''simple docstring''' stooge(a_, 0, len(a_ ) - 1 ) return arr def __magic_name__ ( A : Any, A : str, A : List[Any] ): '''simple docstring''' if i >= ...
713
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def __magic_name__ ( A : List[str] ): '''simple docstring''' a = {} a = tokeni...
662
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __lowerCAmelCase : Optional[int] = { 'configuration_rag': ['RagConfig'], 'retrieval_rag': ['RagRetriever'], 'tokenization_rag': ['RagTokenizer'], } try: ...
714
import json import multiprocessing import os import re from collections import defaultdict import torch from accelerate import Accelerator from accelerate.utils import set_seed from arguments import HumanEvalArguments from datasets import load_dataset, load_metric from torch.utils.data import IterableDataset from...
662
0
'''simple docstring''' from collections import Counter from pathlib import Path from typing import Optional, Tuple import yaml class snake_case__ (yaml.SafeLoader ): """simple docstring""" def __UpperCAmelCase ( self : Any , __lowerCamelCase : ...
715
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCAmelCase : Any = { 'configuration_roc_bert': ['ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RoCBertConfig'], 'tokenization_roc_bert': ['RoCBertTok...
662
0
import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_torch, slow from .test_pipeline...
716
import itertools import json import os import unittest from transformers import AddedToken, LongformerTokenizer, LongformerTokenizerFast from transformers.models.longformer.tokenization_longformer import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_c...
662
0
from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class snake_case__ (__UpperCAmelCase ): """simple d...
717
from typing import TYPE_CHECKING from ....utils import _LazyModule __lowerCAmelCase : int = {'tokenization_tapex': ['TapexTokenizer']} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys __lowerCAmelCase : Tuple = _LazyModule(__name__, globals()['__fi...
662
0
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
718
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) __lowerCAmelCase : Dict = { 'configuration_blip': [ 'BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BlipConfig...
662
0
def __magic_name__ ( A : int ): '''simple docstring''' if upper_limit < 0: raise ValueError("Limit for the Catalan sequence must be ≥ 0" ) a = [0] * (upper_limit + 1) # Base case: C(0) = C(1) = 1 a = 1 if upper_limit > 0: a = ...
719
import math import flax.linen as nn import jax.numpy as jnp def __magic_name__ ( A : jnp.ndarray, A : int, A : float = 1, A : float = 1, A : float = 1.0E4, A : bool = False, A : float = 1.0, ): '''simple docstring''' ...
662
0
from PIL import Image def __magic_name__ ( A : Tuple, A : List[str] ): '''simple docstring''' def brightness(A : Tuple ) -> float: return 128 + level + (c - 128) if not -255.0 <= level <= 255.0: raise ValueError("level must be between ...
720
import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision from transformers.utils imp...
662
0
from __future__ import annotations from collections.abc import Callable __lowerCAmelCase : Optional[Any] = list[list[float | int]] def __magic_name__ ( A : Optional[Any], A : int ): '''simple docstring''' a = len(lowerCamelCase__ ) a =...
721
import argparse import os import re import numpy as np import PIL import torch from timm import create_model from torch.optim.lr_scheduler import OneCycleLR from torch.utils.data import DataLoader, Dataset from torchvision.transforms import Compose, RandomResizedCrop, Resize, ToTensor from accelerate import Acce...
662
0
from __future__ import annotations def __magic_name__ ( A : list[int], A : int ): '''simple docstring''' if len(A ) == 0: return False a = len(A ) // 2 if a_list[midpoint] == item: return True if item < a_list[midpoint]: ...
700
# this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under git won't be considered # since...
662
0
import argparse from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import BigBirdPegasusConfig, BigBirdPegasusForConditionalGeneration __lowerCAmelCase : Any = [ # tf -> hf ('/', '.'), ('layer_', 'layers.'), ('kernel', 'weight'), (...
701
def __magic_name__ ( A : int, A : int, A : int ): '''simple docstring''' if exponent == 1: return base if exponent % 2 == 0: a = _modexpt(A, exponent // 2, A ) % modulo_value return (x * x) % modulo_value else...
662
0