code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' import copy import random from transformers import CLIPTokenizer class _UpperCAmelCase ( lowerCAmelCase_ ): def __init__( self,*__SCREAMING_SNAKE_CASE,**__SCREAMING_SNAKE_CASE ): '''simple docstring''' super().__init__(*__SCREAMING_SNAKE_CASE,...
689
'''simple docstring''' import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGe...
689
1
'''simple docstring''' import warnings from transformers import AutoTokenizer from transformers.utils import is_torch_available from transformers.utils.generic import ExplicitEnum from ...processing_utils import ProcessorMixin if is_torch_available(): import torch class _UpperCAmelCase ...
689
'''simple docstring''' from collections import deque def _lowerCAmelCase ( lowercase ) -> Dict: __lowerCAmelCase = len(lowercase ) __lowerCAmelCase = deque() __lowerCAmelCase = [False for _ in range(lowercase )] __lowerCAmelCase = ...
689
1
'''simple docstring''' from __future__ import annotations import os from typing import Any import requests _a : Tuple = """https://api.github.com""" # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user _a : Optional[Any] ...
689
'''simple docstring''' from argparse import ArgumentParser from .env import EnvironmentCommand def _lowerCAmelCase ( ) -> Union[str, Any]: __lowerCAmelCase = ArgumentParser("""Diffusers CLI tool""" , usage="""diffusers-cli <command> [<args>]""" ) __lowerCAmelCase ...
689
1
'''simple docstring''' from __future__ import annotations from bisect import bisect_left from functools import total_ordering from heapq import merge @total_ordering class _UpperCAmelCase ( lowerCAmelCase_ ): def __lt__( self,__SCREAMING_SNAKE_CASE ): '''simple docstring'...
689
'''simple docstring''' import argparse import fairseq import torch from transformers import UniSpeechSatConfig, UniSpeechSatForCTC, UniSpeechSatForPreTraining, logging logging.set_verbosity_info() _a : List[Any] = logging.get_logger(__name__) _a : int = { ...
689
1
'''simple docstring''' from __future__ import annotations _a : Union[str, Any] = """#""" class _UpperCAmelCase : def __init__( self ): '''simple docstring''' __lowerCAmelCase = {} def lowerCamelCase__ ( self,__SCREAMING_SNA...
689
'''simple docstring''' from scipy.stats import spearmanr import datasets _a : str = """ The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no c...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase , lowercase ) -> Optional[int]: __lowerCAmelCase = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def _lowerCAmelCase ( lowercase , lowercase , lowe...
689
'''simple docstring''' from ..utils import DummyObject, requires_backends class _UpperCAmelCase ( metaclass=lowerCAmelCase_ ): a : List[str] =["""onnx"""] def __init__( self,*__SCREAMING_SNAKE_CASE,**__SCREAMING_SNAKE_CASE ): '''simple docstring''' requires_...
689
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _a : Tuple = { """configuration_roformer""": ["...
689
'''simple docstring''' from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import loggi...
689
1
'''simple docstring''' import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor _a : int = logging.get_logger(__name__) class _UpperCAmelCase ( lowerCAmelCase_ ): def __init__( self,*__SCREAMING_SNAKE_CASE,...
689
'''simple docstring''' def _lowerCAmelCase ( lowercase = 5000_0000 ) -> int: __lowerCAmelCase = set() __lowerCAmelCase = int((limit - 24) ** (1 / 2) ) __lowerCAmelCase = set(range(3 , prime_square_limit + 1 , 2 ) ) primes....
689
1
'''simple docstring''' import os import sys import unittest _a : Optional[Any] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import check_dummies # noqa: E402 from check_dummies import creat...
689
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers...
689
1
'''simple docstring''' import sys _a : Any = ( """73167176531330624919225119674426574742355349194934""" """96983520312774506326239578318016984801869478851843""" """85861560789112949495459501737958331952853208805511""" """12540698747158523863050715693290963295227443043...
689
'''simple docstring''' from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def _lowerCAmelCase ( lowercase ) -> Optional[int]: if not is_accelerate_available(): return method __lowerCAmelCase =...
689
1
'''simple docstring''' import re from filelock import FileLock try: import nltk _a : Optional[Any] = True except (ImportError, ModuleNotFoundError): _a : Optional[int] = False if NLTK_AVAILABLE: with FileLock(""".lock""") as lock: nltk.download("""punk...
689
'''simple docstring''' import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def _lowerCAmelCase ( lowercase , lowercase , lowercase , lowercase , lowercase ) -> Optional[int]: # load base model ...
689
1
'''simple docstring''' import unittest from transformers import DebertaVaTokenizer, DebertaVaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _a : List[Any] = ...
689
'''simple docstring''' from collections import Counter from timeit import timeit def _lowerCAmelCase ( lowercase = "" , ) -> bool: return sum(c % 2 for c in Counter(input_str.replace(""" """ , """""" ).lower() ).values() ) < 2 def _lowerCAmelCase ( lowe...
689
1
'''simple docstring''' import numpy as np def _lowerCAmelCase ( lowercase ) -> np.array: return (2 / (1 + np.exp(-2 * vector ))) - 1 if __name__ == "__main__": import doctest doctest.testmod()
689
'''simple docstring''' import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def _lowerCAmelCa...
689
1
'''simple docstring''' import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def _lowerCAmelCase ( lowercase , lowercase , lowercase , lowercase , lowercase ) -> Optional[int]: # load base model ...
689
'''simple docstring''' import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging _a : Tuple = """\ """ _a : Tuple = """ Perplexity (PPL) is ...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase ) -> float: if edge <= 0 or not isinstance(lowercase , lowercase ): raise ValueError("""Length must be a positive.""" ) return 3 * ((25 + 10 * (5 ** (1 / 2))) ** (1 / 2)) * (edge**2) def _lowerCAmelCase...
689
'''simple docstring''' from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): impor...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase = 1000 ) -> int: __lowerCAmelCase , __lowerCAmelCase = 1, 1 __lowerCAmelCase = [] for i in range(1 , n + 1 ): __lowerCAmelCase = prev_numerator + 2 * prev_denominator __lowe...
689
'''simple docstring''' import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartFor...
689
1
'''simple docstring''' # this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under ...
689
'''simple docstring''' import os import sys import unittest _a : List[Any] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import check_dummies # noqa: E402 from check_dummies import create_du...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase , lowercase ) -> list[int]: __lowerCAmelCase = int(lowercase ) # Initialize Result __lowerCAmelCase = [] # Traverse through all denomination for denomination in reversed(lowercase ): ...
689
'''simple docstring''' def _lowerCAmelCase ( lowercase ) -> tuple[int, int]: try: __lowerCAmelCase = float(lowercase ) except ValueError: raise ValueError("""Please enter a valid number""" ) __lowerCAmelCase = decimal - int(lowercase ) ...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase , lowercase ) -> str: if a < 0 or b < 0: raise ValueError("""the value of both inputs must be positive""" ) __lowerCAmelCase = str(bin(lowercase ) )[2:] # remove the leading "0b" __lowerCAme...
689
'''simple docstring''' from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool from google.protobuf import symbol_database as _symbol_database from google.protobuf.internal import builder as _builder # @@protoc_insertion_point(imports) _a :...
689
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available _a : Union[str, Any] = { """configuration_gpt_neo""": ["""GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTNeoConfig"""...
689
'''simple docstring''' from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block...
689
1
'''simple docstring''' import argparse import logging import os import re import tensorflow as tf from transformers import ( AutoConfig, AutoTokenizer, DataCollatorForLanguageModeling, PushToHubCallback, TFAutoModelForMaskedLM, create_optimizer, ) _a : Optional[Any] ...
689
'''simple docstring''' import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_...
689
1
'''simple docstring''' # Lint as: python3 import sys from collections.abc import Mapping from typing import TYPE_CHECKING import numpy as np import pyarrow as pa from .. import config from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import torch cl...
689
'''simple docstring''' import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def _lowerCAmelCase ( lowercase ) -> Optional[Any]: # vision encoder if "img_encoder.pos_embed" in name: __lowerCAm...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase ) -> int: __lowerCAmelCase = len(lowercase ) __lowerCAmelCase = len(matrix[0] ) __lowerCAmelCase = min(lowercase , lowercase ) for row in range(lowercase ): # Chec...
689
'''simple docstring''' import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGe...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase , lowercase , lowercase ) -> int: def count_of_possible_combinations(lowercase ) -> int: if target < 0: return 0 if target == 0: return 1 return sum(count_of_possible_combinations(...
689
'''simple docstring''' from collections import deque def _lowerCAmelCase ( lowercase ) -> Dict: __lowerCAmelCase = len(lowercase ) __lowerCAmelCase = deque() __lowerCAmelCase = [False for _ in range(lowercase )] __lowerCAmelCase = ...
689
1
'''simple docstring''' from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import loggi...
689
'''simple docstring''' from argparse import ArgumentParser from .env import EnvironmentCommand def _lowerCAmelCase ( ) -> Union[str, Any]: __lowerCAmelCase = ArgumentParser("""Diffusers CLI tool""" , usage="""diffusers-cli <command> [<args>]""" ) __lowerCAmelCase ...
689
1
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers...
689
'''simple docstring''' import argparse import fairseq import torch from transformers import UniSpeechSatConfig, UniSpeechSatForCTC, UniSpeechSatForPreTraining, logging logging.set_verbosity_info() _a : List[Any] = logging.get_logger(__name__) _a : int = { ...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase = 100_0000 ) -> int: __lowerCAmelCase = set(range(3 , lowercase , 2 ) ) primes.add(2 ) for p in range(3 , lowercase , 2 ): if p not in primes: continue ...
689
'''simple docstring''' from scipy.stats import spearmanr import datasets _a : str = """ The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no c...
689
1
'''simple docstring''' import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( """kwargs, expected""" , [ ({"""num_shards""": 0, """max_num_jobs""": 1}, []), ({"""num_shards""": 10, "...
689
'''simple docstring''' from ..utils import DummyObject, requires_backends class _UpperCAmelCase ( metaclass=lowerCAmelCase_ ): a : List[str] =["""onnx"""] def __init__( self,*__SCREAMING_SNAKE_CASE,**__SCREAMING_SNAKE_CASE ): '''simple docstring''' requires_...
689
1
'''simple docstring''' import torch from torch import nn from transformers import CLIPPreTrainedModel, CLIPVisionModel from ...models.attention import BasicTransformerBlock from ...utils import logging _a : List[str] = logging.get_logger(__name__) # pylint: disable=invalid-name ...
689
'''simple docstring''' from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import loggi...
689
1
'''simple docstring''' from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): impor...
689
'''simple docstring''' def _lowerCAmelCase ( lowercase = 5000_0000 ) -> int: __lowerCAmelCase = set() __lowerCAmelCase = int((limit - 24) ** (1 / 2) ) __lowerCAmelCase = set(range(3 , prime_square_limit + 1 , 2 ) ) primes....
689
1
'''simple docstring''' from __future__ import annotations import math def _lowerCAmelCase ( lowercase , lowercase ) -> float: __lowerCAmelCase = u for i in range(1 , lowercase ): __lowerCAmelCase = temp * (u - i) return temp def ...
689
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers...
689
1
'''simple docstring''' import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate impo...
689
'''simple docstring''' from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def _lowerCAmelCase ( lowercase ) -> Optional[int]: if not is_accelerate_available(): return method __lowerCAmelCase =...
689
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, ...
689
'''simple docstring''' import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def _lowerCAmelCase ( lowercase , lowercase , lowercase , lowercase , lowercase ) -> Optional[int]: # load base model ...
689
1
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torc...
689
'''simple docstring''' from collections import Counter from timeit import timeit def _lowerCAmelCase ( lowercase = "" , ) -> bool: return sum(c % 2 for c in Counter(input_str.replace(""" """ , """""" ).lower() ).values() ) < 2 def _lowerCAmelCase ( lowe...
689
1
'''simple docstring''' import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers...
689
'''simple docstring''' import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def _lowerCAmelCa...
689
1
'''simple docstring''' import argparse import os import re import tensorflow as tf import torch from transformers import BertConfig, BertModel from transformers.utils import logging logging.set_verbosity_info() _a : Optional[int] = logging.get_logger(__name__) def _lowerCAm...
689
'''simple docstring''' import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging _a : Tuple = """\ """ _a : Tuple = """ Perplexity (PPL) is ...
689
1
'''simple docstring''' import gc import math import unittest import torch from diffusers import UNetaDModel from diffusers.utils import floats_tensor, logging, slow, torch_all_close, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTes...
689
'''simple docstring''' from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): impor...
689
1
'''simple docstring''' from collections import deque from .hash_table import HashTable class _UpperCAmelCase ( lowerCAmelCase_ ): def __init__( self,*__SCREAMING_SNAKE_CASE,**__SCREAMING_SNAKE_CASE ): '''simple docstring''' super().__init__(*__SCREAMING_SNAKE_CASE,*...
689
'''simple docstring''' import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartFor...
689
1
'''simple docstring''' from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def _lowerCAmelCase ( lowercase ) -> Optional[int]: if not is_accelerate_available(): return method __lowerCAmelCase =...
689
'''simple docstring''' import os import sys import unittest _a : List[Any] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import check_dummies # noqa: E402 from check_dummies import create_du...
689
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _a : List[Any] = { """configuration_gpt_bigcode""": ["""GPT_BIGCODE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTBigCodeConfi...
689
'''simple docstring''' def _lowerCAmelCase ( lowercase ) -> tuple[int, int]: try: __lowerCAmelCase = float(lowercase ) except ValueError: raise ValueError("""Please enter a valid number""" ) __lowerCAmelCase = decimal - int(lowercase ) ...
689
1
'''simple docstring''' from typing import Optional import pyspark from .. import Features, NamedSplit from ..download import DownloadMode from ..packaged_modules.spark.spark import Spark from .abc import AbstractDatasetReader class _UpperCAmelCase ( lowerCAmelCase_ ): def __init__( ...
689
'''simple docstring''' from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool from google.protobuf import symbol_database as _symbol_database from google.protobuf.internal import builder as _builder # @@protoc_insertion_point(imports) _a :...
689
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _a : Optional[Any] = ...
689
'''simple docstring''' from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block...
689
1
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a : str = logging.get_logger(__name__) _a : Optio...
689
'''simple docstring''' import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_...
689
1
'''simple docstring''' import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration...
689
'''simple docstring''' import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def _lowerCAmelCase ( lowercase ) -> Optional[Any]: # vision encoder if "img_encoder.pos_embed" in name: __lowerCAm...
689
1
'''simple docstring''' # this script reports modified .py files under the desired list of top-level sub-dirs passed as a list of arguments, e.g.: # python ./utils/get_modified_files.py utils src tests examples # # it uses git to find the forking point and which files were modified - i.e. files not under ...
689
'''simple docstring''' import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGe...
689
1
'''simple docstring''' from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block...
689
'''simple docstring''' from collections import deque def _lowerCAmelCase ( lowercase ) -> Dict: __lowerCAmelCase = len(lowercase ) __lowerCAmelCase = deque() __lowerCAmelCase = [False for _ in range(lowercase )] __lowerCAmelCase = ...
689
1
'''simple docstring''' from abc import ABC, abstractmethod from typing import List, Optional class _UpperCAmelCase ( lowerCAmelCase_ ): def __init__( self ): '''simple docstring''' self.test() def lowerCamelCase__ ( self ): '''simple docstrin...
689
'''simple docstring''' from argparse import ArgumentParser from .env import EnvironmentCommand def _lowerCAmelCase ( ) -> Union[str, Any]: __lowerCAmelCase = ArgumentParser("""Diffusers CLI tool""" , usage="""diffusers-cli <command> [<args>]""" ) __lowerCAmelCase ...
689
1
'''simple docstring''' import argparse import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_dummies.py _a : Dict = """src/diffusers""" # Matches is_xxx_available() _a : List[Any] ...
689
'''simple docstring''' import argparse import fairseq import torch from transformers import UniSpeechSatConfig, UniSpeechSatForCTC, UniSpeechSatForPreTraining, logging logging.set_verbosity_info() _a : List[Any] = logging.get_logger(__name__) _a : int = { ...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase = 100 ) -> int: __lowerCAmelCase = set() __lowerCAmelCase = 0 __lowerCAmelCase = n + 1 # maximum limit for a in range(2 , lowercase ): for b in range(2 , lowercase ...
689
'''simple docstring''' from scipy.stats import spearmanr import datasets _a : str = """ The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no c...
689
1
'''simple docstring''' import os from typing import Dict, List, Tuple, TypeVar, Union _a : Tuple = TypeVar("""T""") _a : Tuple = Union[List[T], Tuple[T, ...]] _a : List[Any] = Union[T, List[T], Dict[str, T]] _a : Tuple = ...
689
'''simple docstring''' from ..utils import DummyObject, requires_backends class _UpperCAmelCase ( metaclass=lowerCAmelCase_ ): a : List[str] =["""onnx"""] def __init__( self,*__SCREAMING_SNAKE_CASE,**__SCREAMING_SNAKE_CASE ): '''simple docstring''' requires_...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase ) -> list: __lowerCAmelCase = len(lowercase ) for _ in range(lowercase ): for i in range(_ % 2 , arr_size - 1 , 2 ): if arr[i + 1] < arr[i]: __lowerCAmelCase , __lo...
689
'''simple docstring''' from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import loggi...
689
1
'''simple docstring''' from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_st...
689
'''simple docstring''' def _lowerCAmelCase ( lowercase = 5000_0000 ) -> int: __lowerCAmelCase = set() __lowerCAmelCase = int((limit - 24) ** (1 / 2) ) __lowerCAmelCase = set(range(3 , prime_square_limit + 1 , 2 ) ) primes....
689
1
'''simple docstring''' import warnings from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch from ...models import UNetaDModel from ...schedulers import RePaintScheduler from ...utils import PIL_INTERPOLATION, logging, randn_tensor from ..pipeline_utils import Diffusi...
689
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers...
689
1
'''simple docstring''' import argparse import shutil import time from json import JSONDecodeError from logging import getLogger from pathlib import Path from typing import Dict, List import torch from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoModelForSeqaSeqLM,...
689
'''simple docstring''' from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def _lowerCAmelCase ( lowercase ) -> Optional[int]: if not is_accelerate_available(): return method __lowerCAmelCase =...
689
1
'''simple docstring''' from collections import Counter from timeit import timeit def _lowerCAmelCase ( lowercase = "" , ) -> bool: return sum(c % 2 for c in Counter(input_str.replace(""" """ , """""" ).lower() ).values() ) < 2 def _lowerCAmelCase ( lowe...
689
'''simple docstring''' import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def _lowerCAmelCase ( lowercase , lowercase , lowercase , lowercase , lowercase ) -> Optional[int]: # load base model ...
689
1
'''simple docstring''' import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from ...
689
'''simple docstring''' from collections import Counter from timeit import timeit def _lowerCAmelCase ( lowercase = "" , ) -> bool: return sum(c % 2 for c in Counter(input_str.replace(""" """ , """""" ).lower() ).values() ) < 2 def _lowerCAmelCase ( lowe...
689
1
'''simple docstring''' import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class _UpperCAmelCase ( lowerCAmelCase_ ,...
689
'''simple docstring''' import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def _lowerCAmelCa...
689
1
'''simple docstring''' from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_acc...
689
'''simple docstring''' import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging _a : Tuple = """\ """ _a : Tuple = """ Perplexity (PPL) is ...
689
1
'''simple docstring''' from math import factorial def _lowerCAmelCase ( lowercase , lowercase , lowercase ) -> float: if successes > trials: raise ValueError("""successes must be lower or equal to trials""" ) if trials < 0 or successes < 0: raise Value...
689
'''simple docstring''' from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): impor...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase ) -> list: __lowerCAmelCase = int(lowercase ) if n_element < 1: __lowerCAmelCase = ValueError("""a should be a positive number""" ) raise my_error __lowerCAmelCase = [1] __lo...
689
'''simple docstring''' import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartFor...
689
1
'''simple docstring''' _a : Optional[Any] = """Alexander Joslin""" import operator as op from .stack import Stack def _lowerCAmelCase ( lowercase ) -> int: __lowerCAmelCase = {"""*""": op.mul, """/""": op.truediv, """+""": op.add, """-""": op.sub} ...
689
'''simple docstring''' import os import sys import unittest _a : List[Any] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import check_dummies # noqa: E402 from check_dummies import create_du...
689
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _a : Tuple = logging.get_logger(__name__) _a : Tuple = {...
689
'''simple docstring''' def _lowerCAmelCase ( lowercase ) -> tuple[int, int]: try: __lowerCAmelCase = float(lowercase ) except ValueError: raise ValueError("""Please enter a valid number""" ) __lowerCAmelCase = decimal - int(lowercase ) ...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase , lowercase ) -> Optional[Any]: __lowerCAmelCase = 0 __lowerCAmelCase = len(lowercase ) - 1 while left <= right: # avoid divided by 0 during interpolation if sorted_collection[left] == ...
689
'''simple docstring''' from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool from google.protobuf import symbol_database as _symbol_database from google.protobuf.internal import builder as _builder # @@protoc_insertion_point(imports) _a :...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase ) -> int: __lowerCAmelCase = [[0 for _ in range(lowercase )] for _ in range(m + 1 )] for i in range(m + 1 ): __lowerCAmelCase = 1 for n in range(m + 1 ): for k in range(1 , ...
689
'''simple docstring''' from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block...
689
1
'''simple docstring''' import logging import os from typing import Dict, List, Optional, Union import torch import torch.nn as nn from accelerate.utils.imports import ( is_abit_bnb_available, is_abit_bnb_available, is_bnb_available, ) from ..big_modeling import dispatch_model, init_empty_wei...
689
'''simple docstring''' import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_...
689
1
'''simple docstring''' _a : List[Any] = [ """Audio""", """Array2D""", """Array3D""", """Array4D""", """Array5D""", """ClassLabel""", """Features""", """Sequence""", """Value""", """Image""", """Translation""", """TranslationVariableLangu...
689
'''simple docstring''' import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def _lowerCAmelCase ( lowercase ) -> Optional[Any]: # vision encoder if "img_encoder.pos_embed" in name: __lowerCAm...
689
1
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LIC...
689
'''simple docstring''' import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGe...
689
1
'''simple docstring''' from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common impo...
689
'''simple docstring''' from collections import deque def _lowerCAmelCase ( lowercase ) -> Dict: __lowerCAmelCase = len(lowercase ) __lowerCAmelCase = deque() __lowerCAmelCase = [False for _ in range(lowercase )] __lowerCAmelCase = ...
689
1
'''simple docstring''' import json import os import torch from diffusers import UNetaDModel os.makedirs("""hub/hopper-medium-v2/unet/hor32""", exist_ok=True) os.makedirs("""hub/hopper-medium-v2/unet/hor128""", exist_ok=True) os.makedirs("""hub/hopper-medium-v2/value_function""", exist_ok=True) de...
689
'''simple docstring''' from argparse import ArgumentParser from .env import EnvironmentCommand def _lowerCAmelCase ( ) -> Union[str, Any]: __lowerCAmelCase = ArgumentParser("""Diffusers CLI tool""" , usage="""diffusers-cli <command> [<args>]""" ) __lowerCAmelCase ...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase ) -> bool: if not isinstance(lowercase , lowercase ): raise ValueError("""check_bouncy() accepts only integer arguments""" ) __lowerCAmelCase = str(lowercase ) __lowerCAmelCase = """...
689
'''simple docstring''' import argparse import fairseq import torch from transformers import UniSpeechSatConfig, UniSpeechSatForCTC, UniSpeechSatForPreTraining, logging logging.set_verbosity_info() _a : List[Any] = logging.get_logger(__name__) _a : int = { ...
689
1
'''simple docstring''' import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated _a : Dict = collections.named...
689
'''simple docstring''' from scipy.stats import spearmanr import datasets _a : str = """ The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no c...
689
1
'''simple docstring''' import json import os import unittest from transformers.models.ctrl.tokenization_ctrl import VOCAB_FILES_NAMES, CTRLTokenizer from ...test_tokenization_common import TokenizerTesterMixin class _UpperCAmelCase ( lowerCAmelCase_ , unittest.TestCase ): a : ...
689
'''simple docstring''' from ..utils import DummyObject, requires_backends class _UpperCAmelCase ( metaclass=lowerCAmelCase_ ): a : List[str] =["""onnx"""] def __init__( self,*__SCREAMING_SNAKE_CASE,**__SCREAMING_SNAKE_CASE ): '''simple docstring''' requires_...
689
1
'''simple docstring''' from ..utils import DummyObject, requires_backends class _UpperCAmelCase ( metaclass=lowerCAmelCase_ ): a : List[str] =["""onnx"""] def __init__( self,*__SCREAMING_SNAKE_CASE,**__SCREAMING_SNAKE_CASE ): '''simple docstring''' requires_...
689
'''simple docstring''' from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import loggi...
689
1
'''simple docstring''' import json import os import sys import tempfile import unittest from pathlib import Path from shutil import copyfile from huggingface_hub import HfFolder, Repository, create_repo, delete_repo from requests.exceptions import HTTPError import transformers from transformers import ( ...
689
'''simple docstring''' def _lowerCAmelCase ( lowercase = 5000_0000 ) -> int: __lowerCAmelCase = set() __lowerCAmelCase = int((limit - 24) ** (1 / 2) ) __lowerCAmelCase = set(range(3 , prime_square_limit + 1 , 2 ) ) primes....
689
1
'''simple docstring''' import os import re import shutil import sys import tempfile import unittest import black _a : str = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import check_copies #...
689
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers...
689
1
'''simple docstring''' import math from typing import Optional import numpy as np from ...configuration_utils import PretrainedConfig from ...utils import logging _a : str = logging.get_logger(__name__) _a : List[Any] = { """facebook/encodec_24khz""": ""...
689
'''simple docstring''' from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def _lowerCAmelCase ( lowercase ) -> Optional[int]: if not is_accelerate_available(): return method __lowerCAmelCase =...
689
1
'''simple docstring''' from cva import destroyAllWindows, imread, imshow, waitKey def _lowerCAmelCase ( lowercase ) -> Tuple: # getting number of pixels in the image __lowerCAmelCase , __lowerCAmelCase = img.shape[0], img.shape[1] # converting each pixel's color to ...
689
'''simple docstring''' import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def _lowerCAmelCase ( lowercase , lowercase , lowercase , lowercase , lowercase ) -> Optional[int]: # load base model ...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase , lowercase ) -> int: return int((input_a, input_a).count(0 ) == 0 ) def _lowerCAmelCase ( ) -> None: assert and_gate(0 , 0 ) == 0 assert and_gate(0 , 1 ) == 0 assert and_ga...
689
'''simple docstring''' from collections import Counter from timeit import timeit def _lowerCAmelCase ( lowercase = "" , ) -> bool: return sum(c % 2 for c in Counter(input_str.replace(""" """ , """""" ).lower() ).values() ) < 2 def _lowerCAmelCase ( lowe...
689
1
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescal...
689
'''simple docstring''' import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def _lowerCAmelCa...
689
1
'''simple docstring''' import datasets _a : Tuple = """\ @InProceedings{conneau2018xnli, author = \"Conneau, Alexis and Rinott, Ruty and Lample, Guillaume and Williams, Adina and Bowman, Samuel R. ...
689
'''simple docstring''' import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging _a : Tuple = """\ """ _a : Tuple = """ Perplexity (PPL) is ...
689
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, ...
689
'''simple docstring''' from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): impor...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase , lowercase ) -> str: if not isinstance(lowercase , lowercase ): raise ValueError("""iterations must be defined as integers""" ) if not isinstance(lowercase , lowercase ) or not number >= 1: ...
689
'''simple docstring''' import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartFor...
689
1
'''simple docstring''' import inspect import unittest from transformers import RegNetConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_comm...
689
'''simple docstring''' import os import sys import unittest _a : List[Any] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, """utils""")) import check_dummies # noqa: E402 from check_dummies import create_du...
689
1
'''simple docstring''' import unittest from parameterized import parameterized from transformers import OpenLlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_c...
689
'''simple docstring''' def _lowerCAmelCase ( lowercase ) -> tuple[int, int]: try: __lowerCAmelCase = float(lowercase ) except ValueError: raise ValueError("""Please enter a valid number""" ) __lowerCAmelCase = decimal - int(lowercase ) ...
689
1
'''simple docstring''' import webbrowser from sys import argv from urllib.parse import parse_qs, quote import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": _a : Dict = """%20""".join(argv[1:]) if len(argv) > 1 else quote(str(...
689
'''simple docstring''' from google.protobuf import descriptor as _descriptor from google.protobuf import descriptor_pool as _descriptor_pool from google.protobuf import symbol_database as _symbol_database from google.protobuf.internal import builder as _builder # @@protoc_insertion_point(imports) _a :...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase ) -> tuple[int, int]: try: __lowerCAmelCase = float(lowercase ) except ValueError: raise ValueError("""Please enter a valid number""" ) __lowerCAmelCase = decimal - int(lowercase ) ...
689
'''simple docstring''' from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block...
689
1
'''simple docstring''' import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floa...
689
'''simple docstring''' import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_...
689
1
'''simple docstring''' def _lowerCAmelCase ( lowercase ) -> Optional[Any]: __lowerCAmelCase = [] __lowerCAmelCase = set({"""(""", """[""", """{"""} ) __lowerCAmelCase = set({""")""", """]""", """}"""} ) __lowerCAmelCase = {"""{""": ...
689
'''simple docstring''' import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def _lowerCAmelCase ( lowercase ) -> Optional[Any]: # vision encoder if "img_encoder.pos_embed" in name: __lowerCAm...
689
1
'''simple docstring''' import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatch...
689
'''simple docstring''' import argparse from pathlib import Path from typing import Dict, OrderedDict, Tuple import torch from audiocraft.models import MusicGen from transformers import ( AutoFeatureExtractor, AutoTokenizer, EncodecModel, MusicgenDecoderConfig, MusicgenForConditionalGe...
689
1
'''simple docstring''' from string import ascii_lowercase, ascii_uppercase def _lowerCAmelCase ( lowercase ) -> str: if not sentence: return "" __lowerCAmelCase = dict(zip(lowercase , lowercase ) ) return lower_to_upper.get(sentence[0] , se...
689
'''simple docstring''' from collections import deque def _lowerCAmelCase ( lowercase ) -> Dict: __lowerCAmelCase = len(lowercase ) __lowerCAmelCase = deque() __lowerCAmelCase = [False for _ in range(lowercase )] __lowerCAmelCase = ...
689
1
'''simple docstring''' import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin from .feature_extraction_wavaveca import WavaVecaFeatureExtractor from .tokenization_wavaveca import WavaVecaCTCTokenizer class _UpperCAmelCase ( lowerCAmelCase_ ): ...
689
'''simple docstring''' from argparse import ArgumentParser from .env import EnvironmentCommand def _lowerCAmelCase ( ) -> Union[str, Any]: __lowerCAmelCase = ArgumentParser("""Diffusers CLI tool""" , usage="""diffusers-cli <command> [<args>]""" ) __lowerCAmelCase ...
689
1
'''simple docstring''' import json import os from typing import Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a : Union[str, Any] = logging.get_logger(__name__) _a : Optional[int] = { ...
689
'''simple docstring''' import argparse import fairseq import torch from transformers import UniSpeechSatConfig, UniSpeechSatForCTC, UniSpeechSatForPreTraining, logging logging.set_verbosity_info() _a : List[Any] = logging.get_logger(__name__) _a : int = { ...
689
1
'''simple docstring''' import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics ...
689
'''simple docstring''' from scipy.stats import spearmanr import datasets _a : str = """ The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no c...
689
1
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TextClassificationPipeline, pipeline, ) from transformers.testing_utils import is_pipeline_test, nested_simplify, require_tf, require_t...
689
'''simple docstring''' from ..utils import DummyObject, requires_backends class _UpperCAmelCase ( metaclass=lowerCAmelCase_ ): a : List[str] =["""onnx"""] def __init__( self,*__SCREAMING_SNAKE_CASE,**__SCREAMING_SNAKE_CASE ): '''simple docstring''' requires_...
689
1
'''simple docstring''' import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _a : List[str] = logging.get_logger(_...
689
'''simple docstring''' from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import loggi...
689
1
'''simple docstring''' _a : List[str] = """ABCDEFGHIJKLMNOPQRSTUVWXYZ""" def _lowerCAmelCase ( ) -> None: __lowerCAmelCase = input("""Enter message: """ ) __lowerCAmelCase = input("""Enter key [alphanumeric]: """ ) __lowerCAmelCase ...
689
'''simple docstring''' def _lowerCAmelCase ( lowercase = 5000_0000 ) -> int: __lowerCAmelCase = set() __lowerCAmelCase = int((limit - 24) ** (1 / 2) ) __lowerCAmelCase = set(range(3 , prime_square_limit + 1 , 2 ) ) primes....
689
1
'''simple docstring''' import numpy as np import torch from torch.nn import CrossEntropyLoss from transformers import AutoModelForCausalLM, AutoTokenizer import datasets from datasets import logging _a : Tuple = """\ """ _a : Tuple = """ Perplexity (PPL) is ...
689
'''simple docstring''' import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers...
689
1
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from .....
689
'''simple docstring''' from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def _lowerCAmelCase ( lowercase ) -> Optional[int]: if not is_accelerate_available(): return method __lowerCAmelCase =...
689
1
'''simple docstring''' import collections import os from typing import List, Optional, Tuple from transformers.utils import is_jieba_available, requires_backends if is_jieba_available(): import jieba from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a : Union[...
689
'''simple docstring''' import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def _lowerCAmelCase ( lowercase , lowercase , lowercase , lowercase , lowercase ) -> Optional[int]: # load base model ...
689
1
'''simple docstring''' import torch from torch import nn from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin class _UpperCAmelCase ( lowerCAmelCase_ , lowerCAmelCase_ ): @register_to_config def __init__( self,*, __SCREAMI...
689
'''simple docstring''' from collections import Counter from timeit import timeit def _lowerCAmelCase ( lowercase = "" , ) -> bool: return sum(c % 2 for c in Counter(input_str.replace(""" """ , """""" ).lower() ).values() ) < 2 def _lowerCAmelCase ( lowe...
689
1