code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __snake_case : List[Any] = { "configuration_roformer": ...
721
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp impor...
691
0
import copy import tempfile import unittest from huggingface_hub import HfFolder, delete_repo from parameterized import parameterized from requests.exceptions import HTTPError from transformers import AutoConfig, GenerationConfig from transformers.testing_utils import TOKEN, USER, is_staging_test class A (...
700
'''simple docstring''' import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require...
691
0
'''simple docstring''' def _lowercase ( lowerCamelCase__ : list ): if not isinstance(lowerCamelCase__, lowerCamelCase__ ): raise ValueError("Input series is not valid, valid series - [2, 4, 6]" ) if len(lowerCamelCase__ ) == 0: raise ValueError("Input li...
701
'''simple docstring''' import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case : Tuple = logging.get_logger(__name__) __snake_case : int = { "facebook/wav2vec2-base-960h": "https://huggingface.c...
691
0
'''simple docstring''' from __future__ import annotations __snake_case : Dict = [-10, -5, 0, 5, 5.1, 11, 13, 21, 3, 4, -21, -10, -5, -1, 0] __snake_case : List[Any] = [-5, 0, 5, 5.1, 11, 13, 21, -1, 4, -1, -10, -5, -1, 0, -1] def _lowercase ( lowerCamelCase__ ...
702
'''simple docstring''' def _lowercase ( lowerCamelCase__ : int, lowerCamelCase__ : int ): return number | (1 << position) def _lowercase ( lowerCamelCase__ : int, lowerCamelCase__ : int ): return number & ~(1 << position) def _lowercase ( lowerCamelCase_...
691
0
'''simple docstring''' import tempfile import torch from diffusers import ( DEISMultistepScheduler, DPMSolverMultistepScheduler, DPMSolverSinglestepScheduler, UniPCMultistepScheduler, ) from .test_schedulers import SchedulerCommonTest class A ( a ): __UpperCAmelCase : ...
703
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # ...
691
0
'''simple docstring''' import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse("9.1.0"): __snake_case : int = { "linear": PIL.Image.Resampling.BILINEAR, ...
704
'''simple docstring''' def _lowercase ( lowerCamelCase__ : list[list] ): _a = current_set.copy() for row_index, row in enumerate(lowerCamelCase__ ): _a = row[0] for column_index, column in enumerate(lowerCamelCase__ ): if mag...
691
0
'''simple docstring''' from __future__ import annotations from collections.abc import Callable from typing import Generic, TypeVar __snake_case : int = TypeVar("T") __snake_case : List[Any] = TypeVar("U") class A ( Generic[T, U] ): def __init__( self , ...
705
'''simple docstring''' import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unorde...
691
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __snake_case : Dict = { "configuration_time_series_transformer": [ "TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "Time...
706
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings __snake_case : Optional[int] = R"\n [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and\n c...
691
0
'''simple docstring''' from __future__ import annotations def _lowercase ( lowerCamelCase__ : dict, lowerCamelCase__ : str ): _a , _a = set(lowerCamelCase__ ), [start] while stack: _a = stack.pop() explored.add(lowerCamelCase__ ...
707
'''simple docstring''' class A : def __init__( self ) -> List[str]: _a = 0 _a = 0 _a = {} def __lowerCAmelCase ( self , snake_case_ ) -> int: if vertex not in self.adjacency: ...
691
0
def _lowercase ( lowerCamelCase__ : Any, lowerCamelCase__ : Any ): _a = [1] for i in range(2, lowerCamelCase__ ): factorials.append(factorials[-1] * i ) assert 0 <= k < factorials[-1] * n, "k out of bounds" _a = [] _a = li...
708
'''simple docstring''' import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, f...
691
0
import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class A ( unittest.TestCase ): def __lowerCAmelCase ( s...
709
'''simple docstring''' import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) __snake_case : Union[st...
691
0
'''simple docstring''' from .glue import glue_convert_examples_to_features, glue_output_modes, glue_processors, glue_tasks_num_labels from .squad import SquadExample, SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features from .utils import DataProcessor, InputExample, InputFeature...
710
'''simple docstring''' import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput de...
691
0
import argparse import json import os import time import zipfile from get_ci_error_statistics import download_artifact, get_artifacts_links from transformers import logging __snake_case : Optional[Any] = logging.get_logger(__name__) def _lowercase ( lowerCamelCase__ : Optional[Any...
711
'''simple docstring''' def _lowercase ( lowerCamelCase__ : list[int], lowerCamelCase__ : list[int], lowerCamelCase__ : int ): return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(lowerCamelCase__ ) ) def _lowercas...
691
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case : List[Any] = logging.get_logger(__name__) __snake_case : Optional[int] = { "MIT/ast-finetuned-audioset-10-10-0.4593": ( "https://hugging...
712
'''simple docstring''' import heapq as hq import math from collections.abc import Iterator class A : def __init__( self , snake_case_ ) -> Optional[int]: _a = str(id_ ) _a = None _a = None _a = ...
691
0
'''simple docstring''' import copy import os import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np import pyarrow as pa import pyarrow.parquet as pq import pytest from datasets.arrow_writer import ArrowWriter, OptimizedTypedSequence, ParquetWriter, TypedSequence from ...
713
'''simple docstring''' __snake_case : List[str] = "Tobias Carryer" from time import time class A : def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=int(time() ) ) -> str: # noqa: B008 _a = mul...
691
0
'''simple docstring''' def _lowercase ( lowerCamelCase__ : str ): _a = [int(lowerCamelCase__ ) for i in ip_va_address.split("." ) if i.isdigit()] return len(lowerCamelCase__ ) == 4 and all(0 <= int(lowerCamelCase__ ) <= 254 for octet in octets ) ...
714
'''simple docstring''' import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small...
691
0
'''simple docstring''' import requests from bsa import BeautifulSoup def _lowercase ( lowerCamelCase__ : str = "https://www.worldometers.info/coronavirus" ): _a = BeautifulSoup(requests.get(lowerCamelCase__ ).text, "html.parser" ) _a = soup.findAll(...
715
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __snake_case : int = { "configuration_bloom": ["BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP", "BloomConfig", "BloomOnnxConfig"],...
691
0
def _lowercase ( lowerCamelCase__ : int, lowerCamelCase__ : int ): return base * power(lowerCamelCase__, (exponent - 1) ) if exponent else 1 if __name__ == "__main__": print("Raise base to the power of exponent using recursion...") __snake_case : List[Any] = ...
716
'''simple docstring''' from ..utils import DummyObject, requires_backends class A ( metaclass=a ): __UpperCAmelCase : int = ["""torch""", """scipy"""] def __init__( self , *snake_case_ , **snake_case_ ) -> Tuple: requires_backends(self ...
691
0
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from itertools import chain from typing import Optional, Union import datasets import numpy as np import torch from datasets import load_dataset import transformers from transformers import ( AutoConfig, ...
717
'''simple docstring''' __snake_case : Dict = { "Pillow": "Pillow<10.0.0", "accelerate": "accelerate>=0.20.3", "av": "av==9.2.0", "beautifulsoup4": "beautifulsoup4", "black": "black~=23.1", "codecarbon": "codecarbon==1.2.0", "cookiecutter": "cookiecutter==1.7.3", ...
691
0
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_flax_available, is_torch_available, is_transformers_available, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDepend...
718
'''simple docstring''' import os import unittest from transformers import BatchEncoding from transformers.models.bert.tokenization_bert import ( BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.models.prophetnet.tokenization_prophetnet imp...
691
0
'''simple docstring''' from maths.is_square_free import is_square_free from maths.prime_factors import prime_factors def _lowercase ( lowerCamelCase__ : int ): _a = prime_factors(lowerCamelCase__ ) if is_square_free(lowerCamelCase__ ): return -1 if len(l...
719
'''simple docstring''' import argparse from copy import deepcopy import numpy as np from datasets import ClassLabel, DatasetDict, load_dataset from evaluate import load from transformers import ( AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, Trainer, TrainerCal...
691
0
'''simple docstring''' def _lowercase ( lowerCamelCase__ : list ): if len(lowerCamelCase__ ) < 2: return collection def circle_sort_util(lowerCamelCase__ : list, lowerCamelCase__ : int, lowerCamelCase__ : int ) -> bool: _a = False if...
720
'''simple docstring''' # Usage: # ./gen-card-allenai-wmt16.py import os from pathlib import Path def _lowercase ( lowerCamelCase__ : Any, lowerCamelCase__ : Optional[int], lowerCamelCase__ : Dict, lowerCamelCase__ : List[str] ): _a = { "en": "Machine learning is ...
691
0
'''simple docstring''' import timeit import numpy as np import datasets from datasets.arrow_writer import ArrowWriter from datasets.features.features import _ArrayXD def _lowercase ( lowerCamelCase__ : List[str] ): def wrapper(*lowerCamelCase__ : Tuple, **lowerCamelCase__ : Tuple ...
721
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp impor...
691
0
from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, ...
700
'''simple docstring''' import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require...
691
0
'''simple docstring''' from abc import ABC, abstractmethod from argparse import ArgumentParser class A ( a ): @staticmethod @abstractmethod def __lowerCAmelCase ( snake_case_ ) -> str: raise NotImplementedError() @abstractmethod def __lowe...
701
'''simple docstring''' import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case : Tuple = logging.get_logger(__name__) __snake_case : int = { "facebook/wav2vec2-base-960h": "https://huggingface.c...
691
0
'''simple docstring''' from __future__ import annotations import math from collections.abc import Callable def _lowercase ( lowerCamelCase__ : Callable[[int | float], int | float], lowerCamelCase__ : int | float, lowerCamelCase__ : int | float, lowerCamelCase__ : int = 100, ): _a ...
702
'''simple docstring''' def _lowercase ( lowerCamelCase__ : int, lowerCamelCase__ : int ): return number | (1 << position) def _lowercase ( lowerCamelCase__ : int, lowerCamelCase__ : int ): return number & ~(1 << position) def _lowercase ( lowerCamelCase_...
691
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.util...
703
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # ...
691
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __snake_case : str = { "configuration_rembert"...
704
'''simple docstring''' def _lowercase ( lowerCamelCase__ : list[list] ): _a = current_set.copy() for row_index, row in enumerate(lowerCamelCase__ ): _a = row[0] for column_index, column in enumerate(lowerCamelCase__ ): if mag...
691
0
'''simple docstring''' import unittest from transformers import ( MODEL_FOR_OBJECT_DETECTION_MAPPING, AutoFeatureExtractor, AutoModelForObjectDetection, ObjectDetectionPipeline, is_vision_available, pipeline, ) from transformers.testing_utils import ( is_pipeline_test, nested_sim...
705
'''simple docstring''' import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unorde...
691
0
'''simple docstring''' import sacrebleu as scb from packaging import version from sacrebleu import CHRF import datasets __snake_case : int = "\\n@inproceedings{popovic-2015-chrf,\n title = \"chr{F}: character n-gram {F}-score for automatic {MT} evaluation\",\n author = \"Popovi{\'c},...
706
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings __snake_case : Optional[int] = R"\n [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and\n c...
691
0
'''simple docstring''' from datetime import datetime as dt import os from github import Github __snake_case : Union[str, Any] = [ "good first issue", "good second issue", "good difficult issue", "feature request", "new model", "wip", ] def _lowercase ( )...
707
'''simple docstring''' class A : def __init__( self ) -> List[str]: _a = 0 _a = 0 _a = {} def __lowerCAmelCase ( self , snake_case_ ) -> int: if vertex not in self.adjacency: ...
691
0
import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast __snake_case : List[Any] = datasets.utils.logging.get_logger(__name__) @dataclass class A ( dat...
708
'''simple docstring''' import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, f...
691
0
import io import math from typing import Dict, Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import convert_to_rgb, normalize, to_channel_dimension_format, to_pil_image from ...image_util...
709
'''simple docstring''' import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) __snake_case : Union[st...
691
0
'''simple docstring''' import gc import random import unittest import numpy as np import torch from transformers import XLMRobertaTokenizer from diffusers import ( AltDiffusionImgaImgPipeline, AutoencoderKL, PNDMScheduler, UNetaDConditionModel, ) from diffusers.image_processor import VaeImagePr...
710
'''simple docstring''' import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput de...
691
0
from __future__ import annotations def _lowercase ( lowerCamelCase__ : list[int], lowerCamelCase__ : list[int], lowerCamelCase__ : int ): _a = list(range(len(lowerCamelCase__ ) ) ) _a = [v / w for v, w in zip(lowerCamelCase__, lowerCamelCase__ ...
711
'''simple docstring''' def _lowercase ( lowerCamelCase__ : list[int], lowerCamelCase__ : list[int], lowerCamelCase__ : int ): return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(lowerCamelCase__ ) ) def _lowercas...
691
0
'''simple docstring''' import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils impor...
712
'''simple docstring''' import heapq as hq import math from collections.abc import Iterator class A : def __init__( self , snake_case_ ) -> Optional[int]: _a = str(id_ ) _a = None _a = None _a = ...
691
0
'''simple docstring''' import math import os import re import sys import unittest from pathlib import Path from typing import Tuple from unittest.mock import patch from parameterized import parameterized from transformers.testing_utils import ( CaptureStderr, ExtendSysPath, TestCasePlus, execut...
713
'''simple docstring''' __snake_case : List[str] = "Tobias Carryer" from time import time class A : def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=int(time() ) ) -> str: # noqa: B008 _a = mul...
691
0
'''simple docstring''' import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class A ( a ): __UpperCAmelCase : List[Any] = (DDPMParallelScheduler,) def __lowerCAmelCase ( self , **snake_case_ ) -> ...
714
'''simple docstring''' import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small...
691
0
'''simple docstring''' def _lowercase ( lowerCamelCase__ : list ): if len(lowerCamelCase__ ) <= 1: return [tuple(lowerCamelCase__ )] _a = [] def generate(lowerCamelCase__ : int, lowerCamelCase__ : list ): if k == 1: res.a...
715
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __snake_case : int = { "configuration_bloom": ["BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP", "BloomConfig", "BloomOnnxConfig"],...
691
0
def _lowercase ( lowerCamelCase__ : Dict, lowerCamelCase__ : List[str], lowerCamelCase__ : Tuple, lowerCamelCase__ : Optional[int]=None ): _a = (path or []) + [u] for v in graph[u]: if visited_edge[u][v] is False: _a , _a = True, True ...
716
'''simple docstring''' from ..utils import DummyObject, requires_backends class A ( metaclass=a ): __UpperCAmelCase : int = ["""torch""", """scipy"""] def __init__( self , *snake_case_ , **snake_case_ ) -> Tuple: requires_backends(self ...
691
0
'''simple docstring''' import argparse import pytorch_lightning as pl import torch from torch import nn from transformers import LongformerForQuestionAnswering, LongformerModel class A ( pl.LightningModule ): def __init__( self , snake_case_ ) -> List[str]: super()...
717
'''simple docstring''' __snake_case : Dict = { "Pillow": "Pillow<10.0.0", "accelerate": "accelerate>=0.20.3", "av": "av==9.2.0", "beautifulsoup4": "beautifulsoup4", "black": "black~=23.1", "codecarbon": "codecarbon==1.2.0", "cookiecutter": "cookiecutter==1.7.3", ...
691
0
'''simple docstring''' __snake_case : Dict = range(2, 20 + 1) __snake_case : Any = [10**k for k in range(ks[-1] + 1)] __snake_case : dict[int, dict[int, list[list[int]]]] = {} def _lowercase ( lowerCamelCase__ : int, lowerCamelCase__ : Optional[Any], l...
718
'''simple docstring''' import os import unittest from transformers import BatchEncoding from transformers.models.bert.tokenization_bert import ( BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.models.prophetnet.tokenization_prophetnet imp...
691
0
'''simple docstring''' from __future__ import annotations def _lowercase ( lowerCamelCase__ : list[float], lowerCamelCase__ : list[float] ): _a = sorted(numsa + numsa ) _a , _a = divmod(len(lowerCamelCase__ ), 2 ) if mod == 1: ...
719
'''simple docstring''' import argparse from copy import deepcopy import numpy as np from datasets import ClassLabel, DatasetDict, load_dataset from evaluate import load from transformers import ( AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, Trainer, TrainerCal...
691
0
'''simple docstring''' from __future__ import annotations def _lowercase ( lowerCamelCase__ : list[int | str] ): create_state_space_tree(lowerCamelCase__, [], 0, [0 for i in range(len(lowerCamelCase__ ) )] ) def _lowercase ( lowerCamelCase__ : list[int | s...
720
'''simple docstring''' # Usage: # ./gen-card-allenai-wmt16.py import os from pathlib import Path def _lowercase ( lowerCamelCase__ : Any, lowerCamelCase__ : Optional[int], lowerCamelCase__ : Dict, lowerCamelCase__ : List[str] ): _a = { "en": "Machine learning is ...
691
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __snake_case : int = logging.get_logger(__name__) __snake_ca...
721
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp impor...
691
0
import unittest import numpy as np from transformers import is_flax_available from transformers.testing_utils import require_flax from ..test_modeling_flax_common import ids_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.generation import ( Fl...
700
'''simple docstring''' import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require...
691
0
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings __snake_case : Optional[int] = R"\n [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and\n c...
701
'''simple docstring''' import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case : Tuple = logging.get_logger(__name__) __snake_case : int = { "facebook/wav2vec2-base-960h": "https://huggingface.c...
691
0
'''simple docstring''' import inspect import os import unittest from dataclasses import dataclass import torch from accelerate import Accelerator, DistributedDataParallelKwargs, GradScalerKwargs from accelerate.state import AcceleratorState from accelerate.test_utils import execute_subprocess_async, require_cu...
702
'''simple docstring''' def _lowercase ( lowerCamelCase__ : int, lowerCamelCase__ : int ): return number | (1 << position) def _lowercase ( lowerCamelCase__ : int, lowerCamelCase__ : int ): return number & ~(1 << position) def _lowercase ( lowerCamelCase_...
691
0
'''simple docstring''' import math class A : def __init__( self , snake_case_=0 ) -> Tuple: # a graph with Node 0,1,...,N-1 _a = n _a = [ [math.inf for j in range(0 , snake_case_ )] for i in range(0 , snake_ca...
703
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # ...
691
0
'''simple docstring''' import os from datetime import datetime as dt from github import Github __snake_case : Tuple = [ "good first issue", "feature request", "wip", ] def _lowercase ( ): _a = Github(os.environ["GITHUB_TOKEN"] ) _a =...
704
'''simple docstring''' def _lowercase ( lowerCamelCase__ : list[list] ): _a = current_set.copy() for row_index, row in enumerate(lowerCamelCase__ ): _a = row[0] for column_index, column in enumerate(lowerCamelCase__ ): if mag...
691
0
'''simple docstring''' import logging import os import sys from dataclasses import dataclass, field from typing import Optional import numpy as np import torch from datasets import load_dataset from torchvision.transforms import Compose, Lambda, Normalize, RandomHorizontalFlip, RandomResizedCrop, ToTensor impo...
705
'''simple docstring''' import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unorde...
691
0
'''simple docstring''' import argparse import glob import logging import os import sys import time from collections import defaultdict from pathlib import Path from typing import Dict, List, Tuple import numpy as np import pytorch_lightning as pl import torch from callbacks import SeqaSeqLoggingCallback, get_ch...
706
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings __snake_case : Optional[int] = R"\n [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and\n c...
691
0
'''simple docstring''' import argparse import fairseq import torch from torch import nn from transformers import ( MBartaaTokenizer, MBartConfig, MBartForCausalLM, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaModel, ...
707
'''simple docstring''' class A : def __init__( self ) -> List[str]: _a = 0 _a = 0 _a = {} def __lowerCAmelCase ( self , snake_case_ ) -> int: if vertex not in self.adjacency: ...
691
0
import importlib import os import fsspec import pytest from fsspec import register_implementation from fsspec.registry import _registry as _fsspec_registry from datasets.filesystems import COMPRESSION_FILESYSTEMS, HfFileSystem, extract_path_from_uri, is_remote_filesystem from .utils import require_lza, require_zst...
708
'''simple docstring''' import os from pathlib import Path from unittest.mock import patch import pytest import zstandard as zstd from datasets.download.download_config import DownloadConfig from datasets.utils.file_utils import ( OfflineModeIsEnabled, cached_path, fsspec_get, fsspec_head, f...
691
0
from math import factorial def _lowercase ( lowerCamelCase__ : int = 100 ): return sum(map(lowerCamelCase__, str(factorial(lowerCamelCase__ ) ) ) ) if __name__ == "__main__": print(solution(int(input("Enter the Number: ").strip())))
709
'''simple docstring''' import argparse import re import numpy as np import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SamConfig, SamImageProcessor, SamModel, SamProcessor, SamVisionConfig, ) __snake_case : Union[st...
691
0
'''simple docstring''' from typing import Any class A : def __init__( self , snake_case_ ) -> List[str]: _a = data _a = None class A : def __init__( self ) -> int: _a = None ...
710
'''simple docstring''' import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput de...
691
0
import argparse from pathlib import Path import requests import torch from PIL import Image from transformers import ( RobertaTokenizer, TrOCRConfig, TrOCRForCausalLM, TrOCRProcessor, VisionEncoderDecoderModel, ViTConfig, ViTImageProcessor, ViTModel, ) from transformers.utils import ...
711
'''simple docstring''' def _lowercase ( lowerCamelCase__ : list[int], lowerCamelCase__ : list[int], lowerCamelCase__ : int ): return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(lowerCamelCase__ ) ) def _lowercas...
691
0
'''simple docstring''' def _lowercase ( lowerCamelCase__ : list[list] ): _a = current_set.copy() for row_index, row in enumerate(lowerCamelCase__ ): _a = row[0] for column_index, column in enumerate(lowerCamelCase__ ): if...
712
'''simple docstring''' import heapq as hq import math from collections.abc import Iterator class A : def __init__( self , snake_case_ ) -> Optional[int]: _a = str(id_ ) _a = None _a = None _a = ...
691
0
'''simple docstring''' import argparse __snake_case : Union[str, Any] = "docs/source/_static/js/custom.js" def _lowercase ( lowerCamelCase__ : str ): with open(lowerCamelCase__, encoding="utf-8", newline="\n" ) as f: _a = f.readlines() ...
713
'''simple docstring''' __snake_case : List[str] = "Tobias Carryer" from time import time class A : def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=int(time() ) ) -> str: # noqa: B008 _a = mul...
691
0
'''simple docstring''' import argparse import torch from transformers import ( SpeechTaConfig, SpeechTaFeatureExtractor, SpeechTaForSpeechToSpeech, SpeechTaForSpeechToText, SpeechTaForTextToSpeech, SpeechTaProcessor, SpeechTaTokenizer, logging, ) from transformers.tokenization_...
714
'''simple docstring''' import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small...
691
0
'''simple docstring''' from typing import Dict import numpy as np import torch from . import residue_constants as rc from .tensor_utils import tensor_tree_map, tree_map def _lowercase ( lowerCamelCase__ : Dict[str, torch.Tensor] ): _a = [] _a = [] _a ...
715
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __snake_case : int = { "configuration_bloom": ["BLOOM_PRETRAINED_CONFIG_ARCHIVE_MAP", "BloomConfig", "BloomOnnxConfig"],...
691
0
import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unordered, map_nested, tem...
716
'''simple docstring''' from ..utils import DummyObject, requires_backends class A ( metaclass=a ): __UpperCAmelCase : int = ["""torch""", """scipy"""] def __init__( self , *snake_case_ , **snake_case_ ) -> Tuple: requires_backends(self ...
691
0
'''simple docstring''' import hashlib import unittest from typing import Dict import numpy as np from transformers import ( MODEL_FOR_MASK_GENERATION_MAPPING, TF_MODEL_FOR_MASK_GENERATION_MAPPING, is_vision_available, pipeline, ) from transformers.pipelines import MaskGenerationPipeline from tr...
717
'''simple docstring''' __snake_case : Dict = { "Pillow": "Pillow<10.0.0", "accelerate": "accelerate>=0.20.3", "av": "av==9.2.0", "beautifulsoup4": "beautifulsoup4", "black": "black~=23.1", "codecarbon": "codecarbon==1.2.0", "cookiecutter": "cookiecutter==1.7.3", ...
691
0
'''simple docstring''' import argparse import os import sys from unittest.mock import patch import pytorch_lightning as pl import timeout_decorator import torch from distillation import SummarizationDistiller, distill_main from finetune import SummarizationModule, main from transformers import MarianMTModel fr...
718
'''simple docstring''' import os import unittest from transformers import BatchEncoding from transformers.models.bert.tokenization_bert import ( BasicTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.models.prophetnet.tokenization_prophetnet imp...
691
0
'''simple docstring''' import logging from dataclasses import dataclass, field from pathlib import Path from typing import Optional, Union from .generation.configuration_utils import GenerationConfig from .training_args import TrainingArguments from .utils import add_start_docstrings __snake_case : Dict ...
719
'''simple docstring''' import argparse from copy import deepcopy import numpy as np from datasets import ClassLabel, DatasetDict, load_dataset from evaluate import load from transformers import ( AutoModelForSequenceClassification, AutoTokenizer, DataCollatorWithPadding, Trainer, TrainerCal...
691
0
'''simple docstring''' import argparse import json import os import fairseq import torch from fairseq.data import Dictionary from transformers import ( HubertConfig, HubertForCTC, HubertModel, WavaVecaCTCTokenizer, WavaVecaFeatureExtractor, WavaVecaProcessor, logging, ) logging.se...
720
'''simple docstring''' # Usage: # ./gen-card-allenai-wmt16.py import os from pathlib import Path def _lowercase ( lowerCamelCase__ : Any, lowerCamelCase__ : Optional[int], lowerCamelCase__ : Dict, lowerCamelCase__ : List[str] ): _a = { "en": "Machine learning is ...
691
0
'''simple docstring''' from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailabl...
721
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp impor...
691
0
'''simple docstring''' def lowerCAmelCase_ ( __A : int , __A : int , __A : int ): '''simple docstring''' snake_case: str = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total ...
692
'''simple docstring''' import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartForSeq...
692
1
'''simple docstring''' from datasets.utils.patching import _PatchedModuleObj, patch_submodule from . import _test_patching def lowerCAmelCase_ ( ): '''simple docstring''' import os as original_os from os import path as original_path from os import rename as original_rename ...
692
'''simple docstring''' import math def lowerCAmelCase_ ( __A : int ): '''simple docstring''' if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3...
692
1
'''simple docstring''' import shutil import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_tf_cross_test, require_tf, require_torch, require_torchvision, require_vision, ) from transformers.utils import is_tf_available, is_torch_available, is_v...
692
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common...
692
1
'''simple docstring''' import torch from transformers import CamembertForMaskedLM, CamembertTokenizer def lowerCAmelCase_ ( __A : str , __A : Dict , __A : int , __A : Dict=5 ): '''simple docstring''' assert masked_input.count('<mask>' ) == 1 snak...
692
'''simple docstring''' from typing import Any, Dict, Optional import torch import torch.nn.functional as F from torch import nn from ..utils import maybe_allow_in_graph from .activations import get_activation from .attention_processor import Attention from .embeddings import CombinedTimestepLabelEmbeddings ...
692
1
'''simple docstring''' from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import...
692
'''simple docstring''' import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.te...
692
1
'''simple docstring''' from __future__ import annotations def lowerCAmelCase_ ( __A : list[int] ): '''simple docstring''' return len(set(__A ) ) == len(__A ) if __name__ == "__main__": import doctest doctest.testmod()
692
'''simple docstring''' from math import asin, atan, cos, radians, sin, sqrt, tan __UpperCAmelCase = 6378137.0 __UpperCAmelCase = 6356752.314245 __UpperCAmelCase = 6_378_137 def lowerCAmelCase_ ( __A : float , __A : float , __A : float ...
692
1
'''simple docstring''' class SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self ): '''simple docstring''' snake_case: dict[str, TrieNode] = {} # Mapping from char to TrieNode snake_case: Union[str, Any] ...
692
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __UpperCAmelCase = { "configuration_roformer": ["ROFORMER_PR...
692
1
'''simple docstring''' import os from typing import Any, Callable, Dict, List, Optional, Tuple, Union import torch from torch import nn from ...models.controlnet import ControlNetModel, ControlNetOutput from ...models.modeling_utils import ModelMixin from ...utils import logging __UpperCAmelCase ...
692
'''simple docstring''' import argparse import torch from datasets import load_dataset from donut import DonutModel from transformers import ( DonutImageProcessor, DonutProcessor, DonutSwinConfig, DonutSwinModel, MBartConfig, MBartForCausalLM, VisionEncoderDecoderModel, XLMRob...
692
1
'''simple docstring''' import json import os from collections import Counter import torch import torchvision import torchvision.transforms as transforms from PIL import Image from torch import nn from torch.utils.data import Dataset __UpperCAmelCase = {1: (1, 1), 2: (2, 1), 3: (3, 1), 4: (2, 2)...
692
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, trans...
692
1
'''simple docstring''' class SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self ): '''simple docstring''' snake_case: Dict = 0 snake_case: Dict = 0 snake_case: Dict = {} ...
692
'''simple docstring''' import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAtt...
692
1
'''simple docstring''' import os def lowerCAmelCase_ ( __A : str = "matrix.txt" ): '''simple docstring''' with open(os.path.join(os.path.dirname(__A ) , __A ) ) as in_file: snake_case: Tuple = in_file.read() snake_case: List[str] ...
692
'''simple docstring''' from queue import PriorityQueue from typing import Any import numpy as np def lowerCAmelCase_ ( __A : dict , __A : str , __A : set , __A : set , __A : dict , __A : dict , __A : PriorityQueue , __A : dict , __...
692
1
'''simple docstring''' from abc import ABC, abstractmethod from argparse import ArgumentParser class SCREAMING_SNAKE_CASE ( snake_case ): '''simple docstring''' @staticmethod @abstractmethod def _UpperCamelCase ( SCREAMING_SNAKE_CASE__ ): '''sim...
692
'''simple docstring''' import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase ...
692
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available, is_vision_available, ) __UpperCAmelCase = { "configuration_perceiver": ["PERCEIVER_PRETRAINED_CONFIG_A...
692
'''simple docstring''' from operator import delitem, getitem, setitem import pytest from data_structures.hashing.hash_map import HashMap def lowerCAmelCase_ ( __A : Optional[Any] ): '''simple docstring''' return getitem, k def lowerCAmelCase_ ( __A : Any...
692
1
'''simple docstring''' import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAP...
692
'''simple docstring''' import argparse import json import os import fairseq import torch from torch import nn from transformers import ( SpeechaTextaConfig, SpeechaTextaForCausalLM, SpeechaTextaTokenizer, SpeechEncoderDecoderConfig, SpeechEncoderDecoderModel, WavaVecaConfig, Wava...
692
1
'''simple docstring''' import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets __UpperCAmelCase = datasets.logging.get_logger(__name__) __UpperCAmelCase = "\\n@InProceeding...
692
'''simple docstring''' def lowerCAmelCase_ ( __A : int = 1_00 ): '''simple docstring''' snake_case: List[str] = n * (n + 1) * (2 * n + 1) / 6 snake_case: List[Any] = (n * (n + 1) / 2) ** 2 return int(square_of_sum - sum_of_squares ) ...
692
1
'''simple docstring''' import os import sys import unittest __UpperCAmelCase = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, crea...
692
'''simple docstring''' import argparse from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import BigBirdPegasusConfig, BigBirdPegasusForConditionalGeneration __UpperCAmelCase = [ # tf -> hf ("/", "."), ("layer_", "layers."), ("k...
692
1
'''simple docstring''' import torch from diffusers import KDPMaDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE ( snake_case ): '''simple docstring''' __UpperCamelCase = (KDPMaDis...
692
'''simple docstring''' def lowerCAmelCase_ ( __A : List[str] ): '''simple docstring''' snake_case: str = [0] * len(__A ) snake_case: Tuple = [] snake_case: Tuple = [1] * len(__A ) for values in graph.values(): ...
692
1
'''simple docstring''' import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig __UpperCAmelCase = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE : ''...
692
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision...
692
1
'''simple docstring''' import math def lowerCAmelCase_ ( __A : int ): '''simple docstring''' if not isinstance(__A , __A ): snake_case: List[str] = f"""Input value of [number={number}] must be an integer""" raise TypeError(__A ) if num...
692
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { "microsoft/swinv2-tiny-patch4-window8-256": ( "https://huggingface.co/microsoft/swinv2-tiny-patch4...
692
1
'''simple docstring''' from typing import Any class SCREAMING_SNAKE_CASE : '''simple docstring''' def __init__( self , SCREAMING_SNAKE_CASE__ ): '''simple docstring''' snake_case: Optional[Any] = data snake_case: int ...
692
'''simple docstring''' import os import sys import unittest __UpperCAmelCase = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, "utils")) import check_dummies # noqa: E402 from check_dummies import create_dummy_files, crea...
692
1
'''simple docstring''' from __future__ import annotations from typing import Generic, TypeVar __UpperCAmelCase = TypeVar("T") class SCREAMING_SNAKE_CASE ( Generic[T] ): '''simple docstring''' def __init__( self , SCREAMING_SNAKE_CASE__ ): '''simp...
692
'''simple docstring''' import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig __UpperCAmelCase = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE : ''...
692
1
'''simple docstring''' import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def lowerCAmelCas...
692
'''simple docstring''' import importlib import os import fsspec import pytest from fsspec import register_implementation from fsspec.registry import _registry as _fsspec_registry from datasets.filesystems import COMPRESSION_FILESYSTEMS, HfFileSystem, extract_path_from_uri, is_remote_filesystem from .utils ...
692
1
'''simple docstring''' import inspect import tempfile from collections import OrderedDict, UserDict from collections.abc import MutableMapping from contextlib import ExitStack, contextmanager from dataclasses import fields from enum import Enum from typing import Any, ContextManager, List, Tuple import numpy...
692
'''simple docstring''' import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartForSeq...
692
1
'''simple docstring''' from __future__ import annotations from cmath import sqrt def lowerCAmelCase_ ( __A : int , __A : int , __A : int ): '''simple docstring''' if a == 0: raise ValueError('Coefficient \'a\' must not be zero.' ) snake_case: T...
692
'''simple docstring''' import math def lowerCAmelCase_ ( __A : int ): '''simple docstring''' if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3...
692
1
'''simple docstring''' from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) # TODO Update this __UpperCAmelCase = { "facebook/esm-1b...
692
'''simple docstring''' import json import os import re import shutil import tempfile import unittest from typing import Tuple from transformers import AddedToken, BatchEncoding, ByTaTokenizer from transformers.utils import cached_property, is_tf_available, is_torch_available from ...test_tokenization_common...
692
1
'''simple docstring''' # limitations under the License. from typing import Optional, Tuple, Union import torch from diffusers import DiffusionPipeline, ImagePipelineOutput class SCREAMING_SNAKE_CASE ( snake_case ): '''simple docstring''' def __init__( self , SCREAMING_...
692
'''simple docstring''' from typing import Any, Dict, Optional import torch import torch.nn.functional as F from torch import nn from ..utils import maybe_allow_in_graph from .activations import get_activation from .attention_processor import Attention from .embeddings import CombinedTimestepLabelEmbeddings ...
692
1
'''simple docstring''' import math def lowerCAmelCase_ ( __A : int = 1_00 ): '''simple docstring''' snake_case: int = sum(i * i for i in range(1 , n + 1 ) ) snake_case: Dict = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ...
692
'''simple docstring''' import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.te...
692
1
'''simple docstring''' import colorsys from PIL import Image # type: ignore def lowerCAmelCase_ ( __A : float , __A : float , __A : int ): '''simple docstring''' snake_case: Any = x snake_case: Any = y for step in ...
692
'''simple docstring''' from math import asin, atan, cos, radians, sin, sqrt, tan __UpperCAmelCase = 6378137.0 __UpperCAmelCase = 6356752.314245 __UpperCAmelCase = 6_378_137 def lowerCAmelCase_ ( __A : float , __A : float , __A : float ...
692
1
'''simple docstring''' import argparse import requests import torch # pip3 install salesforce-lavis # I'm actually installing a slightly modified version: pip3 install git+https://github.com/nielsrogge/LAVIS.git@fix_lavis from lavis.models import load_model_and_preprocess from PIL import Image from transfo...
692
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __UpperCAmelCase = { "configuration_roformer": ["ROFORMER_PR...
692
1
'''simple docstring''' import os import unittest from transformers.models.phobert.tokenization_phobert import VOCAB_FILES_NAMES, PhobertTokenizer from ...test_tokenization_common import TokenizerTesterMixin class SCREAMING_SNAKE_CASE ( snake_case , unittest.TestCase ): '''simple do...
692
'''simple docstring''' import argparse import torch from datasets import load_dataset from donut import DonutModel from transformers import ( DonutImageProcessor, DonutProcessor, DonutSwinConfig, DonutSwinModel, MBartConfig, MBartForCausalLM, VisionEncoderDecoderModel, XLMRob...
692
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available __UpperCAmelCase = { "configuration_gpt_neo": ["GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP", "GPTNeoConfig", "GPTNeoOnnxConfig"], } try...
692
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, trans...
692
1
'''simple docstring''' import json import os import shutil import tempfile import unittest import numpy as np import pytest from transformers import BertTokenizer, BertTokenizerFast from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES from transformers.testing_utils import require_vision...
692
'''simple docstring''' import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAtt...
692
1
'''simple docstring''' def lowerCAmelCase_ ( __A : int , __A : int ): '''simple docstring''' snake_case: Optional[Any] = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): snake_case: Optional[int] = ...
692
'''simple docstring''' from queue import PriorityQueue from typing import Any import numpy as np def lowerCAmelCase_ ( __A : dict , __A : str , __A : set , __A : set , __A : dict , __A : dict , __A : PriorityQueue , __A : dict , __...
692
1