code
stringlengths
82
53.2k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' from typing import Optional, Union import torch from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention from ...modeli...
585
'''simple docstring''' from __future__ import annotations def __SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): if len(_SCREAMING_SNAKE_CASE ) == 0: return [] _snake_case, _snake_case = min(_SCREAMING_SNAKE_CASE ), max(_SCREAMING_SNAKE_CASE ) _snake_case = int(...
585
1
"""simple docstring""" import unittest from transformers import SPIECE_UNDERLINE, ReformerTokenizer, ReformerTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, require_torch, slow from transformers.utils import cached_property from ...test_tokeni...
706
"""simple docstring""" import argparse import os import re import tensorflow as tf import torch from transformers import BertConfig, BertModel from transformers.utils import logging logging.set_verbosity_info() a = logging.get_logger(__name__) def lowercase (snake_case__ : s...
529
0
import shutil import tempfile import unittest from transformers import ClapFeatureExtractor, ClapProcessor, RobertaTokenizer, RobertaTokenizerFast from transformers.testing_utils import require_sentencepiece, require_torchaudio from .test_feature_extraction_clap import floats_list @require_torchaudio @require_sente...
548
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.u...
437
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available SCREAMING_SNAKE_CASE__ : Union[str, Any] = { """configuration_canine""": ["""CANINE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Can...
233
'''simple docstring''' from __future__ import annotations import os from typing import Any import requests SCREAMING_SNAKE_CASE__ : Any = """https://api.github.com""" # https://docs.github.com/en/free-pro-team@latest/rest/reference/users#get-the-authenticated-user SCREAMING_SNAKE_CASE__ : ...
233
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class __a (metaclass=UpperCamelCase_): '''simple docstring''' _SCREAMING_SNAKE_CASE :Optional[Any] = ["""flax""", """transformers"""] def __init__( self , *_a , **_a ) -> List[str]...
680
"""simple docstring""" import numpy as np import qiskit def _lowercase ( __lowerCAmelCase = 8 , __lowerCAmelCase = None ) -> str: SCREAMING_SNAKE_CASE__ : List[Any] = np.random.default_rng(seed=__lowerCAmelCase ) # Roughly 25% of the qubits will contrib...
680
1
import logging import os from dataclasses import dataclass from enum import Enum from typing import List, Optional, Union from filelock import FileLock from transformers import PreTrainedTokenizer, is_tf_available, is_torch_available snake_case_ = logging.getLogger(__name__) @datacla...
262
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case_ = { '''configuration_instructblip''': [ '''INSTRUCTBLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InstructBlipConfig''', '''InstructBli...
262
1
import math UpperCamelCase = 10 UpperCamelCase = 7 UpperCamelCase = BALLS_PER_COLOUR * NUM_COLOURS def __magic_name__ ( SCREAMING_SNAKE_CASE = 20 ) -> str: _lowercase : Tuple = math.comb(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ...
66
"""simple docstring""" _snake_case = 6_5521 def lowerCAmelCase__ ( UpperCamelCase__ ): '''simple docstring''' _a : List[str] = 1 _a : Optional[int] = 0 for plain_chr in plain_text: _a : Dict = (a + ord(UpperCamelC...
389
0
import os from math import logaa def snake_case_ ( lowercase__ : str = "base_exp.txt" ): '''simple docstring''' _lowerCAmelCase =0 _lowerCAmelCase =0 for i, line in enumerate(open(os.path.join(os.path.dirname(lowercase__ ) , lowercase__ ) ...
701
from __future__ import annotations import unittest from transformers import EsmConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attentio...
149
0
from ..utils import DummyObject, requires_backends class _a ( metaclass=UpperCamelCase__): """simple docstring""" UpperCamelCase__ = ["""transformers""", """torch""", """note_seq"""] def __init__( self: List[str] , *__lowerCamelCase...
380
def lowerCAmelCase_ ( A_): UpperCamelCase__: Optional[int] = len(A_) for i in range(1 ,A_): UpperCamelCase__: List[Any] = collection[i] UpperCamelCase__: Tuple = 0 UpperCamelCase__: Union[str, Any] = i - 1 while low...
380
1
"""simple docstring""" def UpperCAmelCase__ (snake_case__ : int = 1_00_00_00 ): """simple docstring""" _snake_case : Tuple = set(range(3 , snake_case__ , 2 ) ) primes.add(2 ) for p in range(3 , snake_case__ , 2 ): if p n...
28
"""simple docstring""" import os import sys import unittest A_ = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, '''utils''')) import check_dummies # noqa: E402 from check_dummies import create_dummy_files...
28
1
'''simple docstring''' import random def lowerCamelCase (_SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : float , _SCREAMING_SNAKE_CASE : bool = False ): __a : dict = {i: [] for i in range(_SCREAMING_SNAKE_CASE )} # if probability is greater or equal than 1, t...
476
'''simple docstring''' from __future__ import annotations import math def lowerCamelCase (_SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : int , _SCREAMING_SNAKE_CASE : bool , _SCREAMING_SNAKE_CASE : list[int] , _SCREAMING_SNAKE_CASE : float ): if depth < 0: raise Valu...
476
1
from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo a = '\\n@misc{wu2016googles,\n title={Google\'s Neural Machine Translation System: Bridging the Gap between Human and Machine Translation},\n author={Yonghui Wu and Mike Schuste...
650
import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class UpperCamelCase__ ( __magic_name__ ): __SCREAMING_SNAKE_CASE : str = (UnCLIPScheduler,) def UpperCAmelCase__ ( self : int , **UpperCamelC...
650
1
import os import warnings from typing import List, Optional from ...tokenization_utils_base import BatchEncoding from ...utils import logging from .configuration_rag import RagConfig a__ = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE_ : """simple docstring""...
279
"""simple docstring""" from __future__ import annotations from fractions import Fraction def UpperCAmelCase ( snake_case : int , snake_case : int ): return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) ...
227
0
import warnings from typing import List, Optional, Union from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class lowerCA...
516
import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor _lowerCamelCase : Optional[Any] = logging.get_logger(__name__) class lowerCAmelCase__ ( __magic_name__ ): '''simple docstring''' def __init__( self , ...
516
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { """google/switch-base-8""": """https://huggingface.co/google/switch-base-8/blob/main/config.json""", } c...
88
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...
61
0
from ..utils import DummyObject, requires_backends class UpperCamelCase ( metaclass=snake_case__ ): """simple docstring""" snake_case = ["torch"] def __init__( self : int ,*_SCREAMING_SNAKE_CASE : int ,**_SCREAMING_SNAKE_CASE : Tuple ) -> Optiona...
110
from __future__ import annotations from collections.abc import Generator import requests from bsa import BeautifulSoup lowerCAmelCase_ = 'https://www.indeed.co.in/jobs?q=mobile+app+development&l=' def snake_case ( UpperCAmelCase : str = "mumbai" ): A = BeautifulS...
110
1
"""simple docstring""" import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class lowercase__ ( A ): ...
573
"""simple docstring""" def UpperCAmelCase ( A : int , A : int ): '''simple docstring''' return int((input_a, input_a).count(1 ) != 0 ) def UpperCAmelCase ( ): '''simple docstring''' assert or_gate(0 , 0 ...
573
1
import os import re from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _SCREAMING_SNAKE_CASE : Any = logging.get_logger(__name__) _SCREAMING_SNA...
472
from importlib import import_module from .logging import get_logger _SCREAMING_SNAKE_CASE : Optional[int] = get_logger(__name__) class UpperCAmelCase__ : """simple docstring""" def __init__( self : Optional[int] , __lowerCamelCase : Optional[Any] , __lowerCam...
472
1
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp import M...
325
'''simple docstring''' import sys import webbrowser import requests from bsa import BeautifulSoup from fake_useragent import UserAgent if __name__ == "__main__": print("Googling.....") __A = "https://www.google.com/search?q=" + " ".join(sys.argv[1:]) __A = requests.get(url, heade...
325
1
"""simple docstring""" import inspect import unittest import warnings from math import ceil, floor from transformers import LevitConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils impor...
366
"""simple docstring""" import argparse import numpy as np import torch from transformers import SpeechTaHifiGan, SpeechTaHifiGanConfig, logging logging.set_verbosity_info() __A = logging.get_logger('''transformers.models.speecht5''') def lowercase_ ( _lowerCamelCase: Any , _lowerCamelCase: ...
366
1
'''simple docstring''' import gc import random import unittest import numpy as np import torch from PIL import Image from diffusers import ( DDIMScheduler, KandinskyVaaImgaImgPipeline, KandinskyVaaPriorPipeline, UNetaDConditionModel, VQModel, ) from diffuser...
207
'''simple docstring''' from typing import Callable, Optional from .. import Features from ..packaged_modules.generator.generator import Generator from .abc import AbstractDatasetInputStream class UpperCAmelCase ( snake_case_ ): def __init__( self :...
207
1
from __future__ import annotations import random # Maximum size of the population. Bigger could be faster but is more memory expensive. UpperCamelCase = 200 # Number of elements selected in every generation of evolution. The selection takes # place from best to worst of that generation and must be ...
152
import gc import unittest import numpy as np import torch from diffusers import AutoencoderKL, DDIMScheduler, DiTPipeline, DPMSolverMultistepScheduler, TransformeraDModel from diffusers.utils import is_xformers_available, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determin...
152
1
'''simple docstring''' from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar __lowerCAmelCase = TypeVar('T') class _lowerCAmelCase ( Generic[T] ): '''simple docstring''' def __init__(self , UpperCAmelCase ) ...
585
'''simple docstring''' import os import sys import warnings from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xope...
585
1
'''simple docstring''' # tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]'...
710
import inspect import unittest from transformers import MobileViTConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...t...
249
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unles...
217
def __lowerCAmelCase ( a__ , a__ ) -> bool: return numa ^ numa < 0 if __name__ == "__main__": import doctest doctest.testmod()
219
0
"""simple docstring""" import time import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, torch_device from ..test_modeling_common import ids_tensor if is_torch_available(): import torch from transformers.generation impor...
135
"""simple docstring""" import unittest from knapsack import knapsack as k class _A ( unittest.TestCase ): """simple docstring""" def __snake_case ( self : Dict): a : Tuple = 0 a : An...
135
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a = { "configuration_timesformer": ["TIMESFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "TimesformerConfig"], } try: if not is_torch_available(): raise Opti...
518
import argparse import os from pathlib import Path from typing import Dict import tensorflow as tf import torch from tqdm import tqdm from transformers import PegasusConfig, PegasusForConditionalGeneration, PegasusTokenizer from transformers.models.pegasus.configuration_pegasus import DEFAULTS, ta...
518
1
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging __lowerCAmelCase = logging.ge...
335
from collections import OrderedDict from typing import Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...feature_extraction_utils import FeatureExtractionMixin from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...tokeniza...
335
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) A_ : str = {"configuration_xlnet": ["XLNET_PRETRAINED_CO...
38
# We ignore warnings about stepping the scheduler since we step it ourselves during gradient accumulation import warnings from .state import AcceleratorState, GradientState warnings.filterwarnings('ignore', category=UserWarning, module='torch.optim.lr_scheduler') class _a : '''simple docstring''' ...
520
0
'''simple docstring''' import numpy # List of input, output pairs lowerCAmelCase: List[Any] = ( ((5, 2, 3), 1_5), ((6, 5, 9), 2_5), ((1_1, 1_2, 1_3), 4_1), ((1, 1, 1), 8), ((1_1, 1_2, 1_3), 4_1), ) lowerCAmelCase: str = (((5_1_5, 2_2, 1_3), 5_5_5), ((6_1, 3_5, 4_9), 1_5_0)...
195
'''simple docstring''' from math import factorial, pi def lowerCamelCase__ ( _A , _A = 30 ): if not isinstance(_A , (int, float) ): raise ValueError('maclaurin_sin() requires either an int or float for theta' ) if not isinstance(_A , _A ) or accuracy <= ...
195
1
'''simple docstring''' import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger UpperCAmelCase_ : List[Any] = '<<<<<<< This should probably be modified because it mentions: ...
533
'''simple docstring''' import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed f...
533
1
"""simple docstring""" import inspect import unittest import torch import torch.nn as nn from accelerate.hooks import ( AlignDevicesHook, ModelHook, SequentialHook, add_hook_to_module, attach_align_device_hook, remove_hook_from_module, remove_hook_from_submodules, ) from accelera...
498
"""simple docstring""" from timeit import timeit def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->int: if number < 0: raise ValueError("""the value of input must not be negative""" ) A__ : Optional[int] = 0 while number: ...
498
1
import pytest import datasets # Import fixture modules as plugins snake_case_ = ['''tests.fixtures.files''', '''tests.fixtures.hub''', '''tests.fixtures.fsspec'''] def snake_case__ ( SCREAMING_SNAKE_CASE_ : Union[str, Any] , SCREAMING_SNAKE_CASE_ : Dict )...
164
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available __SCREAMING_SNAKE_CASE :Dict = { '''configuration_chinese_clip''': [ '''CHINESE_CLIP_PRETRAINED_CONFIG_ARCHIVE_MAP''', ...
236
0
import requests from bsa import BeautifulSoup def UpperCamelCase__ ( A__ , A__ ) -> str: snake_case__ : Dict = BeautifulSoup(requests.get(A__ , params=A__ ).content , 'html.parser' ) snake_case__ : Any = soup.find('d...
699
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ : Tuple = logging.get_logger(__name__) lowerCAmelCase__ : Union[str, An...
699
1
from __future__ import annotations import unittest from transformers import XGLMConfig, XGLMTokenizer, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor,...
81
'''simple docstring''' from typing import List, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ : Union[str, Any] = logging.get_logger(__name__) lowerCAmelCase_ : Optional[int] = { "huggingface/informer-to...
489
0
"""simple docstring""" import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def ...
628
"""simple docstring""" import copy from typing import Dict, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto import CONFIG_MAPPING from ..detr import DetrConfig from ..swin import SwinConfig lowerCAmelCase__ = { 'facebook/maskformer-swin-base-ade...
628
1
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf ...
159
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _A = { 'configuration_squeezebert': [ 'SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SqueezeBertConf...
159
1
from itertools import count def lowerCamelCase_ ( A : int = 50 ): """simple docstring""" lowerCAmelCase_ = [1] * min_block_length for n in count(A ): fill_count_functions.append(1 ) for block_length in range(A , n + 1 ): for b...
701
import torch def lowerCamelCase_ ( ): """simple docstring""" if torch.cuda.is_available(): lowerCAmelCase_ = torch.cuda.device_count() else: lowerCAmelCase_ = 0 print(F'Successfully ran on {num_gpus} GPUs' ) if __name__ == "__main__": mai...
413
0
from __future__ import annotations import math import numpy as np from numpy.linalg import norm def __lowerCAmelCase ( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) ) ) def __lowerCAmelCase...
413
import numpy as np from cva import COLOR_BGR2GRAY, cvtColor, imread from numpy import array, uinta from PIL import Image from digital_image_processing import change_contrast as cc from digital_image_processing import convert_to_negative as cn from digital_image_processing import sepia as sp from digital_image_proc...
413
1
import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def _lowercase ( ) -> Any: '''simple docstring''' with offline(OfflineSimulationMode.CON...
713
import unittest from transformers import AutoConfig, AutoTokenizer, BertConfig, TensorType, is_flax_available from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, require_flax, slow if is_flax_available(): import jax from transformers.models.auto.modeling_flax_auto import FlaxAutoModel ...
184
0
"""simple docstring""" from __future__ import annotations from typing import Dict from ...configuration_utils import PretrainedConfig A_ : Union[str, Any] = { "susnato/ernie-m-base_pytorch": "https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/config.json", ...
196
"""simple docstring""" A_ : Any = "0.21.0" from .accelerator import Accelerator from .big_modeling import ( cpu_offload, cpu_offload_with_hook, disk_offload, dispatch_model, init_empty_weights, init_on_device, load_checkpoint_and_dispatch, ) from...
196
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = { "RWKV/rwkv-4-169m-pile": "https://huggingface.co/RWKV/rwkv-4-169m-pile/resolve/main/config.json"...
707
"""simple docstring""" import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig...
122
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __UpperCamelCase = logging.get_logger(__name__) __UpperCamelCase = { 'google/switch-base-8': 'https://huggingface.co/google/switch-base-8/blob/main/config.json', } class lowerCamelCase__ ...
551
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __UpperCamelCase = { 'configuration_xlm_roberta_xl': [ 'XLM_ROBERTA_XL_PRETRAINED_CONFIG_ARCHIVE_MAP', 'XLMRobertaXLConfig', 'XLMRobertaXLOnnxConfig', ]...
551
1
"""simple docstring""" def lowerCAmelCase (__UpperCamelCase : int , __UpperCamelCase : int ): """simple docstring""" if not isinstance(__UpperCamelCase , __UpperCamelCase ): raise ValueError('''iterations must be defined as integers''' ...
296
"""simple docstring""" def lowerCAmelCase (__UpperCamelCase : int = 1_0_0_0_0_0_0 ): """simple docstring""" __UpperCamelCase =1 __UpperCamelCase =1 __UpperCamelCase ={1: 1} for inputa in range(2 , __UpperCamelCase ): __Upper...
296
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __lowercase = { '''configuration_llama''': ['''LLAMA_PRETRAINED_CONFIG_ARCHIVE_M...
370
'''simple docstring''' import os from distutils.util import strtobool def snake_case__ ( _A: int , _A: List[Any] ) -> Union[str, Any]: '''simple docstring''' for e in env_keys: lowerCAmelCase = int(os.environ.get(_A , -1 ) ) if val >= 0: r...
370
1
import argparse import random import joblib import numpy as np import torch from igf.igf import ( SecondaryLearner, collect_objective_set, compute_perplexity, generate_datasets, load_gpta, recopy_gpta, set_seed, train_secondary_learner, ) from torch.utils.data import D...
409
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) __a = {'configuration_mbart': ['MBART_PRETRAINED...
409
1
"""simple docstring""" def snake_case ( A__ ,A__ ): if a < 0 or b < 0: raise ValueError("the value of both inputs must be positive" ) UpperCAmelCase_ : str = str(bin(A__ ) )[2:] # remove the leading "0b" UpperCAmelCase_ : Optional[int] = str(bin(A__ ) )[2...
95
from math import atan, cos, radians, sin, tan from .haversine_distance import haversine_distance __UpperCamelCase : Optional[Any] = 6_378_137.0 __UpperCamelCase : Any = 6_356_752.314_245 __UpperCamelCase : Optional[int] = 6378137 def _UpperCAmelCase ( Up...
519
0
from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Sequence, Value from .base import TaskTemplate @dataclass(frozen=snake_case_ ) class SCREAMING_SNAKE_CASE_ ( snake_case_ ): # `task` is not a ClassVar since we want it t...
534
def SCREAMING_SNAKE_CASE__ ( __a ): if not isinstance(__a , __a ): raise ValueError('Input must be an integer' ) if input_num <= 0: raise ValueError('Input must be positive' ) return sum( divisor for divisor in range(1 , input_num // 2 + 1 ) if input_num % divi...
534
1
import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class lowerCamelCase (unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self ) -> Un...
406
import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class __A ( ctypes.Structure ): # _fields is a specific attr expected by ctypes lowerCAmelCase_ : str = [("size", ctypes.c_int), ("visible", cty...
343
0
"""simple docstring""" import sys import turtle def __lowerCAmelCase ( lowercase : tuple[float, float] , lowercase : tuple[float, float] ) -> tuple[float, float]: """simple docstring""" return (pa[0] + pa[0]) / 2, (pa[1] + pa[1]) / 2 def ...
117
"""simple docstring""" import os import tempfile import unittest from pathlib import Path from transformers import AutoConfig, is_torch_available from transformers.testing_utils import require_torch, torch_device if is_torch_available(): from transformers import PyTorchBenchmark, PyTorchBenchmarkArguments ...
117
1
'''simple docstring''' import os from datetime import datetime as dt from github import Github __lowerCAmelCase = [ 'good first issue', 'feature request', 'wip', ] def _UpperCAmelCase ( ): a_ : Optional[int] = Github(os.envi...
466
'''simple docstring''' import logging import os from dataclasses import dataclass, field from functools import partial from pathlib import Path from tempfile import TemporaryDirectory from typing import List, Optional import faiss import torch from datasets import Features, Seque...
466
1
"""simple docstring""" from __future__ import annotations import math from collections.abc import Callable def _snake_case ( snake_case__ : Callable[[int | float], int | float] , snake_case__ : int | float , snake_case__ : int | float , snake_case__ : int = 100...
22
"""simple docstring""" import sys from collections import defaultdict class lowerCAmelCase_ : '''simple docstring''' def __init__( self : Optional[Any] ) -> int: A = [] def _SCREAMING_SNAKE_CASE ( self : Union[str, Any] ...
22
1
import baseaa def _lowerCamelCase( __snake_case ) -> bytes: return baseaa.aaaencode(string.encode("utf-8" ) ) def _lowerCamelCase( __snake_case ) -> str: return baseaa.aaadecode(__snake_case ).decode("utf-8" ) if __name__ == "__main__": import doctest docte...
524
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_xlnet import XLNetT...
524
1
def __lowerCamelCase ( A__ : int ) -> Union[str, Any]: if not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): raise ValueError("""multiplicative_persistence() only accepts integral values""" ) if num < 0: raise ValueError("""multiplicative_persistence() does not a...
710
from typing import Optional, Tuple import jax import jax.numpy as jnp from flax import linen as nn from flax.core.frozen_dict import FrozenDict from transformers import CLIPConfig, FlaxPreTrainedModel from transformers.models.clip.modeling_flax_clip import FlaxCLIPVisionModule def __lowerCamelCase ( ...
171
0
"""simple docstring""" def _snake_case ( __snake_case : int ): """simple docstring""" return str(__snake_case ) == str(__snake_case )[::-1] def _snake_case ( __snake_case : int ): """simple docstring""" ...
88
'''simple docstring''' from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import ScoreSdeVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class _snake_case (...
407
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available snake_case : str = { 'configuration_time_series_transformer': [ 'TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'TimeSeriesTransformerConfig', ], } ...
710
import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class lowerCAmelCase__ ( ctypes.Structure ): # _fields is a specific attr expected by ctypes __A : Optional[Any] = [('size', ...
182
0
from __future__ import annotations from typing import Any class snake_case_ : def __init__( self , __lowerCAmelCase = 6 ): SCREAMING_SNAKE_CASE_ : Node | None = None SCREAMING_SNAKE_CASE_ : Node | None = None self.create_linked_list(__Uppe...
345
import glob import os import random from string import ascii_lowercase, digits import cva __lowerCamelCase : Union[str, Any] = "" __lowerCamelCase : Dict = "" __lowerCamelCase : Optional[int] = "" __lowerCamelCase : Optional[A...
416
0
def __UpperCAmelCase ( __A , __A ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = [1] for i in range(2 , __A ): factorials.append(factorials[-1] * i ) assert 0 <= k < factorials[-1] * n, "k out of boun...
701
import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor A = logging.get_logger(__name__) class lowercase__ ( __SCREAMING_SNAKE_CASE ): def __init__( self : Union[str, Any] , *_lowercase : Any , *...
277
0
def a__ ( A__, A__ ): return int((input_a, input_a).count(1 ) != 0 ) def a__ ( ): assert or_gate(0, 0 ) == 0 assert or_gate(0, 1 ) == 1 assert or_gate(1, 0 ) == 1 assert or_gate(1, 1 ) == 1 if __name__ == "__...
101
"""simple docstring""" from collections.abc import Sequence def _lowerCAmelCase(a : Sequence[float] , a : float ) -> float: return sum(c * (x**i) for i, c in enumerate(a ) ) def _lowerCAmelCase(a : Sequence[float] , a : float ) -> ...
255
0
'''simple docstring''' def _snake_case ( lowercase ) -> int: __a : Any = [] for data in source_data: for i, el in enumerate(__lowerCAmelCase ): if len(__lowerCAmelCase ) < i + 1: data_lists.append([] ) ...
709
'''simple docstring''' import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand __SCREAMING_SNAKE_CASE : List[str] = ( '4S 3H 2C 7S 5H', '9D 8H 2C 6S 7H', '2D 6D 9D TH 7D', 'TC 8C 2S JH 6C', ...
697
0
'''simple docstring''' import warnings from contextlib import contextmanager from ...processing_utils import ProcessorMixin class SCREAMING_SNAKE_CASE ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] = '''Speech2TextFe...
517
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging _lowercase = logging.get_logger(__name__) _lowercase = { """edbeeching/decision-transformer-gym-hopper-medium""": ( """https://huggingface.co/edbeeching/decision-transfo...
5
0
def UpperCamelCase__ ( lowerCAmelCase__ ,lowerCAmelCase__ ): return "\n".join( f"""{number} * {i} = {number * i}""" for i in range(1 ,number_of_terms + 1 ) ) if __name__ == "__main__": print(multiplication_table(number=5, number_of_terms=10))
700
import collections.abc from typing import Optional, Tuple, Union import torch import torch.utils.checkpoint from torch import nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import BaseModelOutputWithNoAttention, ImageClassifierOutp...
72
0
"""simple docstring""" import os import unittest from huggingface_hub.utils import are_progress_bars_disabled import transformers.models.bart.tokenization_bart from transformers import logging from transformers.testing_utils import CaptureLogger, mockenv, mockenv_context from transformers.utils.loggi...
224
from sympy import diff, lambdify, symbols from sympy.functions import * # noqa: F403 def snake_case__ ( lowerCamelCase_ , lowerCamelCase_ , lowerCamelCase_ = "x" , lowerCamelCase_ = 10**-10 , lowerCamelCase_ = 1 , ): A : int = ...
542
0
from ..utils import DummyObject, requires_backends class lowerCamelCase (metaclass=__lowerCamelCase ): """simple docstring""" UpperCAmelCase_ = ["flax", "transformers"] def __init__( self : Tuple, *_UpperCAmelCase ...
157
import math from ...configuration_utils import PretrainedConfig from ...utils import logging _lowerCamelCase : Dict = logging.get_logger(__name__) _lowerCamelCase : Union[str, Any] = { '''facebook/data2vec-base-960h''': '''https://hugg...
157
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) __snake_case : int = { """configuration_llama""": ["""LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Llam...
540
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, StableDiffusionSAGPipeline, UNetaDConditionModel, ) from diffusers.utils import slow, torch_device from diffusers.util...
540
1
"""simple docstring""" import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def a__ ( *_SCREAMING_SNAKE_CASE ): """simple docstring""" if not isinstance(lowerCamelCase__ , lowerCamelCase__ ): UpperCamelCase = l...
715
"""simple docstring""" lowerCAmelCase__ = ''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' lowe...
544
0
"""simple docstring""" from graphs.minimum_spanning_tree_kruskal import kruskal def A ( ): """simple docstring""" snake_case_ :Optional[Any] = 9 snake_case_ :Optional[int] = [ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, ...
584
"""simple docstring""" class __lowerCAmelCase : '''simple docstring''' def __init__( self ): """simple docstring""" snake_case_ :Optional[Any] = 0 snake_case_ :Dict = 0 snake_case_ :Any = {} ...
584
1
def _lowerCamelCase ( _a ): """simple docstring""" _lowerCamelCase = 0 # if input_string is "aba" than new_input_string become "a|b|a" _lowerCamelCase = '''''' _lowerCamelCase = '''''' # append each character + "|" in new_string for range(0, length-1) fo...
719
def _lowerCamelCase ( _a ): """simple docstring""" if bit_count < 0: raise ValueError('''The given input must be positive''' ) # get the generated string sequence _lowerCamelCase = gray_code_sequence_string(_a ) # # convert them to integers for i in range(len(_a ...
297
0
'''simple docstring''' def __snake_case ( _UpperCAmelCase : list[int], _UpperCAmelCase : list[int]): UpperCamelCase = len(_UpperCAmelCase) print('''The following activities are selected:''') # The first activity is always selected UpperCamelCase = 0 pri...
212
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : Optional[Any] = logging.get_logger(__name__) snake_case_ : Any = { 'naver-clova-ix/donut-base': 'https://huggingface.co/naver-clova-ix/donut-base/resolve/main/conf...
212
1
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging a : str = logging.get_logger...
711
"""simple docstring""" import re from filelock import FileLock try: import nltk a : str = True except (ImportError, ModuleNotFoundError): a : List[str] = False if NLTK_AVAILABLE: with FileLock('''.lock''') as lock: nltk.download('''punkt''', quiet=True) def _UpperCamelCase ( ...
19
0
'''simple docstring''' from ..utils import DummyObject, requires_backends class A__ ( metaclass=__snake_case ): A__ = ['onnx'] def __init__( self : Union[str, Any] , *_a : Tuple , **_a : Tuple ) -> Any: '''simple d...
405
from math import pi, sqrt, tan def __lowerCAmelCase( _SCREAMING_SNAKE_CASE ) -> float: """simple docstring""" if side_length < 0: raise ValueError('surface_area_cube() only accepts non-negative values' ) return 6 * side_length**2 ...
27
0
from random import randint from tempfile import TemporaryFile import numpy as np def _snake_case( SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Tuple ) -> str: '''simple docstring''' ...
703
import argparse import json from pathlib import Path import torch import torchaudio from datasets import load_dataset from huggingface_hub import hf_hub_download from transformers import ASTConfig, ASTFeatureExtractor, ASTForAudioClassification from transformers.utils import logging logging.set_ver...
586
0
from __future__ import annotations def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> int: if len(__lowerCAmelCase ) < k or k < 0: raise ValueError('''Invalid Input''' ) snake_case__ = snake_case__ = sum(array[:k] ) for i in range...
33
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available a_ : int = { 'configuration_rag': ['RagConfig'], 'retrieval_rag': ['RagRetriever'], 'tokenization_rag': ['RagTokenizer'], } try: if not is_to...
73
0
"""simple docstring""" import functools def lowerCamelCase__ ( UpperCAmelCase_ , UpperCAmelCase_ )-> int: """simple docstring""" UpperCamelCase = len(UpperCAmelCase_ ) UpperCamelCase = len(UpperCAmelCase_ ) @functool...
718
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..auto.configuration_auto import CONFIG_MAPPING SCREAMING_SNAKE_CASE = logging.get_logger(__name__) class __a ( _lowerCAmelCase ): UpperCamelCase_ : ...
556
0
import gc import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, EulerAncestralDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPanoramaPipeline, ...
32
"""simple docstring""" from __future__ import annotations import math from collections.abc import Callable def __magic_name__ ( _lowerCamelCase : Callable[[int | float], int | float] , _lowerCamelCase : int | float , _lowerCamelCase : int | float , _l...
581
0
'''simple docstring''' from __future__ import annotations def UpperCAmelCase_ (__a : int ): """simple docstring""" _a : Optional[int] = 2 _a : str = [] while i * i <= n: if n % i: i += 1 else: n //= i fac...
319
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __lowerCAmelCase = { """configuration_squeezebert""": [ """SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Sque...
319
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices A = logging.get_logger(__name__) A = { '''shi-labs/dinat-min...
52
import math from collections.abc import Callable def _a ( UpperCAmelCase , UpperCAmelCase , UpperCAmelCase ) -> float: """simple docstring""" lowerCamelCase__ : float = xa lowerCamelCase__ : float = xa while True: if x_n == x_na or ...
315
0
'''simple docstring''' from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def _lowercase ( ) -> tuple[list[int], int]: """simple docstring""" __UpperCAmelCase : str = [ran...
10
'''simple docstring''' import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel _a : Union[str, Any] = HfApi() _a : int = {} # fmt: off _a : Optional[int] = torch.tensor([ -0.7_515, -1.6_883, 0.2_420, 0.0_300, 0.6_347,...
10
1
from ...configuration_utils import PretrainedConfig from ...utils import logging __a = logging.get_logger(__name__) __a = { 'facebook/xglm-564M': 'https://huggingface.co/facebook/xglm-564M/resolve/main/config.json', # See all XGLM models at https://huggingface.co/models?filter=x...
97
from __future__ import annotations from math import pi, sqrt def a ( snake_case__: float , snake_case__: float ): '''simple docstring''' if inductance <= 0: raise ValueError('''Inductance cannot be 0 or negative''' ) elif capacitance <= 0: rais...
97
1
'''simple docstring''' import copy import tempfile import unittest from huggingface_hub import HfFolder, delete_repo from parameterized import parameterized from requests.exceptions import HTTPError from transformers import AutoConfig, GenerationConfig from transformers.testing_utils import TOKEN, USER, is_s...
701
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _A: List[str] = { """configuration_luke""": ["""LUKE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """LukeConfig"""], """tokenization_luke""": ["""LukeTokeni...
617
0
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image i...
86
"""simple docstring""" def SCREAMING_SNAKE_CASE ( lowercase__ , lowercase__ ) -> float: if digit_amount > 0: return round(number - int(lowercase__ ) , lowercase__ ) return number - int(lowercase__ ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) pr...
453
0
'''simple docstring''' import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_avai...
566
'''simple docstring''' import logging import numpy as np import pytest from scipy.linalg import eigh logging.basicConfig(level=logging.INFO, format='''%(message)s''') def snake_case__ ( a ) -> np.ndarray: '''simple docstring''' return input_array.reshape((input_array.size,...
566
1
import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if...
542
import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all image processors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from...
542
1
'''simple docstring''' from .glue import GlueDataset, GlueDataTrainingArguments from .language_modeling import ( LineByLineTextDataset, LineByLineWithRefDataset, LineByLineWithSOPTextDataset, TextDataset, TextDatasetForNextSentencePrediction, ) from .squad import SquadDataset, SquadDataTraini...
454
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # ...
454
1
from typing import Dict, List, Optional, Tuple, Union import torch from ...models import AutoencoderKL, TransformeraDModel from ...schedulers import KarrasDiffusionSchedulers from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class lowerCamelCase ( ...
201
import torch from diffusers import DDPMParallelScheduler from .test_schedulers import SchedulerCommonTest class lowerCamelCase ( __lowerCamelCase ): UpperCamelCase_ : int = (DDPMParallelScheduler,) def snake_case__ ( self :Any , **lowercase :str )...
201
1
"""simple docstring""" from __future__ import annotations from cmath import sqrt def __a ( _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) ->tuple[complex, complex]: if a == 0: raise ValueError('Coefficient \'a\' must not be zero.' ) a__: Optional[Any] ...
217
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase__ = logging.get_logger(__name__) lowercase__ = { 'facebook/s2t-small-librispeech-asr': ( 'https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/config.json' ...
217
1
# limitations under the License. # NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from .pipelines import DiffusionPipeline, ImagePipelineOutput # noqa: F401 from .utils import deprecate d...
403
def __lowerCamelCase (UpperCAmelCase__ : list[int] , UpperCAmelCase__ : list[int] , UpperCAmelCase__ : int ): return not any( neighbour == 1 and colored_vertices[i] == color for i, neighbour in enumerate(UpperCAmelCase__ ) ) def __lowerCamelCase ...
403
1
import argparse import torch from safetensors.torch import load_file from diffusers import StableDiffusionPipeline def lowerCAmelCase_ ( lowercase: Optional[int] , lowercase: Any , lowercase: str , lowercase: List[str] , lowercase: Optional[int] ) -...
264
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCAmelCase_ = {'''configuration_xlnet''': ['''XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP'''...
264
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _snake_case : Optional[int] = { "configuration_poolformer": [ "POOLFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "PoolFormerConfig", "P...
81
import json import os import unittest from transformers import BatchEncoding, MvpTokenizer, MvpTokenizerFast from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers, require_torch from transformers.utils import cached_property from .....
551
0
def lowerCAmelCase_( lowercase_ : Optional[Any] = 1_00_00_00 ) -> str: _lowerCamelCase = limit + 1 _lowerCamelCase = [0] * limit for first_term in range(1 , snake_case_ ): for n in range(snake_case_ , snake_case_ , sna...
710
"""simple docstring""" def lowerCAmelCase_( lowercase_ : str , lowercase_ : str ) -> bool: _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = len(lowercase_ ) _lowerCamelCase = [[False for _ in range(m + 1 ...
623
0
from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case : List[Any] = logging.get_logger(__name__) snake_case : str = { 'google/switch-base-8': 'https://huggingface.co/google/switch-base-8/blob/main/config.json', } class __lowerc...
605
'''simple docstring''' import math import flax.linen as nn import jax.numpy as jnp def _lowerCAmelCase ( __magic_name__ : jnp.ndarray , __magic_name__ : int , __magic_name__ : float = 1 , __magic_name__ : float = 1 , __magic_name__ : float = 1.0...
92
0
import argparse import torch from transformers import MobileBertConfig, MobileBertForPreTraining, load_tf_weights_in_mobilebert from transformers.utils import logging logging.set_verbosity_info() def A_ ( lowercase_ , lowercase_ , lowercase_ ) ->List[str]: """simple docstring...
702
import tempfile import unittest from transformers import AutoModelForSeqaSeqLM, AutoTokenizer from transformers.testing_utils import ( is_torch_available, require_optimum, require_torch, slow, ) if is_torch_available(): import torch @require_torch @require_optimum @slow class a_( unit...
259
0
'''simple docstring''' import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.dat...
98
'''simple docstring''' import inspect from typing import Callable, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import DiffusionPipeline from diffusers.models import AutoencoderKL, UNetaDConditionModel from diffusers.pipelines.stabl...
98
1
import argparse import collections import numpy as np import torch from flax import traverse_util from tax import checkpoints from transformers import MTaConfig, UMTaEncoderModel, UMTaForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() def _a ( SC...
429
import math def _a ( SCREAMING_SNAKE_CASE = 1_00 ): """simple docstring""" lowercase__ = sum(i * i for i in range(1 , n + 1 ) ) lowercase__ = int(math.pow(sum(range(1 , n + 1 ) ) , 2 ) ) return square_of_sum - sum_of_squares if __name__ ==...
429
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { '''roberta-base''': '''https://huggingface....
154
def lowerCAmelCase ( UpperCAmelCase ) ->bool: """simple docstring""" if p < 2: raise ValueError('''p should not be less than 2!''' ) elif p == 2: return True __magic_name__ : Tuple = 4 __magic_name...
154
1
'''simple docstring''' from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": snake_case_ = input("""Enter image url: """).strip() print(f'''Downloading image from {url} ...''') snake_case_ = BeautifulSoup(requests.get(url).content, ""...
537
'''simple docstring''' def _lowerCamelCase( UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : str ) -> Optional[Any]: A : Tuple = 0 while b > 0: if b & 1: res += a a += a b >>= 1 return res def _l...
537
1
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _A ...
100
import os def __A ( ) -> Dict: with open(os.path.dirname(__lowerCamelCase ) + """/p022_names.txt""" ) as file: a = str(file.readlines()[0] ) a = names.replace("""\"""" , """""" ).split(""",""" ) names.sort() ...
468
0
import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import datasets import datasets.config from .utils import require_beam class UpperCAmelCase_ ( datasets.BeamBasedBuilder ): '''simple docstring''' def _lo...
114
import logging import os from dataclasses import dataclass from typing import List, Optional, Union import tqdm from filelock import FileLock from transformers import ( BartTokenizer, BartTokenizerFast, DataProcessor, PreTrainedTokenizer, RobertaTokenizer, RobertaTokenizerFast, XLMR...
114
1
import math class __SCREAMING_SNAKE_CASE : def __init__( self : Optional[Any] , A : Any=0 ) ->Optional[Any]: # a graph with Node 0,1,...,N-1 lowerCamelCase__ : Dict = n lowerCamelCase__ : List[Any] = [ [math.inf for j...
315
import unittest from diffusers.models.unet_ad_blocks import * # noqa F403 from diffusers.utils import torch_device from .test_unet_blocks_common import UNetBlockTesterMixin class __SCREAMING_SNAKE_CASE ( lowerCAmelCase_ ,unittest.TestCase ): _UpperCAmelCase : Dict = Dow...
315
1
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('''>=''', '''4.25.0''')): raise OptionalDependencyNotAvailable() exce...
313
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by app...
313
1