python_code stringlengths 0 679k | repo_name stringlengths 9 41 | file_path stringlengths 6 149 |
|---|---|---|
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Test TE Paddle Recompute"""
from pathlib import Path
import re
import subprocess
import numpy as np
import pytest
from transformer_engine.paddle.fp8 import is_fp8_available
test_root = Path(_... | TransformerEngine-main | tests/paddle/test_recompute.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Test TE Paddle Parallel"""
from pathlib import Path
import unittest
from dist_launcher import TestDistributed
from utils import is_devices_enough
from transformer_engine.paddle.fp8 import is_f... | TransformerEngine-main | tests/paddle/test_parallel.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Utils for testing"""
import random
import numpy as np
import paddle
from paddle.distributed import fleet
from paddle.distributed.fleet.meta_parallel import get_rng_state_tracker
import transfo... | TransformerEngine-main | tests/paddle/utils.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Test basic installation of Paddle extensions"""
def test_import():
"""
Test if Paddle extension can be imported normally
"""
import transformer_engine.paddle # pylint: disabl... | TransformerEngine-main | tests/paddle/test_install.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Helper functions to launch distributed tests"""
import copy
import os
from pathlib import Path
import subprocess
import time
import unittest
from paddle import fluid
from paddle.distributed.uti... | TransformerEngine-main | tests/paddle/dist_launcher.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Test TransformerLayer encoder recompute"""
import sys
import paddle
import transformer_engine.paddle as te
class Net(paddle.nn.Layer):
"""Network use for recompute testing"""
def __in... | TransformerEngine-main | tests/paddle/recompute_tests/recompute_transformer_encoder.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Unittest for LayerNormMLP layer in tensor parallel"""
import unittest
import paddle
from paddle.distributed import fleet
from utils import assert_allclose, assert_shape, set_random_seed
import... | TransformerEngine-main | tests/paddle/parallel_tests/layernorm_mlp_tp.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Unittest for Transformer layer in tensor parallel"""
import unittest
import paddle
from paddle.distributed import fleet
from utils import assert_allclose, set_random_seed
import transformer_en... | TransformerEngine-main | tests/paddle/parallel_tests/transformer_tp.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Unittest for Linear layer in pipeline parallel"""
import unittest
import numpy as np
import paddle
from paddle.distributed import fleet
from paddle.distributed.fleet.meta_parallel import (
... | TransformerEngine-main | tests/paddle/parallel_tests/linear_pp.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Unittest for Linear layer in tensor parallel"""
import unittest
import paddle
from paddle.distributed import fleet
from paddle.distributed.fleet.layers.mpu import mp_ops
from utils import asse... | TransformerEngine-main | tests/paddle/parallel_tests/linear_tp.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Unittest for LayerNormLinear layer in tensor parallel"""
import unittest
import paddle
from paddle.distributed import fleet
from paddle.distributed.fleet.layers.mpu import mp_ops
from utils im... | TransformerEngine-main | tests/paddle/parallel_tests/layernorm_linear_tp.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Unittest for group sharding"""
import unittest
import paddle
from paddle.distributed import fleet
from paddle.distributed.fleet.meta_optimizers.dygraph_optimizer import (
DygraphShardingOpt... | TransformerEngine-main | tests/paddle/parallel_tests/group_sharding.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Unittest for Linear layer in tensor parallel"""
import unittest
import paddle
from paddle.distributed import fleet
from utils import assert_allclose, set_random_seed
import transformer_engine.... | TransformerEngine-main | tests/paddle/parallel_tests/amax_reduction.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Tests for the cpp extensions."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import numpy as np
import transformer_engine # pyli... | TransformerEngine-main | tests/tensorflow/test_extensions.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Tests for the fp8 layers."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import tensorflow as tf
import transformer_engine.tenso... | TransformerEngine-main | tests/tensorflow/test_layers.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Tests for the MHA layer."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import tensorflow as tf
import transformer_engine.tensor... | TransformerEngine-main | tests/tensorflow/test_mha.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import transformer_engine.tensorflow
print("OK")
| TransformerEngine-main | tests/tensorflow/test_sanity_import.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Tests for the Transformer layer."""
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
import tensorflow as tf
import transformer_engin... | TransformerEngine-main | tests/tensorflow/test_transformer.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""
This file contains tests for exporting TransformerEngine models to ONNX.
The purpose of these tests is validation that TE models are converted to their correct ONNX
representation. Toward this... | TransformerEngine-main | tests/pytorch/test_onnx_export.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""
This file contains tests for saving and loading TransformerEngine torch checkpoints.
The purpose of this test is to validate the TransformerEngine hooks for saving FP8 metadata
in torch checkp... | TransformerEngine-main | tests/pytorch/test_torch_save_load.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
from typing import Tuple
import pytest
import torch
import transformer_engine.pytorch as te
# Model names for test_torch_dynamo
_model_names = ["Linear", "LayerNorm", "LayerNormLinear", "LayerNo... | TransformerEngine-main | tests/pytorch/test_jit.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import torch
import pytest
from transformer_engine.pytorch.fp8 import fp8_autocast, FP8GlobalStateManager
from transformer_engine.pytorch.utils import (
init_method_normal,
scaled_init_met... | TransformerEngine-main | tests/pytorch/test_sanity.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import torch
import pytest
from transformer_engine.pytorch.utils import (
init_method_normal,
scaled_init_method_normal,
get_device_compute_capability,
)
from transformer_engine.pytorc... | TransformerEngine-main | tests/pytorch/test_fused_attn.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import math
import os
import contextlib
from typing import List, Optional
import pytest
import copy
import torch
import torch.nn as nn
from torch.nn import Parameter
from torch import _C
from torc... | TransformerEngine-main | tests/pytorch/test_numerics.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import transformer_engine.pytorch
print("OK")
| TransformerEngine-main | tests/pytorch/test_sanity_import.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import os
import re
import glob
import datetime
from prettytable import PrettyTable
from matplotlib import pyplot as plt
NUM_MOST_RECENT_RUNS = 100
te_path = os.getenv("TE_PATH", "/opt/transform... | TransformerEngine-main | tests/pytorch/distributed/print_logs.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
from typing import List, Tuple, Union
import pytest
import subprocess
import os
from dataclasses import dataclass, asdict
from functools import lru_cache
import torch
@dataclass()
class ModelCon... | TransformerEngine-main | tests/pytorch/distributed/test_convergence.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import unittest
import flax
import jax
import jax.numpy as jnp
import numpy as np
from utils import assert_allclose
from transformer_engine.common.recipe import DelayedScaling
from transformer_en... | TransformerEngine-main | tests/jax/test_helper.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
from functools import partial
import flax
import jax
import jax.numpy as jnp
import pytest
from transformer_engine.common.recipe import Format
from transformer_engine.jax.flax import TransformerL... | TransformerEngine-main | tests/jax/test_layer.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import jax
import numpy as np
import pytest
from utils import is_devices_enough
from transformer_engine.jax.flax import extend_logical_axis_rules
from transformer_engine.jax.sharding import get_do... | TransformerEngine-main | tests/jax/test_sharding.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
from functools import partial
from typing import Dict
import flax
import jax
import jax.numpy as jnp
from praxis import pax_fiddle
from praxis.base_layer import WeightInit, DEFAULT_INIT_MUTABLE_LI... | TransformerEngine-main | tests/jax/test_praxis_layers.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import pytest
import jax.numpy as jnp
from jax.core import ShapedArray
from transformer_engine_jax import DType
from transformer_engine.jax.cpp_extensions import te_dtype_to_jax_dtype
from transfo... | TransformerEngine-main | tests/jax/test_custom_call_shape.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import functools
import operator
from typing import Any, Callable, Tuple, Sequence, Union, Iterable, Optional
import jax
import jax.numpy as jnp
import numpy as np
from flax import linen as nn
fro... | TransformerEngine-main | tests/jax/utils.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import functools
import operator
import jax
import jax.numpy as jnp
import numpy as np
import pytest
from jax import lax
from jax import jit, value_and_grad
from flax import linen as nn
from util... | TransformerEngine-main | tests/jax/test_custom_call_compute.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Tests for fused attention"""
import os
from enum import Enum
from math import sqrt
import jax
import jax.numpy as jnp
import numpy as np
import pytest
from flax.linen import combine_masks
from... | TransformerEngine-main | tests/jax/test_fused_attn.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import transformer_engine.jax
print("OK")
| TransformerEngine-main | tests/jax/test_sanity_import.py |
#!/usr/bin/env python
# coding: utf-8
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import os
import sys
import json
import datetime
if len(sys.argv) < 2:
print("Usage: python copyright_checker.py <path>")
path = sys.argv[1]
config_pa... | TransformerEngine-main | qa/L0_license/copyright_checker.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import os
import sys
import sphinx_rtd_theme
from sphinx.ext.autodoc.mock import mock
from sphinx.ext.autodoc import between, ClassDocumenter, AttributeDocumenter
from sphinx.util import inspect
fr... | TransformerEngine-main | docs/conf.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import math
from typing import Callable, Optional
import torch
import transformer_engine.pytorch as te
from transformer_engine.pytorch.fp8 import DelayedScaling, dist_group_type
def speedometer(
... | TransformerEngine-main | docs/examples/quickstart_utils.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""MNIST example of Transformer Engine Paddle"""
import argparse
import os
import unittest
import paddle
from paddle import nn
import paddle.nn.functional as F
from paddle.vision.transforms impor... | TransformerEngine-main | examples/paddle/mnist/test_single_gpu_mnist.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
from transformer_engine.tensorflow import Format, DelayedScaling
import argparse
import tensorflow as tf
import time
import transformer_engine.tensorflow as te
from keras import layers
from keras ... | TransformerEngine-main | examples/tensorflow/transformer_layer.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import argparse
import tensorflow as tf
import tensorflow_datasets as tfds
import transformer_engine.tensorflow as te
class MNIST(tf.keras.Model):
def __init__(self, use_te=False):
su... | TransformerEngine-main | examples/tensorflow/mnist.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
import argparse
import torch
import torch.nn as nn
import torch.nn.functional as F
import torch.optim as optim
from torchvision import datasets, transforms
from torch.optim.lr_scheduler import Step... | TransformerEngine-main | examples/pytorch/mnist/main.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
""" MNIST training on single GPU"""
import argparse
import unittest
from functools import partial
import jax
import jax.numpy as jnp
import numpy as np
import optax
from datasets import load_datase... | TransformerEngine-main | examples/jax/mnist/test_single_gpu_mnist.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Encoder training on multi-GPU with tesnor parallelism"""
import argparse
import unittest
from functools import partial
import flax
import jax
import jax.numpy as jnp
import nltk
import numpy as ... | TransformerEngine-main | examples/jax/encoder/test_model_parallel_encoder.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Encoder training with multi-GPU, multiprocessing, and tensor parallelism"""
import argparse
import multiprocessing as mp
import os
import unittest
from functools import partial
import flax
impor... | TransformerEngine-main | examples/jax/encoder/test_multiprocessing_encoder.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Encoder training on single GPU"""
import argparse
import unittest
from functools import partial
import flax
import jax
import jax.numpy as jnp
import nltk
import numpy as np
import optax
from da... | TransformerEngine-main | examples/jax/encoder/test_single_gpu_encoder.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Encoder training on multi-GPU with data parallelism"""
import argparse
import unittest
from functools import partial
import flax
import jax
import jax.numpy as jnp
import nltk
import numpy as np... | TransformerEngine-main | examples/jax/encoder/test_multigpu_encoder.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Top level package"""
from . import common
try:
from . import pytorch
except ImportError as e:
pass
try:
from . import jax
except ImportError as e:
pass
try:
from . import ... | TransformerEngine-main | transformer_engine/__init__.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""FP8 utilities for TransformerEngine"""
from contextlib import contextmanager
from typing import Tuple, Optional, Dict, Any, Union
import numpy as np
import paddle
import transformer_engine_pad... | TransformerEngine-main | transformer_engine/paddle/fp8.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Utils for profiling"""
from contextlib import contextmanager
from paddle.fluid import core
@contextmanager
def nvtx_range(msg):
"""Context to insert NVTX"""
core.nvprof_nvtx_push(msg)... | TransformerEngine-main | transformer_engine/paddle/profile.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Constants"""
from enum import Enum
import paddle
import transformer_engine_paddle as tex
class FP8FwdTensors(Enum):
"""Used as named indices on the `scale`, `scale_inv`,
and `amax` t... | TransformerEngine-main | transformer_engine/paddle/constants.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Transformer Engine bindings for Paddle"""
from .fp8 import fp8_autocast
from .layer import (Linear, LayerNorm, LayerNormLinear, LayerNormMLP, FusedScaleMaskSoftmax,
DotProduc... | TransformerEngine-main | transformer_engine/paddle/__init__.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Methods needed for distributed training."""
from contextlib import contextmanager
from typing import Optional, Union, Tuple
import paddle
import paddle.distributed.fleet.base.topology as tp
fr... | TransformerEngine-main | transformer_engine/paddle/distributed.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Utility functions for Transformer Engine modules"""
from typing import Optional, Tuple, Union
import paddle
import paddle.nn.functional as F
def cast_if_needed(tensor: Union[paddle.Tensor, No... | TransformerEngine-main | transformer_engine/paddle/utils.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""TE FP8 extensions and GEMMs"""
import math
from typing import Optional, Tuple, Union
import paddle
import transformer_engine_paddle as tex
from .constants import TE_DType, FP8FwdTensors, FP8BwdT... | TransformerEngine-main | transformer_engine/paddle/cpp_extensions.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""FP8 meta buffer for FP8 amax reduction"""
from abc import ABC, abstractmethod
from collections import deque
from functools import partial
import os
from typing import Dict, Any, List, Union
imp... | TransformerEngine-main | transformer_engine/paddle/fp8_buffer.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Methods needed for recompute."""
import os
import inspect
from paddle.distributed import fleet
from .constants import RecomputeFunctionNames
from .fp8 import get_global_fp8_state
__all__ = ['... | TransformerEngine-main | transformer_engine/paddle/recompute.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Attntion API"""
import math
import os
import warnings
from typing import Optional, Tuple, Union
import paddle
import paddle.nn.functional as F
from .layernorm_linear import LayerNormLinear
fro... | TransformerEngine-main | transformer_engine/paddle/layer/attention.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""LayerNormMLP API"""
import os
from typing import Union, Tuple, Dict, Any, Optional
import paddle
import paddle.nn.functional as F
from paddle.nn.initializer import Constant
from .base import T... | TransformerEngine-main | transformer_engine/paddle/layer/layernorm_mlp.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Linear API"""
import os
from typing import Union, Tuple
import paddle
import paddle.nn.functional as F
from paddle.nn.initializer import Constant
from ..constants import TE_DType
from ..cpp_ex... | TransformerEngine-main | transformer_engine/paddle/layer/layernorm.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Linear API"""
from typing import Union, Tuple, Dict, Any, Optional
import paddle
import paddle.nn.functional as F
from paddle.nn.initializer import Constant
from .base import (
Transformer... | TransformerEngine-main | transformer_engine/paddle/layer/linear.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Layer level Paddle APIs"""
from .attention import DotProductAttention, MultiHeadAttention
from .layernorm import LayerNorm
from .layernorm_linear import LayerNormLinear
from .layernorm_mlp impor... | TransformerEngine-main | transformer_engine/paddle/layer/__init__.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""LayerNormLinear API"""
import os
from typing import Union, Tuple, Dict, Any, Optional
import paddle
import paddle.nn.functional as F
from paddle.nn.initializer import Constant
from ..cpp_exten... | TransformerEngine-main | transformer_engine/paddle/layer/layernorm_linear.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Transformer"""
from typing import Optional, Union
import paddle
from . import LayerNormMLP, LayerNorm, MultiHeadAttention
from ..constants import AttnMaskTypes, LayerTypes, dist_group_type
fro... | TransformerEngine-main | transformer_engine/paddle/layer/transformer.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Fused scaled masked softmax functions"""
import os
import warnings
from typing import Callable, Tuple, Union, Optional
import paddle
from transformer_engine.paddle.cpp_extensions import (
... | TransformerEngine-main | transformer_engine/paddle/layer/softmax.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Base modules and utilities for TransformerEngine Paddle API"""
from abc import ABC, abstractmethod
from contextlib import contextmanager
import os
import pickle
from typing import Generator, Dic... | TransformerEngine-main | transformer_engine/paddle/layer/base.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""FW agnostic user-end APIs"""
import ctypes
import os
import platform
import subprocess
import sys
def get_te_path():
"""Find Transformer Engine install path using pip"""
command = [sy... | TransformerEngine-main | transformer_engine/common/__init__.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""The utilities for Transformer Engine"""
import inspect
import warnings
from enum import Enum
warnings.simplefilter('default')
class DeprecatedEnum: # pylint: disable=too-few-public-methods
... | TransformerEngine-main | transformer_engine/common/utils.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""This module provides predefined FP8 recipes."""
from __future__ import annotations
from enum import Enum
from typing import Literal, Optional, Union, Callable, NamedTuple
from pydantic.dataclass... | TransformerEngine-main | transformer_engine/common/recipe.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""FP8 utilies for TransformerEngine"""
from contextlib import contextmanager
from typing import Generator, Optional, Dict, Any
import tensorflow as tf
import transformer_engine_tensorflow as tex
... | TransformerEngine-main | transformer_engine/tensorflow/fp8.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""XLA functions and JIT utilities"""
from typing import Callable
import tensorflow as tf
@tf.function(jit_compile=True)
def _bgrad_dgelu_fused(grad_output, inp):
"""Bgrad-Dgelu fused"""
... | TransformerEngine-main | transformer_engine/tensorflow/jit.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Enums for e2e transformer"""
import tensorflow as tf
import transformer_engine_tensorflow as tex
"""
This is a map: tf.dtype -> int
Used for passing dtypes into cuda
extension. Has one to one ... | TransformerEngine-main | transformer_engine/tensorflow/constants.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Transformer Engine bindings for Tensorflow"""
from transformer_engine.common.recipe import DelayedScaling
from transformer_engine.common.recipe import Format
from .constants import TE_DType
fro... | TransformerEngine-main | transformer_engine/tensorflow/__init__.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Utility functions for Transformer Engine modules"""
import tensorflow as tf
def attention_mask_func(
attention_scores: tf.Tensor, attention_mask: tf.Tensor
) -> tf.Tensor:
"""Get atten... | TransformerEngine-main | transformer_engine/tensorflow/utils.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Transformer."""
from contextlib import nullcontext
from typing import Callable, Optional, Tuple, Union
import os
from keras import backend, layers, initializers
import tensorflow as tf
from t... | TransformerEngine-main | transformer_engine/tensorflow/transformer.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Top level Transformer Engine PyTorch modules"""
from typing import Union, Callable
from keras import backend, layers, initializers
import tensorflow as tf
import transformer_engine_tensorflow a... | TransformerEngine-main | transformer_engine/tensorflow/module.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Fused scaled masked softmax functions"""
from typing import Callable
import os
import transformer_engine_tensorflow as tex
import tensorflow as tf
from .module import get_stream_id
THREADS_P... | TransformerEngine-main | transformer_engine/tensorflow/softmax.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Attention."""
import os
import warnings
import math
from importlib.metadata import version
from contextlib import nullcontext
from typing import Any, Callable, Optional, Tuple, Union, Dict
from ... | TransformerEngine-main | transformer_engine/pytorch/attention.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Utilities for debugging numerical issues with FP8"""
from typing import Tuple
import torch
from transformer_engine.common import recipe
_NUMERICS_DEBUG = False
def debug(enabled: bool = True)... | TransformerEngine-main | transformer_engine/pytorch/numerics_debug.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""FP8 utilities for TransformerEngine"""
import os
from contextlib import contextmanager
from collections import deque
from typing import Callable, List, Optional, Dict, Any, Tuple, Union
import ... | TransformerEngine-main | transformer_engine/pytorch/fp8.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""NVFuser functions and JIT utilities"""
import os
from typing import Callable, Optional, Tuple
import torch
jit_fuser = torch.jit.script
if torch.__version__ >= "2" and bool(int(os.getenv("NVTE... | TransformerEngine-main | transformer_engine/pytorch/jit.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Enums for e2e transformer"""
import torch
import torch.distributed
import transformer_engine_extensions as tex
"""
This is a map: torch.dtype -> int
Used for passing dtypes into cuda
extension... | TransformerEngine-main | transformer_engine/pytorch/constants.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Transformer Engine bindings for pyTorch"""
from .module import LayerNormLinear
from .module import Linear
from .module import LayerNormMLP
from .module import LayerNorm
from .module import RMSNo... | TransformerEngine-main | transformer_engine/pytorch/__init__.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Export utilities for TransformerEngine"""
from contextlib import contextmanager
_IN_ONNX_EXPORT_MODE = False
@contextmanager
def onnx_export(
enabled: bool = False,
) -> None:
"""
... | TransformerEngine-main | transformer_engine/pytorch/export.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""
ONNX symbolic functions for Transformer Engine
Warnings of the type pasted below are a known Pytorch issue
(https://github.com/pytorch/pytorch/issues/81693):
tests/test_onnx_export.py::test_e... | TransformerEngine-main | transformer_engine/pytorch/te_onnx_extensions.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Methods needed for distributed training (DP/TP)."""
from contextlib import contextmanager
from typing import Any, Dict, Union, Optional, Callable, Tuple
import torch
from torch.cuda import _laz... | TransformerEngine-main | transformer_engine/pytorch/distributed.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Utility functions for Transformer Engine modules"""
import math
from typing import Any, Callable, Optional, Tuple
import torch
def get_device_compute_capability() -> float:
"""Returns the ... | TransformerEngine-main | transformer_engine/pytorch/utils.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Transformer."""
import os
import warnings
from contextlib import nullcontext
from typing import Any, Callable, Optional, Tuple, Union
import torch
import transformer_engine_extensions as tex
f... | TransformerEngine-main | transformer_engine/pytorch/transformer.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Fused scaled masked softmax functions"""
import os
from typing import Callable, Tuple, Union, Optional
import torch
from torch import nn
import torch._C._onnx as _C_onnx
from torch.onnx import _... | TransformerEngine-main | transformer_engine/pytorch/softmax.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Python interface for cast extensions"""
from typing import Optional, Union
import torch
import transformer_engine_extensions as tex
__all__ = ['cast_to_fp8',
'cast_from_fp8']
def ... | TransformerEngine-main | transformer_engine/pytorch/cpp_extensions/cast.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Python interface for GEMM extensions"""
from typing import Optional, Tuple, Union
import torch
import transformer_engine_extensions as tex
from ..constants import TE_DType
from ..utils import as... | TransformerEngine-main | transformer_engine/pytorch/cpp_extensions/gemm.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Python interface for c++ extensions"""
from transformer_engine_extensions import *
from .fused_attn import *
from .gemm import *
from .transpose import *
from .activation import *
from .normali... | TransformerEngine-main | transformer_engine/pytorch/cpp_extensions/__init__.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Python interface for fused attention extensions"""
import math
from typing import Tuple, List, Union
import torch
import transformer_engine_extensions as tex
from transformer_engine_extensions i... | TransformerEngine-main | transformer_engine/pytorch/cpp_extensions/fused_attn.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Python interface for activation extensions"""
from typing import Union
import torch
import transformer_engine_extensions as tex
__all__ = ['gelu', 'relu', 'reglu', 'geglu', 'swiglu']
def gel... | TransformerEngine-main | transformer_engine/pytorch/cpp_extensions/activation.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Python interface for transpose extensions"""
from typing import Optional, Tuple, Union
import torch
import transformer_engine_extensions as tex
from ..constants import TE_DType
__all__ = ['fp8... | TransformerEngine-main | transformer_engine/pytorch/cpp_extensions/transpose.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""Python interface for normalization extensions"""
from typing import Optional, Tuple, Union
import torch
import transformer_engine_extensions as tex
__all__ = ['layernorm_fwd_fp8',
'... | TransformerEngine-main | transformer_engine/pytorch/cpp_extensions/normalization.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""RMSNorm API"""
import os
from typing import Union, Tuple, Optional
import torch
from torch.nn.parameter import Parameter
from torch.nn import init
from .base import TransformerEngineBaseModule... | TransformerEngine-main | transformer_engine/pytorch/module/rmsnorm.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""LayerNormMLP API"""
import os
from typing import Union, Optional, Callable, Tuple, List, Dict, Any
import torch
from torch.nn.parameter import Parameter
from torch.nn import init
from .base im... | TransformerEngine-main | transformer_engine/pytorch/module/layernorm_mlp.py |
# Copyright (c) 2022-2023, NVIDIA CORPORATION & AFFILIATES. All rights reserved.
#
# See LICENSE for license information.
"""LayerNorm API"""
import os
from typing import Union, Tuple, Any, Mapping, Optional
import torch
from torch.nn.parameter import Parameter
from torch.nn import init
import transformer_engine_ext... | TransformerEngine-main | transformer_engine/pytorch/module/layernorm.py |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.