python_code
stringlengths
0
992k
repo_name
stringlengths
8
46
file_path
stringlengths
5
162
from .discriminative_reranking_model import DiscriminativeNMTReranker __all__ = [ "DiscriminativeNMTReranker", ]
EXA-1-master
exa/models/unilm-master/edgelm/examples/discriminative_reranking_nmt/models/__init__.py
from dataclasses import dataclass, field import os import torch import torch.nn as nn from fairseq import utils from fairseq.dataclass import ChoiceEnum, FairseqDataclass from fairseq.models import ( BaseFairseqModel, register_model, ) from fairseq.models.roberta.model import RobertaClassificationHead from ...
EXA-1-master
exa/models/unilm-master/edgelm/examples/discriminative_reranking_nmt/models/discriminative_reranking_model.py
#!/usr/bin/env python import argparse from multiprocessing import Pool from pathlib import Path import sacrebleu import sentencepiece as spm def read_text_file(filename): with open(filename, "r") as f: output = [line.strip() for line in f] return output def get_bleu(in_sent, target_sent): ble...
EXA-1-master
exa/models/unilm-master/edgelm/examples/discriminative_reranking_nmt/scripts/prep_data.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import math from dataclasses import dataclass, field import torch import torch.nn.functional as F from fairseq import metrics, utils from fa...
EXA-1-master
exa/models/unilm-master/edgelm/examples/discriminative_reranking_nmt/criterions/discriminative_reranking_criterion.py
from .discriminative_reranking_criterion import KLDivergenceRerankingCriterion __all__ = [ "KLDivergenceRerankingCriterion", ]
EXA-1-master
exa/models/unilm-master/edgelm/examples/discriminative_reranking_nmt/criterions/__init__.py
#!/usr/bin/env python3 -u # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. """ Translate pre-processed data with a trained model. """ import numpy as np import torch from fairseq import check...
EXA-1-master
exa/models/unilm-master/edgelm/examples/criss/save_encoder.py
#!/usr/bin/env python3 -u # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import argparse import glob from subprocess import check_call try: import faiss has_faiss = True except Imp...
EXA-1-master
exa/models/unilm-master/edgelm/examples/criss/mining/mine.py
#!/usr/bin/env python3 -u # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import argparse import glob import numpy as np DIM = 1024 def compute_dist(source_embs, target_embs, k=5, return...
EXA-1-master
exa/models/unilm-master/edgelm/examples/criss/sentence_retrieval/encoder_analysis.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import torch from fairseq.search import Search class NoisyChannelBeamSearch(Search): def __init__(self, tgt_dict): super().__in...
EXA-1-master
exa/models/unilm-master/edgelm/examples/fast_noisy_channel/noisy_channel_beam_search.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. from . import noisy_channel_translation # noqa from . import noisy_channel_sequence_generator # noqa from . import noisy_channel_beam_search...
EXA-1-master
exa/models/unilm-master/edgelm/examples/fast_noisy_channel/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. from typing import Dict, List, Optional import math import numpy as np import torch import torch.nn.functional as F from torch import Tensor...
EXA-1-master
exa/models/unilm-master/edgelm/examples/fast_noisy_channel/noisy_channel_sequence_generator.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. from fairseq.tasks.translation import TranslationTask from fairseq.tasks.language_modeling import LanguageModelingTask from fairseq import che...
EXA-1-master
exa/models/unilm-master/edgelm/examples/fast_noisy_channel/noisy_channel_translation.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import argparse import os import os.path as op from collections import namedtuple from multiprocessing import cpu_count from typing import Li...
EXA-1-master
exa/models/unilm-master/edgelm/examples/byte_level_bpe/get_bitext.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the r...
EXA-1-master
exa/models/unilm-master/edgelm/examples/byte_level_bpe/gru_transformer.py
#!/usr/bin/env python """Helper script to compare two argparse.Namespace objects.""" from argparse import Namespace # noqa def main(): ns1 = eval(input("Namespace 1: ")) ns2 = eval(input("Namespace 2: ")) def keys(ns): ks = set() for k in dir(ns): if not k.startswith("_"): ...
EXA-1-master
exa/models/unilm-master/edgelm/scripts/compare_namespaces.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. """ Split a large file into a train and valid set while respecting document boundaries. Documents should be separated by...
EXA-1-master
exa/models/unilm-master/edgelm/scripts/split_train_valid_docs.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. """ Use this script in order to build symmetric alignments for your translation dataset. This script depends on fast_align and mosesdecoder too...
EXA-1-master
exa/models/unilm-master/edgelm/scripts/build_sym_alignment.py
#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. from __future__ import absolute_import, division, print_function, unicode_literals import argparse ...
EXA-1-master
exa/models/unilm-master/edgelm/scripts/spm_decode.py
EXA-1-master
exa/models/unilm-master/edgelm/scripts/__init__.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import argparse import os import re import shutil import sys pt_regexp = re.compile(r"checkpoint(\d+|_\d+_\d+|_[a-z]+...
EXA-1-master
exa/models/unilm-master/edgelm/scripts/rm_pt.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. """ Count the number of documents and average number of lines and tokens per document in a large file. Documents should ...
EXA-1-master
exa/models/unilm-master/edgelm/scripts/count_docs.py
#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. from __future__ import absolute_import, division, print_function, unicode_literals import argparse i...
EXA-1-master
exa/models/unilm-master/edgelm/scripts/spm_encode.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. """ Split a large file into shards while respecting document boundaries. Documents should be separated by a single empty...
EXA-1-master
exa/models/unilm-master/edgelm/scripts/shard_docs.py
#!/usr/bin/env python # Copyright (c) Facebook, Inc. and its affiliates. # All rights reserved. # # This source code is licensed under the license found in the # LICENSE file in the root directory of this source tree. from __future__ import absolute_import, division, print_function, unicode_literals import sys impor...
EXA-1-master
exa/models/unilm-master/edgelm/scripts/spm_train.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import argparse import collections import os import re import torch from fairseq.file_io import PathManager def aver...
EXA-1-master
exa/models/unilm-master/edgelm/scripts/average_checkpoints.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import argparse from fairseq.data import Dictionary, data_utils, indexed_dataset def get_parser(): parser = argp...
EXA-1-master
exa/models/unilm-master/edgelm/scripts/read_binarized.py
#!/usr/bin/env python3 # # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import sys """Reads in a fairseq output file, and verifies that the constraints (C- lines) are present in the outpu...
EXA-1-master
exa/models/unilm-master/edgelm/scripts/constraints/validate.py
#!/usr/bin/env python3 # # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. """Extracts random constraints from reference files.""" import argparse import random import sys from sacrebleu imp...
EXA-1-master
exa/models/unilm-master/edgelm/scripts/constraints/extract.py
# -------------------------------------------------------- # Image as a Foreign Language: BEiT Pretraining for Vision and Vision-Language Tasks (https://arxiv.org/abs/2208.10442) # Github source: https://github.com/microsoft/unilm/tree/master/beit3 # Copyright (c) 2023 Microsoft # Licensed under The MIT License [see LI...
EXA-1-master
exa/models/unilm-master/beit3/engine_for_finetuning.py
# -------------------------------------------------------- # Image as a Foreign Language: BEiT Pretraining for Vision and Vision-Language Tasks (https://arxiv.org/abs/2208.10442) # Github source: https://github.com/microsoft/unilm/tree/master/beit3 # Copyright (c) 2023 Microsoft # Licensed under The MIT License [see LI...
EXA-1-master
exa/models/unilm-master/beit3/datasets.py
import cv2 import numpy as np ## aug functions def identity_func(img): return img def autocontrast_func(img, cutoff=0): ''' same output as PIL.ImageOps.autocontrast ''' n_bins = 256 def tune_channel(ch): n = ch.size cut = cutoff * n // 100 if cut == 0: ...
EXA-1-master
exa/models/unilm-master/beit3/randaug.py
# -------------------------------------------------------- # Image as a Foreign Language: BEiT Pretraining for Vision and Vision-Language Tasks (https://arxiv.org/abs/2208.10442) # Github source: https://github.com/microsoft/unilm/tree/master/beit3 # Copyright (c) 2023 Microsoft # Licensed under The MIT License [see LI...
EXA-1-master
exa/models/unilm-master/beit3/utils.py
# -------------------------------------------------------- # Image as a Foreign Language: BEiT Pretraining for Vision and Vision-Language Tasks (https://arxiv.org/abs/2208.10442) # Github source: https://github.com/microsoft/unilm/tree/master/beit3 # Copyright (c) 2023 Microsoft # Licensed under The MIT License [see LI...
EXA-1-master
exa/models/unilm-master/beit3/run_beit3_finetuning.py
# -------------------------------------------------------- # Image as a Foreign Language: BEiT Pretraining for Vision and Vision-Language Tasks (https://arxiv.org/abs/2208.10442) # Github source: https://github.com/microsoft/unilm/tree/master/beit3 # Copyright (c) 2023 Microsoft # Licensed under The MIT License [see LI...
EXA-1-master
exa/models/unilm-master/beit3/modeling_utils.py
# -------------------------------------------------------- # Image as a Foreign Language: BEiT Pretraining for Vision and Vision-Language Tasks (https://arxiv.org/abs/2208.10442) # Github source: https://github.com/microsoft/unilm/tree/master/beit3 # Copyright (c) 2023 Microsoft # Licensed under The MIT License [see LI...
EXA-1-master
exa/models/unilm-master/beit3/modeling_finetune.py
# -------------------------------------------------------- # Image as a Foreign Language: BEiT Pretraining for Vision and Vision-Language Tasks (https://arxiv.org/abs/2208.10442) # Github source: https://github.com/microsoft/unilm/tree/master/beit3 # Copyright (c) 2023 Microsoft # Licensed under The MIT License [see LI...
EXA-1-master
exa/models/unilm-master/beit3/optim_factory.py
import re contractions = { "aint": "ain't", "arent": "aren't", "cant": "can't", "couldve": "could've", "couldnt": "couldn't", "couldn'tve": "couldn't've", "couldnt've": "couldn't've", "didnt": "didn't", "doesnt": "doesn't", "dont": "don't", "hadnt": "hadn't", "hadnt've":...
EXA-1-master
exa/models/unilm-master/beit3/glossary.py
#!/usr/bin/env python3 from setuptools import find_packages, setup setup( name="layoutlmv3", version="0.1", author="LayoutLM Team", url="https://github.com/microsoft/unilm/tree/master/layoutlmv3", packages=find_packages(), python_requires=">=3.7", extras_require={"dev": ["flake8", "isort", "...
EXA-1-master
exa/models/unilm-master/layoutlmv3/setup.py
from .models import ( LayoutLMv3Config, LayoutLMv3ForTokenClassification, LayoutLMv3ForQuestionAnswering, LayoutLMv3ForSequenceClassification, LayoutLMv3Tokenizer, )
EXA-1-master
exa/models/unilm-master/layoutlmv3/layoutlmft/__init__.py
from .layoutlmv3 import ( LayoutLMv3Config, LayoutLMv3ForTokenClassification, LayoutLMv3ForQuestionAnswering, LayoutLMv3ForSequenceClassification, LayoutLMv3Tokenizer, )
EXA-1-master
exa/models/unilm-master/layoutlmv3/layoutlmft/models/__init__.py
# coding=utf-8 # Copyright 2018 The Open AI Team Authors and The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # ...
EXA-1-master
exa/models/unilm-master/layoutlmv3/layoutlmft/models/layoutlmv3/tokenization_layoutlmv3.py
from transformers import AutoConfig, AutoModel, AutoModelForTokenClassification, \ AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer from transformers.convert_slow_tokenizer import SLOW_TO_FAST_CONVERTERS, RobertaConverter from .configuration_layoutlmv3 import LayoutLMv3Config from ....
EXA-1-master
exa/models/unilm-master/layoutlmv3/layoutlmft/models/layoutlmv3/__init__.py
# coding=utf-8 # Copyright 2018 The Open AI Team Authors and The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # ...
EXA-1-master
exa/models/unilm-master/layoutlmv3/layoutlmft/models/layoutlmv3/tokenization_layoutlmv3_fast.py
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a cop...
EXA-1-master
exa/models/unilm-master/layoutlmv3/layoutlmft/models/layoutlmv3/modeling_layoutlmv3.py
# coding=utf-8 from transformers.models.bert.configuration_bert import BertConfig from transformers.utils import logging logger = logging.get_logger(__name__) LAYOUTLMV3_PRETRAINED_CONFIG_ARCHIVE_MAP = { "layoutlmv3-base": "https://huggingface.co/microsoft/layoutlmv3-base/resolve/main/config.json", "layoutlm...
EXA-1-master
exa/models/unilm-master/layoutlmv3/layoutlmft/models/layoutlmv3/configuration_layoutlmv3.py
import torchvision.transforms.functional as F import warnings import math import random import numpy as np from PIL import Image import torch from detectron2.data.detection_utils import read_image from detectron2.data.transforms import ResizeTransform, TransformList def normalize_bbox(bbox, size): return [ ...
EXA-1-master
exa/models/unilm-master/layoutlmv3/layoutlmft/data/image_utils.py
import os import json import torch from torch.utils.data.dataset import Dataset from torchvision import transforms from PIL import Image from layoutlmft.data.image_utils import Compose, RandomResizedCropAndInterpolationWithTwoPic XFund_label2ids = { "O":0, 'B-HEADER':1, 'I-HEADER':2, 'B-QUESTION':3, ...
EXA-1-master
exa/models/unilm-master/layoutlmv3/layoutlmft/data/xfund.py
''' Reference: https://huggingface.co/datasets/pierresi/cord/blob/main/cord.py ''' import json import os from pathlib import Path import datasets from layoutlmft.data.image_utils import load_image, normalize_bbox logger = datasets.logging.get_logger(__name__) _CITATION = """\ @article{park2019cord, title={CORD: A C...
EXA-1-master
exa/models/unilm-master/layoutlmv3/layoutlmft/data/cord.py
# flake8: noqa from .data_collator import DataCollatorForKeyValueExtraction
EXA-1-master
exa/models/unilm-master/layoutlmv3/layoutlmft/data/__init__.py
import torch from dataclasses import dataclass from typing import Any, Dict, List, Optional, Tuple, Union from transformers import BatchEncoding, PreTrainedTokenizerBase from transformers.data.data_collator import ( DataCollatorMixin, _torch_collate_batch, ) from transformers.file_utils import PaddingStrategy ...
EXA-1-master
exa/models/unilm-master/layoutlmv3/layoutlmft/data/data_collator.py
# coding=utf-8 ''' Reference: https://huggingface.co/datasets/nielsr/funsd/blob/main/funsd.py ''' import json import os import datasets from layoutlmft.data.image_utils import load_image, normalize_bbox logger = datasets.logging.get_logger(__name__) _CITATION = """\ @article{Jaume2019FUNSDAD, title={FUNSD: A Da...
EXA-1-master
exa/models/unilm-master/layoutlmv3/layoutlmft/data/funsd.py
#!/usr/bin/env python # coding=utf-8 import logging import os import sys from dataclasses import dataclass, field from typing import Optional import numpy as np from datasets import ClassLabel, load_dataset, load_metric import transformers from layoutlmft.data import DataCollatorForKeyValueExtraction from layoutlmft...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/run_xfund.py
#!/usr/bin/env python # coding=utf-8 import logging import os import sys from dataclasses import dataclass, field from typing import Optional import numpy as np from datasets import ClassLabel, load_dataset, load_metric import transformers from layoutlmft.data import DataCollatorForKeyValueExtraction from transforme...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/run_funsd_cord.py
import os from PIL import Image import xml.etree.ElementTree as ET import numpy as np import json from PIL import Image from shutil import copyfile def convert(ROOT, TRACK, SPLIT): coco_data = { "images": [], "annotations": [], "categories": [{"id": 1, "name": "table"}, ], } DATA_D...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/convert_to_coco_format.py
#!/usr/bin/env python # -------------------------------------------------------------------------------- # MPViT: Multi-Path Vision Transformer for Dense Prediction # Copyright (c) 2022 Electronics and Telecommunications Research Institute (ETRI). # All Rights Reserved. # Written by Youngwan Lee # ---------------------...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/train_net.py
import argparse import os import cv2 import tqdm def convert(fn): # given a file name, convert it into binary and store at the same position img = cv2.imread(fn) gim = cv2.cvtColor(img, cv2.COLOR_BGR2GRAY) gim = cv2.adaptiveThreshold(gim, 255, cv2.ADAPTIVE_THRESH_GAUSSIAN_C, cv2.THRESH_BINARY, 45, 11)...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/adaptive_binarize.py
""" Mostly copy-paste from DINO and timm library: https://github.com/facebookresearch/dino https://github.com/rwightman/pytorch-image-models/blob/master/timm/models/vision_transformer.py """ import warnings import math import torch import torch.nn as nn import torch.utils.checkpoint as checkpoint from timm.models.laye...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/ditod/deit.py
import copy import itertools import os import os.path as osp import shutil from collections import OrderedDict from xml.dom.minidom import Document import detectron2.utils.comm as comm import torch from detectron2.evaluation import COCOEvaluator from detectron2.utils.file_io import PathManager from .table_evaluation....
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/ditod/icdar_evaluation.py
""" Vision Transformer (ViT) in PyTorch A PyTorch implement of Vision Transformers as described in 'An Image Is Worth 16 x 16 Words: Transformers for Image Recognition at Scale' - https://arxiv.org/abs/2010.11929 The official jax code is released and available at https://github.com/google-research/vision_transformer ...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/ditod/beit.py
from detectron2.config import CfgNode as CN def add_vit_config(cfg): """ Add config for VIT. """ _C = cfg _C.MODEL.VIT = CN() # CoaT model name. _C.MODEL.VIT.NAME = "" # Output features from CoaT backbone. _C.MODEL.VIT.OUT_FEATURES = ["layer3", "layer5", "layer7", "layer11"] ...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/ditod/config.py
from detectron2.checkpoint import DetectionCheckpointer from typing import Any import torch import torch.nn as nn from fvcore.common.checkpoint import _IncompatibleKeys, _strip_prefix_if_present, TORCH_VERSION, quantization, \ ObserverBase, FakeQuantizeBase from torch import distributed as dist from scipy import i...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/ditod/mycheckpointer.py
# Copyright (c) Facebook, Inc. and its affiliates. import logging import numpy as np from typing import Dict, List, Optional, Tuple import torch from torch import nn from detectron2.config import configurable from detectron2.structures import ImageList, Instances from detectron2.utils.events import get_event_storage ...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/ditod/rcnn_vl.py
# -------------------------------------------------------------------------------- # VIT: Multi-Path Vision Transformer for Dense Prediction # Copyright (c) 2022 Electronics and Telecommunications Research Institute (ETRI). # All Rights Reserved. # Written by Youngwan Lee # This source code is licensed(Dual License(GPL...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/ditod/backbone.py
# -------------------------------------------------------------------------------- # MPViT: Multi-Path Vision Transformer for Dense Prediction # Copyright (c) 2022 Electronics and Telecommunications Research Institute (ETRI). # All Rights Reserved. # Written by Youngwan Lee # This source code is licensed(Dual License(G...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/ditod/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. All Rights Reserved # from https://github.com/facebookresearch/detr/blob/main/d2/detr/dataset_mapper.py import copy import logging import numpy as np import torch from detectron2.data import detection_utils as utils from detectron2.data import transforms as T from...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/ditod/dataset_mapper.py
# -*- coding: utf-8 -*- # Copyright (c) Facebook, Inc. and its affiliates. """ This file contains components with some default boilerplate logic user may need in training / testing. They will not work for everyone, but many users may find them useful. The behavior of functions/classes in this file is subject to chang...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/ditod/mytrainer.py
from .evaluate import calc_table_score
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/ditod/table_evaluation/__init__.py
""" Evaluation of -.tar.gz file. Yu Fang - March 2019 """ import os import xml.dom.minidom # from eval import eval if os.path.exists("/mnt/localdata/Users/junlongli/projects/datasets/icdar2019"): PATH = "/mnt/localdata/Users/junlongli/projects/datasets/icdar2019/trackA_modern/test" else: PATH = "/mnt/data/dat...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/ditod/table_evaluation/evaluate.py
""" Data structures used by the evaluation process. Yu Fang - March 2019 """ from collections import Iterable import numpy as np from shapely.geometry import Polygon # helper functions def flatten(lis): for item in lis: if isinstance(item, Iterable) and not isinstance(item, str): for x in fl...
EXA-1-master
exa/models/unilm-master/layoutlmv3/examples/object_detection/ditod/table_evaluation/data_structure.py
# -------------------------------------------------------- # BEATs: Audio Pre-Training with Acoustic Tokenizers (https://arxiv.org/abs/2212.09058) # Github source: https://github.com/microsoft/unilm/tree/master/beats # Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] # Based on fa...
EXA-1-master
exa/models/unilm-master/beats/backbone.py
# -------------------------------------------------------- # BEATs: Audio Pre-Training with Acoustic Tokenizers (https://arxiv.org/abs/2212.09058) # Github source: https://github.com/microsoft/unilm/tree/master/beats # Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] # Based on VQ...
EXA-1-master
exa/models/unilm-master/beats/quantizer.py
# -------------------------------------------------------- # BEATs: Audio Pre-Training with Acoustic Tokenizers (https://arxiv.org/abs/2212.09058) # Github source: https://github.com/microsoft/unilm/tree/master/beats # Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] # Based on fa...
EXA-1-master
exa/models/unilm-master/beats/BEATs.py
# -------------------------------------------------------- # BEATs: Audio Pre-Training with Acoustic Tokenizers (https://arxiv.org/abs/2212.09058) # Github source: https://github.com/microsoft/unilm/tree/master/beats # Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] # Based on fa...
EXA-1-master
exa/models/unilm-master/beats/modules.py
# -------------------------------------------------------- # BEATs: Audio Pre-Training with Acoustic Tokenizers (https://arxiv.org/abs/2212.09058) # Github source: https://github.com/microsoft/unilm/tree/master/beats # Copyright (c) 2022 Microsoft # Licensed under The MIT License [see LICENSE for details] # Based on fa...
EXA-1-master
exa/models/unilm-master/beats/Tokenizers.py
import os import sys import time import logging from tqdm import tqdm import torch from fairseq import utils, tasks, options from fairseq.checkpoint_utils import load_model_ensemble_and_task from fairseq.dataclass.utils import convert_namespace_to_omegaconf from torch import Tensor from typing import Dict, List, Opti...
EXA-1-master
exa/models/unilm-master/decoding/GAD/inference_paper.py
#!/usr/bin/env python3 # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import os import subprocess import sys from setuptools import setup, find_packages, Extension from setuptools import E...
EXA-1-master
exa/models/unilm-master/decoding/GAD/setup.py
#!/usr/bin/env python3 -u # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. """ Legacy entry point. Use fairseq_cli/train.py or fairseq-train instead. """ from fairseq_cli.train import cli_mai...
EXA-1-master
exa/models/unilm-master/decoding/GAD/train.py
import os import sys import time import logging from tqdm import tqdm import torch from fairseq import utils, tasks, options from fairseq.checkpoint_utils import load_model_ensemble_and_task from fairseq.dataclass.utils import convert_namespace_to_omegaconf from torch import Tensor from typing import Dict, List, Opti...
EXA-1-master
exa/models/unilm-master/decoding/GAD/inference.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. """isort:skip_file""" import functools import importlib dependencies = [ "dataclasses", "hydra", "numpy", "omegaconf", "...
EXA-1-master
exa/models/unilm-master/decoding/GAD/hubconf.py
from .criterions import * from .models import * from .tasks import * print("GAD plugins loaded...")
EXA-1-master
exa/models/unilm-master/decoding/GAD/block_plugins/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. from dataclasses import dataclass, field from math import log import torch from fairseq import utils from fairseq.data import LanguagePairData...
EXA-1-master
exa/models/unilm-master/decoding/GAD/block_plugins/tasks/translation_lev_modified.py
from .translation_lev_modified import *
EXA-1-master
exa/models/unilm-master/decoding/GAD/block_plugins/tasks/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import torch from fairseq import utils from fairseq.iterative_refinement_generator import DecoderOut from fairseq.models import register_model...
EXA-1-master
exa/models/unilm-master/decoding/GAD/block_plugins/models/BlockNAT.py
from .BlockNAT import *
EXA-1-master
exa/models/unilm-master/decoding/GAD/block_plugins/models/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import math from math import log import torch import torch.nn.functional as F from fairseq import metrics, utils from fairseq.criterions impo...
EXA-1-master
exa/models/unilm-master/decoding/GAD/block_plugins/criterions/glat_loss.py
from .glat_loss import *
EXA-1-master
exa/models/unilm-master/decoding/GAD/block_plugins/criterions/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import argparse from typing import Callable, List, Optional import torch from fairseq import utils from fairseq.data.indexed_dataset import g...
EXA-1-master
exa/models/unilm-master/decoding/GAD/fairseq/options.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. from collections import namedtuple import numpy as np import torch from fairseq import utils DecoderOut = namedtuple( "IterativeRefinem...
EXA-1-master
exa/models/unilm-master/decoding/GAD/fairseq/iterative_refinement_generator.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import logging import torch logger = logging.getLogger(__name__) class NanDetector: """ Detects the first NaN or Inf in forward a...
EXA-1-master
exa/models/unilm-master/decoding/GAD/fairseq/nan_detector.py
__version__ = "1.0.0a0"
EXA-1-master
exa/models/unilm-master/decoding/GAD/fairseq/version.py
# Originally from Microsoft Corporation. # Licensed under the MIT License. """ Wrapper for ngram_repeat_block cuda extension """ import torch from torch import nn import math from typing import Dict, List, Optional import warnings try: from fairseq import ngram_repeat_block_cuda EXTENSION_BUILT = True excep...
EXA-1-master
exa/models/unilm-master/decoding/GAD/fairseq/ngram_repeat_block.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. from argparse import Namespace from typing import Union from fairseq.dataclass import FairseqDataclass from fairseq.dataclass.utils import po...
EXA-1-master
exa/models/unilm-master/decoding/GAD/fairseq/registry.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. """isort:skip_file""" import os import sys try: from .version import __version__ # noqa except ImportError: version_txt = os.path.jo...
EXA-1-master
exa/models/unilm-master/decoding/GAD/fairseq/__init__.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import math from typing import Dict, List, Optional import torch import torch.nn as nn from fairseq import search, utils from fairseq.data im...
EXA-1-master
exa/models/unilm-master/decoding/GAD/fairseq/sequence_generator.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import multiprocessing import os import pdb import sys __all__ = ["set_trace"] _stdin = [None] _stdin_lock = multiprocessing.Lock() try: ...
EXA-1-master
exa/models/unilm-master/decoding/GAD/fairseq/pdb.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import re SPACE_NORMALIZER = re.compile(r"\s+") def tokenize_line(line): line = SPACE_NORMALIZER.sub(" ", line) line = line.strip(...
EXA-1-master
exa/models/unilm-master/decoding/GAD/fairseq/tokenizer.py
#!/usr/bin/env python3 -u # Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import argparse import copy import logging import os from typing import Any, Dict, Iterator, List import torch from...
EXA-1-master
exa/models/unilm-master/decoding/GAD/fairseq/hub_utils.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import sys import torch from fairseq import utils class SequenceScorer(object): """Scores the target for a given source sentence.""" ...
EXA-1-master
exa/models/unilm-master/decoding/GAD/fairseq/sequence_scorer.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import uuid from typing import Dict, Optional from torch import Tensor class FairseqIncrementalState(object): def __init__(self, *args,...
EXA-1-master
exa/models/unilm-master/decoding/GAD/fairseq/incremental_decoding_utils.py
# Copyright (c) Facebook, Inc. and its affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. import argparse import contextlib import copy import importlib import logging import os import sys import tempfile import warnings from iterto...
EXA-1-master
exa/models/unilm-master/decoding/GAD/fairseq/utils.py