| | from concurrent.futures import ThreadPoolExecutor |
| | import os |
| | import random |
| | import subprocess |
| | import timeit |
| |
|
| | import datasets |
| |
|
| | """Example |
| | git init |
| | git remote add origin https://github.com/huggingface/evaluate.git |
| | git fetch --depth 2 origin 9b056cdd5eb95459ae80142014865263e7dd75b8 |
| | # Get file after change |
| | git checkout FETCH_HEAD -- README.md |
| | # Get file before change |
| | git checkout FETCH_HEAD^ -- README.md |
| | """ |
| |
|
| | |
| | |
| |
|
| | NUM_THREADS = 64 |
| | NUM_PROC = 64 |
| | |
| |
|
| | CWD = os.getcwd() |
| |
|
| | |
| | def run_in_shell(cmd: str, cwd=None, timeout=60): |
| | return subprocess.run([cmd], capture_output=True, shell=True, cwd=cwd, timeout=timeout) |
| |
|
| | def get_file_contents(commit, old_file, new_file, repo, cwd=None): |
| | completed = run_in_shell("git init", cwd=cwd) |
| | completed = run_in_shell("git remote add origin " + repo, cwd=cwd) |
| | completed = run_in_shell("git fetch --depth 2 origin " + commit, cwd=cwd) |
| | |
| | if completed.returncode != 0: |
| | |
| | return ("", "") |
| | |
| | |
| | completed = run_in_shell("git checkout FETCH_HEAD -- " + new_file, cwd=cwd) |
| | new_contents = run_in_shell("cat " + new_file, cwd=cwd).stdout.decode(errors='ignore') |
| | completed = run_in_shell("git checkout FETCH_HEAD^ -- " + old_file, cwd=cwd) |
| | |
| | if completed.returncode != 0: |
| | |
| | return (new_contents, "") |
| | old_contents = run_in_shell("cat " + old_file, cwd=cwd).stdout.decode(errors='ignore') |
| | return (new_contents, old_contents) |
| |
|
| | def get_diff(ex): |
| | commit_id = ex["commit"] |
| | repos = list(set(ex["repos"].split(","))) |
| | old_file = ex["old_file"] |
| | new_file = ex["new_file"] |
| | for i, repo in enumerate(repos): |
| | repo = "https://xxx:xxx@github.com/" + repo + ".git" |
| | |
| | random_dir = CWD + "/" + str(random.randint(0, 1000000)) |
| | |
| | run_in_shell("mkdir " + random_dir, timeout=300) |
| | try: |
| | new_contents, old_contents = get_file_contents(commit_id, old_file, new_file, repo, cwd=random_dir) |
| | except Exception as e: |
| | |
| | |
| | if i > 10: |
| | break |
| | continue |
| | finally: |
| | run_in_shell("rm -rf " + random_dir) |
| | ex["new_contents"] = new_contents |
| | ex["old_contents"] = old_contents |
| | return ex |
| | |
| | ex["new_contents"] = "" |
| | ex["old_contents"] = "" |
| | return ex |
| |
|
| | def get_diff_multi_threaded_processed(batch): |
| | with ThreadPoolExecutor(max_workers=NUM_THREADS) as executor: |
| | |
| | results = list(executor.map(get_diff, [dict(zip(batch,t)) for t in zip(*batch.values())])) |
| | |
| | return {k: [dic[k] for dic in results] for k in results[0]} |
| |
|
| | if __name__ == "__main__": |
| | |
| | ds = datasets.load_dataset("./github-commits", use_auth_token=True)["train"] |
| |
|
| | |
| | |
| | java = tuple([".java"]) |
| | javascript = tuple([ |
| | ".js", |
| | "._js", |
| | ".bones", |
| | ".es6", |
| | ".jake", |
| | ".jsb", |
| | ".jscad", |
| | ".jsfl", |
| | ".jsm", |
| | ".jss", |
| | ".njs", |
| | ".pac", |
| | ".sjs", |
| | ".ssjs", |
| | ".xsjs", |
| | ".xsjslib" |
| | ]) |
| | python = tuple([ |
| | ".py", |
| | ".bzl", |
| | ".gyp", |
| | ".lmi", |
| | ".pyde", |
| | ".pyp", |
| | ".pyt", |
| | ".pyw", |
| | ".tac", |
| | ".wsgi", |
| | ".xpy", |
| | ]) |
| |
|
| | import json |
| | with open("programming-languages.json", "r") as f: |
| | extensions = json.load(f) |
| | suffices = tuple([suffix for suffices in extensions.values() for suffix in suffices] ) |
| | def filter_extension(ex): |
| | return ex["new_file"].endswith(suffices) |
| | |
| | def filter_extension_python(ex): |
| | return ex["new_file"].endswith(python) |
| | |
| | def filter_update(ex): |
| | return ex["message"] != "Update " + ex["old_file"] |
| |
|
| | filter_msg = ["initial commit", "please\n", "please", "lalala"] |
| | |
| | def filter_misc(ex): |
| | return ex["message"] not in filter_msg |
| |
|
| | |
| | ds = ds.filter(filter_extension, num_proc=NUM_PROC) |
| | print("After Extension filter", len(ds)) |
| | |
| | ds = ds.filter(filter_update, num_proc=NUM_PROC) |
| | print("After Update filter", len(ds)) |
| | |
| | |
| | ds = ds.filter(filter_misc, num_proc=NUM_PROC) |
| | print("After Misc filter", len(ds)) |
| | |
| | START = 0 |
| | samples_per_instance = 64 * 64 * 64 * 32 |
| | select_start = START * samples_per_instance |
| | select_end = START * samples_per_instance + samples_per_instance |
| | ds = ds.select(range(select_start, select_end)) |
| | print(f"Going from {select_start} till {select_end}") |
| | |
| | |
| | |
| | |
| | def run_multi_processing_threading(): |
| | ds.map(get_diff_multi_threaded_processed, num_proc=NUM_PROC, batch_size=NUM_THREADS, batched=True).to_json(f"diffs_{select_start}_{select_end}.jsonl", num_proc=NUM_PROC)) |
| |
|
| | |
| | |
| | |
| | |
| | |
| | |
| | run_multi_processing_threading() |
| |
|