proof-pile / fetch_books_and_formal.py
Zhangir Azerbayev
converted books and formal to jsonl gz
c91a5cf
raw
history blame
18.8 kB
import sys
import os
import json
import ndjson
import re
from pathlib import Path
from tqdm import tqdm
import requests
from tqdm import tqdm
import base64
def jsonl_of_path(path, jsonl_train_path, jsonl_val_path,
train_split_key, val_split_key):
train_instances = []
val_instances = []
print("CREATING JSONL.GZ")
with open("splits.json") as f:
splits = json.load(f)
for root, dirs, files in tqdm(os.walk(path)):
for name in files:
this_path = os.path.join(root, name)
with open(this_path) as f:
text = f.read()
instance = {"text": text,
"meta": {
"subset_name": "curated",
"file": os.path.join(root, name)
}
}
if this_path in splits[train_split_key]:
train_instances.append(instance)
elif this_path in splits[val_split_key]:
val_instances.append(instance)
else:
raise KeyError("key not found in splits.json")
with open(jsonl_train_path, "w") as f:
ndjson.dump(train_instances, f)
os.system("gzip " + jsonl_train_path)
with open(jsonl_val_path, "w") as f:
ndjson.dump(val_instances, f)
os.system("gzip " + jsonl_val_path)
os.system("rm -r " + path)
print("succesful conversion to jsonl")
def check_encoding(path):
for f in os.listdir(path):
f_path = os.path.join(path, f)
if os.path.isfile(f_path):
with open(f_path, encoding="utf-8") as fle:
try:
fle.read()
except UnicodeDecodeError:
print(f"{f_path} is not unicode")
elif os.path.isdir(f_path):
check_encoding(f_path)
def _get_dir_from_repo(author, repo, sha, repo_dir, save_path, creds):
"""
This super unelegant solution is to get around the github api rate limit
repo_dir must be top-level in the repo.
"""
Path(save_path).mkdir(parents=True, exist_ok=True)
archive_path = os.path.join(save_path, "archive.tar.gz")
tarball_url = (
"https://github.com/" + author + "/" + repo + "/archive/" + sha + ".tar.gz"
)
os.system("wget -O " + archive_path + " " + tarball_url)
os.system("tar -xzf " + archive_path + " -C " + save_path)
export_name = repo + "-" + sha
os.system(
"cp -r " + os.path.join(save_path, export_name, repo_dir, "*") + " " + save_path
)
os.system("rm -r " + os.path.join(save_path, export_name) + " " + archive_path)
def _delete_files_except_pattern(path, pattern):
"""
recursively
"""
for f in os.listdir(path):
f_path = os.path.join(path, f)
if os.path.isfile(f_path):
if not re.search(pattern, f):
os.remove(f_path)
else: # debugging
with open(f_path, encoding="utf-8") as f:
try:
f.read()
except:
print(f"{f_path} not unicode encoded")
elif os.path.islink(f_path):
os.remove(f_path)
elif os.path.isdir(f_path):
_delete_files_except_pattern(f_path, pattern)
def _download_with_progress_bar(url):
response = requests.get(url, stream=True)
total_size_in_bytes = int(response.headers.get("content-length", 0))
block_size = 1024 # 1 Kibibyte
progress_bar = tqdm(total=total_size_in_bytes, unit="iB", unit_scale=True)
to_return = bytearray()
for data in response.iter_content(block_size):
progress_bar.update(len(data))
to_return += data
progress_bar.close()
if total_size_in_bytes != 0 and progress_bar.n != total_size_in_bytes:
raise AssertionError("ERROR, something went wrong")
return to_return
def _blob_to_text(blob, creds):
resp = requests.get(blob["url"], auth=creds)
if resp.status_code != 200:
raise AssertionError("Failed to fetch from Github API")
resp_json = json.loads(resp.content.decode("utf-8"))
return base64.b64decode(resp_json["content"])
def lean(creds):
save_dir = "formal/lean"
Path(save_dir).mkdir(parents=True, exist_ok=True)
sources = [
{
"author": "leanprover-community",
"repo": "mathlib",
"sha": "63138639ca195344ae96aa77f3a02b90a3ac5c68",
"repo_dir": "src",
"save_path": os.path.join(save_dir, "mathlib"),
},
{
"author": "leanprover-community",
"repo": "lean-liquid",
"sha": "9701fc4a29514852b599e9732c2409f34153ce2a",
"repo_dir": "src",
"save_path": os.path.join(save_dir, "liquid"),
},
{
"author": "leanprover-community",
"repo": "sphere-eversion",
"sha": "cb378966c3c02d9e4ee83040d20c51782fa351ae",
"repo_dir": "src",
"save_path": os.path.join(save_dir, "sphere-eversion"),
},
{
"author": "leanprover-community",
"repo": "lftcm2020",
"sha": "8b9f7c47b546227b7b6c877315e45eaccc2a0d70",
"repo_dir": "src",
"save_path": os.path.join(save_dir, "lftcm"),
},
{
"author": "leanprover-community",
"repo": "lean-perfectoid-spaces",
"sha": "95a6520ce578b30a80b4c36e36ab2d559a842690",
"repo_dir": "src",
"save_path": os.path.join(save_dir, "perfectoid"),
},
{
"author": "leanprover-community",
"repo": "mathzoo",
"sha": "87e9b492daeb929838706942aaa2437621b34a0e",
"repo_dir": "src",
"save_path": os.path.join(save_dir, "mathzoo"),
},
]
for source in sources:
_get_dir_from_repo(**source, creds=creds)
_delete_files_except_pattern(source["save_path"], r".*\.lean")
# we also don't want meta code
to_delete = ["tactic", "meta"]
os.system(
"rm -r " + " ".join([os.path.join(save_dir, "mathlib", x) for x in to_delete])
)
jsonl_of_path(save_dir, "formal/lean_train.jsonl", "formal/lean_val.jsonl",
"formal-train", "formal-valid")
def coq(creds):
save_dir = "formal/coq"
Path(save_dir).mkdir(parents=True, exist_ok=True)
sources = [
{
"author": "math-comp",
"repo": "analysis",
"sha": "2ae3b628d12cacdc000c4cd70e6f3cae26ecf429",
"repo_dir": "theories",
"save_path": os.path.join(save_dir, "analysis"),
},
{
"author": "math-comp",
"repo": "math-comp",
"sha": "65519a110ffdad7869b2a7cd08a2ddb51161b377",
"repo_dir": "mathcomp",
"save_path": os.path.join(save_dir, "math-comp"),
},
{
"author": "math-comp",
"repo": "odd-order",
"sha": "833261a01fd0c62b05ccbadfc0c682e0bc16a5e9",
"repo_dir": "theories",
"save_path": os.path.join(save_dir, "odd-order"),
},
{
"author": "math-comp",
"repo": "Abel",
"sha": "61d79aeb0acc1855e22882c484b73645df53b746",
"repo_dir": "theories",
"save_path": os.path.join(save_dir, "abel"),
},
]
for source in sources:
_get_dir_from_repo(**source, creds=creds)
_delete_files_except_pattern(source["save_path"], r".*\.v")
jsonl_of_path(save_dir, "formal/coq_train.jsonl", "formal/coq_val.jsonl",
"formal-train", "formal-valid")
def trench():
save_dir = "books/trench"
archive_path = os.path.join(save_dir, "trench.zip")
Path(save_dir).mkdir(parents=True, exist_ok=True)
print("DOWNLOADING TRENCH")
os.system(
"wget -O "
+ archive_path
+ ' "https://digitalcommons.trinity.edu/cgi/viewcontent.cgi?filename=2&article=1006&context=mono&type=additional"'
)
print("DONE DOWNLOADING TRENCH")
os.system("unzip " + archive_path + " -d " + save_dir)
to_delete = ["trench.zip", "wtrench.sty", "SETEPS.TEX", "EPS"]
os.system("rm -r " + " ".join([os.path.join(save_dir, f) for f in to_delete]))
jsonl_of_path(save_dir, "books/trench_train.jsonl", "books/trench_val.jsonl",
"books-train", "books-valid")
def setmm(creds):
save_dir = "formal/setmm"
Path(save_dir).mkdir(parents=True, exist_ok=True)
headers = {
"Accept": "application/vnd.git-lfs+json",
}
json_data = {
"operation": "download",
"transfer": [
"basic",
],
"objects": [
{
"oid": "ff1a12d49d4c68a05245bfd369af358b93c51b8c141419085bb5cef830f6eb7a",
"size": 182269314,
},
],
}
response = requests.post(
"https://github.com/zhangir-azerbayev/mm-extract.git/info/lfs/objects/batch",
headers=headers,
json=json_data,
)
resp_json = response.json()
download_url = resp_json["objects"][0]["actions"]["download"]["href"]
encoded_src = _download_with_progress_bar(download_url)
src = encoded_src.decode("utf-8")
with open(os.path.join(save_dir, "set.mm"), "w") as f:
f.write(src)
jsonl_of_path(save_dir, "formal/setmm_train.jsonl", "formal/setmm_val.jsonl",
"formal-train", "formal-valid")
def stein(creds):
save_dir = "books/stein"
Path(save_dir).mkdir(parents=True, exist_ok=True)
print("DOWNLOADING STEIN")
resp = _download_with_progress_bar(
"https://api.github.com/repos/williamstein/ent/git/blobs/a70578277b1222c94dc395f7d5baaf9862afd166"
)
print("DONE DOWNLOADING STEIN")
resp_json = json.loads(resp.decode("utf-8"))
src_encoded = base64.b64decode(resp_json["content"])
src = src_encoded.decode("utf-8")
with open(os.path.join(save_dir, "stein.tex"), "w") as f:
f.write(src)
jsonl_of_path(save_dir, "books/stein_train.jsonl", "books/stein_val.jsonl",
"books-train", "books-valid")
def cam():
save_dir = "books/cam"
archive_path = os.path.join(save_dir, "cam.tar.gz")
Path(save_dir).mkdir(parents=True, exist_ok=True)
os.system("wget -O " + archive_path + " https://github.com/dalcde/cam-notes/archive/06b2239.tar.gz")
os.system ("tar -xf " + archive_path + " -C " + save_dir)
export_name = "cam-notes-06b2239b006f14d833cca2434190ebbf9a304bc6/"
os.system(
"cp -r "
+ os.path.join(
save_dir, export_name, "* ")
+ save_dir
)
os.system("rm -r " + os.path.join(save_dir, export_name))
os.remove(archive_path)
os.remove(os.path.join(save_dir, "header.tex"))
_delete_files_except_pattern(save_dir, r".*\.tex")
jsonl_of_path(save_dir, "books/cam_train.jsonl", "books/cam_val.jsonl",
"books-train", "books-valid")
def hol(testing=False):
save_dir = "formal/hol"
archive_path = os.path.join(save_dir, "hol.zip")
Path(save_dir).mkdir(parents=True, exist_ok=True)
if not testing:
os.system(
"wget -O "
+ archive_path
+ " https://github.com/jrh13/hol-light/archive/538c62f.tar.gz"
)
os.system("tar -xvf " + archive_path + " -C " + save_dir)
os.system(
"mv "
+ os.path.join(
save_dir, "hol-light-538c62f7cdb0df146752c83f85fa672ae3906b03/* "
)
+ save_dir
)
os.system(
"rm -r "
+ os.path.join(save_dir, "hol-light-538c62f7cdb0df146752c83f85fa672ae3906b03")
)
os.system("rm " + archive_path)
# all top level files are metaprogramming, so delete them
for f in os.listdir(save_dir):
f_path = os.path.join(save_dir, f)
if os.path.isfile(f_path):
os.remove(f_path)
os.system("rm -r formal/hol/Proofrecording")
_delete_files_except_pattern(save_dir, r".*\.ml|.*\.doc")
jsonl_of_path(save_dir, "formal/hol_train.jsonl", "formal/hol_val.jsonl",
"formal-train", "formal-valid")
def afp(testing=False):
save_dir = "formal/afp"
archive_path = os.path.join(save_dir, "afp.zip")
Path(save_dir).mkdir(parents=True, exist_ok=True)
if not testing:
os.system(
"wget -O "
+ archive_path
+ " https://github.com/isabelle-prover/mirror-afp-2021-1/archive/5a85b23.tar.gz"
)
os.system("tar -xf " + archive_path + " -C " + save_dir)
os.system(
"mv "
+ os.path.join(
save_dir,
"mirror-afp-2021-1-5a85b23fb030c472d9a7b2d65a61e428f4eb8233/thys/* ",
)
+ save_dir
)
os.system(
"rm -r "
+ os.path.join(
save_dir, "mirror-afp-2021-1-5a85b23fb030c472d9a7b2d65a61e428f4eb8233"
)
)
os.system("rm " + archive_path)
_delete_files_except_pattern(save_dir, r".*\.thy|.*\.tex")
jsonl_of_path(save_dir, "formal/afp_train.jsonl", "formal/afp_val.jsonl",
"formal-train", "formal-valid")
def mizar(creds):
save_dir = "formal/mizar"
Path(save_dir).mkdir(parents=True, exist_ok=True)
resp = requests.get(
"https://api.github.com/repos/zhangir-azerbayev/mizar-mirror/git/trees/ce8e9735fd7a4d3488069c48da76bc622aec46ec"
)
if resp.status_code != 200:
raise AssertionError("Failed to fetch mizar from Github API")
resp_json = resp.json()
tree = resp_json["tree"]
print("DOWNLOADING MIZAR")
for blob in tqdm(tree):
assert blob["type"] == "blob"
src = _blob_to_text(blob, creds)
src = src.decode("utf-8")
# mml files have licensing information from lines 2-12
src = "\n".join(
[x for i, x in enumerate(src.split("\n")) if i not in range(2, 13)]
)
save_path = os.path.join(save_dir, blob["path"])
with open(save_path, "w") as f:
f.write(src)
print("DONE DOWNLOADING MIZAR")
jsonl_of_path(save_dir, "formal/mizar_train.jsonl", "formal/mizar_val.jsonl",
"formal-train", "formal-valid")
def hott(creds):
save_dir = "books/hott"
Path(save_dir).mkdir(parents=True, exist_ok=True)
resp = requests.get(
"https://api.github.com/repos/HoTT/book/git/trees/781565e93979f926001a353bf4ee1284ffa4fcb0",
auth=creds,
)
if resp.status_code != 200:
raise AssertionError("Failed to fetch HoTT book from Github API")
resp_json = resp.json()
tree = resp_json["tree"]
blobs = [blob for blob in tree if blob["type"] == "blob"]
banned = [
"back.tex",
"bmpsize-hack.tex",
"main.tex",
]
banned_rgx = r"opt|cover|front|hott"
print("DOWNLOADING HOTT BOOK")
for blob in tqdm(blobs):
if (
blob["path"][-4:] == ".tex"
and blob["path"] not in banned
and not re.match(banned_rgx, blob["path"])
):
src_enc = _blob_to_text(blob, creds)
src = src_enc.decode("utf-8")
save_path = os.path.join(save_dir, blob["path"])
with open(save_path, "w") as f:
f.write(src)
print("DONE DOWNLOADING HOTT BOOK")
jsonl_of_path(save_dir, "books/hott_train.jsonl", "books/hott_val.jsonl",
"books-train", "books-valid")
def stacks(creds):
save_dir = "books/stacks"
Path(save_dir).mkdir(parents=True, exist_ok=True)
resp = requests.get(
"https://api.github.com/repos/stacks/stacks-project/git/trees/0a847ff5e41b47795be075e130e7810173b35933",
auth=creds,
)
resp_json = json.loads(resp.content.decode("utf-8"))
# assumes everything we need is a top level file, which is true for this commit.
blobs = resp_json["tree"]
print("DOWNLOADING STACKS")
for blob in tqdm(blobs):
if (
blob["type"] == "blob"
and blob["path"][-4:] == ".tex"
and blob["path"] != "fdl.tex"
):
decoded_content = _blob_to_text(blob, creds)
with open(os.path.join(save_dir, blob["path"]), "wb") as f:
f.write(decoded_content)
print("DONE DOWNLOADING STACKS")
jsonl_of_path(save_dir, "books/stacks_train.jsonl", "books/stacks_val.jsonl",
"books-train", "books-valid")
def cring(creds):
save_dir = "books/cring"
Path(save_dir).mkdir(parents=True, exist_ok=True)
resp = requests.get(
"https://api.github.com/repos/aisejohan/cring/git/trees/2db2618ff70831002aeefbb16885ee42d5198db3",
auth=creds,
)
if resp.status_code != 200:
raise AssertionError("Failed to catch cring from Github API")
trees = json.loads(resp.content.decode("utf-8"))["tree"]
print("DOWNLOADING CRING")
for blob in tqdm(trees):
if blob["type"] == "blob" and blob["path"] != "license.tex":
decoded_content = _blob_to_text(blob, creds)
with open(os.path.join(save_dir, blob["path"]), "wb") as f:
f.write(decoded_content)
print("DONE DOWNLOADING CRING")
jsonl_of_path(save_dir, "books/cring_train.jsonl", "books/cring_val.jsonl",
"books-train", "books-valid")
def napkin(creds):
save_dir = "books/napkin"
Path(save_dir).mkdir(parents=True, exist_ok=True)
resp = requests.get(
"https://api.github.com/repos/vEnhance/napkin/git/trees/4f56c2ef5d0faf132ee14c15d96fb0f134d58bf0",
auth=creds,
)
if resp.status_code != 200:
raise AssertionError("Failed to catch napkin tree from Github API")
trees = json.loads(resp.content.decode("utf-8"))["tree"]
# We are assuming that we only want the files exactly two levels deep
print("DOWNLOADING NAPKIN")
for tree in tqdm(trees):
if tree["type"] == "tree":
resp = requests.get(tree["url"], auth=creds)
blobs = json.loads(resp.content.decode("utf-8"))["tree"]
for blob in blobs:
if blob["type"] == "blob":
decoded_content = _blob_to_text(blob, creds)
with open(os.path.join(save_dir, blob["path"]), "wb") as f:
f.write(decoded_content)
print("DONE DOWNLOADING NAPKIN")
jsonl_of_path(save_dir, "books/napkin_train.jsonl", "books/napkin_val.jsonl",
"books-train", "books-valid")
def main():
creds = ("zhangir-azerbayev", os.environ["GITHUB_TOKEN"])
napkin(creds)
cring(creds)
stacks(creds)
mizar(creds)
afp(testing=False)
setmm(creds)
trench()
hott(creds)
stein(creds)
coq(creds)
lean(creds)
hol()
cam()
if __name__ == "__main__":
main()