|
|
|
import asyncio |
|
import pathlib |
|
|
|
import httpx |
|
import markdownify |
|
import orjson |
|
from bs4 import BeautifulSoup |
|
|
|
root = pathlib.Path( |
|
"data" |
|
) |
|
root.mkdir(exist_ok=True, parents=True) |
|
|
|
|
|
class md_nolinks(markdownify.MarkdownConverter): |
|
def convert_a(self, _, text, __): |
|
_, _, text = markdownify.chomp(text) |
|
if not text: |
|
return "" |
|
return text |
|
|
|
def convert_img(self, el, text, convert_as_inline): |
|
return "" |
|
|
|
|
|
md = md_nolinks() |
|
|
|
|
|
async def index(): |
|
session = httpx.AsyncClient() |
|
session.headers["user-agent"] = ( |
|
"Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Safari/537.36" |
|
) |
|
if not (root / pathlib.Path("dev_index.json")).exists(): |
|
dev_urls = set() |
|
for idx in range(0, 20): |
|
c = await session.get( |
|
f"https://devopedia.org/site-map/list-articles?page={idx}&action=next-page&tag=0&ajax=1" |
|
) |
|
soup = BeautifulSoup(c.content, "lxml") |
|
for href in [ |
|
a["href"] for a in soup.select("div.dev-events div.uk-panel > div > a") |
|
]: |
|
dev_urls.add(f"https://devopedia.org{href}") |
|
(root / pathlib.Path("dev_index.json")).write_bytes( |
|
orjson.dumps(list(dev_urls), option=orjson.OPT_INDENT_2) |
|
) |
|
else: |
|
dev_urls = orjson.loads(pathlib.Path("dev_index.json").read_bytes()) |
|
final_content = [] |
|
for url in dev_urls: |
|
c = await session.get(url, timeout=None) |
|
soup = BeautifulSoup(c.text, "lxml") |
|
main = soup.select_one("main div.article-middle") |
|
print(url) |
|
dev_content = [] |
|
if main: |
|
h1 = soup.find("h1", attrs={"class": "uk-article-title"}) |
|
for section in main.select("section"): |
|
[i.decompose() for i in section.find_all("figure")] |
|
[ |
|
i.decompose() |
|
for i in section.find_all("sup", attrs={"class": "inline-citation"}) |
|
] |
|
|
|
if section.get("id") and "summary" in section.get("id").lower(): |
|
fmt = f"## Summary\n\n{md.convert_soup(section).rstrip()}" |
|
dev_content.append(fmt) |
|
if section.get("id") and "discussion" in section.get("id").lower(): |
|
z = "## Discussion" |
|
for qa in section.find("ul", recursive=False).find_all( |
|
"li", recursive=False |
|
): |
|
q = qa.find("article-question") |
|
a = qa.find("article-answer") |
|
fmt = f"### {q.get_text()}\n\n{md.convert_soup(a)}" |
|
z += f"\n\n{fmt}" |
|
dev_content.append(z) |
|
if section.get("id") and "milestone" in section.get("id").lower(): |
|
section.find("h2").decompose() |
|
fmt = f"\n\n## Milestones\n\n{md.convert_soup(section).strip()}" |
|
dev_content.append(fmt) |
|
final_content.append( |
|
orjson.dumps( |
|
{ |
|
"text": f"# {h1.get_text()}\n\n" |
|
+ ("\n\n".join(dev_content)) |
|
.replace("\n\n\n", "\n\n") |
|
.replace("\n\n\n\n", "\n\n") |
|
.replace("\r\n", "\n") |
|
.replace("\t", " "), |
|
"meta": { |
|
"title": h1.get_text(), |
|
"href": f"{url.split('/')[-1]}", |
|
}, |
|
} |
|
) |
|
) |
|
else: |
|
raise Exception |
|
pathlib.Path("dev_files.jsonl").write_bytes(b"\n".join(final_content)) |
|
|
|
|
|
if __name__ == "__main__": |
|
asyncio.run(index()) |
|
|