LLMBB-Agent / agent /utils /doc_parser.py
ff_li
目录调整
f67d239
raw
history blame contribute delete
No virus
2.12 kB
import re
from agent.utils.tokenization_qwen import count_tokens
def rm_newlines(text):
text = re.sub(r'(?<=[^\.。::])\n', ' ', text)
return text
def rm_cid(text):
text = re.sub(r'\(cid:\d+\)', '', text)
return text
def rm_hexadecimal(text):
text = re.sub(r'[0-9A-Fa-f]{21,}', '', text)
return text
def deal(text):
text = rm_newlines(text)
text = rm_cid(text)
text = rm_hexadecimal(text)
return text
def parse_doc(path):
if '.pdf' in path.lower():
from langchain.document_loaders import PDFMinerLoader
loader = PDFMinerLoader(path)
pages = loader.load_and_split()
elif '.docx' in path.lower():
from langchain.document_loaders import Docx2txtLoader
loader = Docx2txtLoader(path)
pages = loader.load_and_split()
elif '.pptx' in path.lower():
from langchain.document_loaders import UnstructuredPowerPointLoader
loader = UnstructuredPowerPointLoader(path)
pages = loader.load_and_split()
else:
from langchain.document_loaders import UnstructuredFileLoader
loader = UnstructuredFileLoader(path)
pages = loader.load_and_split()
res = []
for page in pages:
dealed_page_content = deal(page.page_content)
res.append({
'page_content': dealed_page_content,
'token': count_tokens(dealed_page_content),
'metadata': page.metadata
})
return res
def pre_process_html(s):
# replace multiple newlines
s = re.sub('\n+', '\n', s)
# replace special string
s = s.replace("Add to Qwen's Reading List", '')
return s
def parse_html_bs(path):
from langchain.document_loaders import BSHTMLLoader
loader = BSHTMLLoader(path, open_encoding='utf-8')
pages = loader.load_and_split()
res = []
for page in pages:
dealed_page_content = pre_process_html(page.page_content)
res.append({
'page_content': dealed_page_content,
'token': count_tokens(dealed_page_content),
'metadata': page.metadata
})
return res