Commit
·
05a6587
1
Parent(s):
c7ad8d9
yolo
Browse files- output_dataset.parquet +2 -2
- process.py +34 -43
- viz.py +3 -3
output_dataset.parquet
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4d75658b3f088a0b4a93eb5aedd74199e3b8428d132565088b9ed11aee069213
|
3 |
+
size 22498564
|
process.py
CHANGED
@@ -4,11 +4,20 @@ import pyarrow as pa
|
|
4 |
import pyarrow.parquet as pq
|
5 |
import argparse
|
6 |
import re
|
|
|
7 |
|
8 |
-
def
|
9 |
-
"""
|
10 |
-
|
11 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
12 |
|
13 |
def extract_images(markdown_content):
|
14 |
"""Extract PHOTO_IDs from markdown files and return as a list."""
|
@@ -16,55 +25,37 @@ def extract_images(markdown_content):
|
|
16 |
|
17 |
def collect_data(directory):
|
18 |
data = {}
|
19 |
-
|
20 |
-
|
21 |
-
for filename in os.listdir(directory):
|
22 |
-
if filename.endswith('.jpg'):
|
23 |
-
photo_id = re.search(r'(\d+)', filename)
|
24 |
-
if photo_id:
|
25 |
-
image_files[photo_id.group(1)] = os.path.join(directory, filename)
|
26 |
|
27 |
-
# Identify
|
28 |
-
for
|
29 |
-
|
30 |
-
|
31 |
-
|
32 |
-
|
|
|
33 |
|
34 |
-
#
|
35 |
for filename in os.listdir(directory):
|
36 |
-
|
37 |
-
|
38 |
-
|
39 |
-
|
40 |
-
|
41 |
-
if extension == 'jpg':
|
42 |
-
# Load binary data instead of encoding
|
43 |
-
content = load_binary(file_path)
|
44 |
-
data[base_name]['Images'].append(content)
|
45 |
-
else:
|
46 |
-
try:
|
47 |
-
with open(file_path, 'r', encoding='utf-8') as file:
|
48 |
-
content = file.read()
|
49 |
-
except UnicodeDecodeError as e:
|
50 |
-
print(f"Error decoding file {file_path}: {e}")
|
51 |
-
continue
|
52 |
|
53 |
-
|
54 |
-
|
55 |
-
|
56 |
image_ids = extract_images(content)
|
57 |
-
data[
|
58 |
-
|
59 |
-
data[base_name][extension] = content
|
60 |
|
61 |
return list(data.values())
|
62 |
|
63 |
def create_parquet_file(data, output_file):
|
64 |
df = pd.DataFrame(data)
|
65 |
-
|
66 |
-
df['Images'] = df['Images'].apply(lambda x: [memoryview(b) for b in x])
|
67 |
-
table = pa.Table.from_pandas(df, preserve_index=False)
|
68 |
pq.write_table(table, output_file)
|
69 |
|
70 |
def main():
|
|
|
4 |
import pyarrow.parquet as pq
|
5 |
import argparse
|
6 |
import re
|
7 |
+
import base64
|
8 |
|
9 |
+
def encode_file(file_path):
|
10 |
+
"""Encode text files or base64 encode image files."""
|
11 |
+
if file_path.endswith('.jpg'):
|
12 |
+
with open(file_path, "rb") as image_file:
|
13 |
+
return base64.b64encode(image_file.read()).decode('utf-8')
|
14 |
+
else:
|
15 |
+
try:
|
16 |
+
with open(file_path, 'r', encoding='utf-8') as file:
|
17 |
+
return file.read()
|
18 |
+
except UnicodeDecodeError as e:
|
19 |
+
print(f"Error decoding file {file_path}: {e}")
|
20 |
+
return None
|
21 |
|
22 |
def extract_images(markdown_content):
|
23 |
"""Extract PHOTO_IDs from markdown files and return as a list."""
|
|
|
25 |
|
26 |
def collect_data(directory):
|
27 |
data = {}
|
28 |
+
image_files = {re.search(r'(\d+)', filename).group(1): filename
|
29 |
+
for filename in os.listdir(directory) if filename.endswith('.jpg')}
|
|
|
|
|
|
|
|
|
|
|
30 |
|
31 |
+
# Identify Markdown and solution Markdown files and establish correct problem IDs
|
32 |
+
markdown_files = [f for f in os.listdir(directory) if f.endswith('.md') or f.endswith('.sol.md')]
|
33 |
+
for mfile in markdown_files:
|
34 |
+
# Adjust the pattern if problem IDs include characters before "sol"
|
35 |
+
problem_id = re.sub(r'sol$', '', mfile.split('.')[0]) # Strip "sol" from end
|
36 |
+
if problem_id not in data:
|
37 |
+
data[problem_id] = {'Problem ID': problem_id, 'Images': [], 'in': None, 'out': None, 'cpp': None, 'md': None, 'sol.md': None}
|
38 |
|
39 |
+
# Now associate other files with these problem IDs
|
40 |
for filename in os.listdir(directory):
|
41 |
+
problem_id = re.sub(r'sol$', '', filename.split('.')[0]) # Strip "sol" from end if present
|
42 |
+
if problem_id in data: # Only process if the problem_id is recognized
|
43 |
+
file_type = filename.split('.')[-1]
|
44 |
+
file_path = os.path.join(directory, filename)
|
45 |
+
content = encode_file(file_path) if not filename.endswith('.jpg') else None
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
46 |
|
47 |
+
if file_type in ['in', 'out', 'cpp', 'md', 'sol.md']:
|
48 |
+
data[problem_id][file_type] = content
|
49 |
+
if file_type in ['md', 'sol.md']:
|
50 |
image_ids = extract_images(content)
|
51 |
+
data[problem_id]['Images'] += [image_files[id] for id in image_ids if id in image_files]
|
52 |
+
data[problem_id]['Images'] = list(set(data[problem_id]['Images'])) # Remove duplicates
|
|
|
53 |
|
54 |
return list(data.values())
|
55 |
|
56 |
def create_parquet_file(data, output_file):
|
57 |
df = pd.DataFrame(data)
|
58 |
+
table = pa.Table.from_pandas(df)
|
|
|
|
|
59 |
pq.write_table(table, output_file)
|
60 |
|
61 |
def main():
|
viz.py
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
import pandas as pd
|
2 |
-
pd.set_option('display.max_columns', None) # None means no limit
|
3 |
-
pd.set_option('display.width', None) # None means use the current terminal width
|
4 |
|
5 |
# Load the Parquet file
|
6 |
df = pd.read_parquet('output_dataset.parquet')
|
@@ -8,7 +8,7 @@ df = pd.read_parquet('output_dataset.parquet')
|
|
8 |
print(df.columns)
|
9 |
|
10 |
# Display the first few rows of the dataframe
|
11 |
-
print(df.loc[
|
12 |
|
13 |
|
14 |
# Basic statistics for numerical columns
|
|
|
1 |
import pandas as pd
|
2 |
+
# pd.set_option('display.max_columns', None) # None means no limit
|
3 |
+
# pd.set_option('display.width', None) # None means use the current terminal width
|
4 |
|
5 |
# Load the Parquet file
|
6 |
df = pd.read_parquet('output_dataset.parquet')
|
|
|
8 |
print(df.columns)
|
9 |
|
10 |
# Display the first few rows of the dataframe
|
11 |
+
print(df.loc[0, "cpp"])
|
12 |
|
13 |
|
14 |
# Basic statistics for numerical columns
|