Spaces:
Sleeping
Sleeping
victormiller
commited on
Commit
•
9054008
1
Parent(s):
2973f7e
Update overview.py
Browse files- overview.py +21 -0
overview.py
CHANGED
@@ -209,14 +209,35 @@ dataset_sources = pd.DataFrame(
|
|
209 |
table_html = dataset_sources.to_html(index=False, border=0)
|
210 |
table_div1 = Div(NotStr(table_html), style="margin: 40px;")
|
211 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
212 |
def overview():
|
213 |
return Div(Section(
|
|
|
|
|
|
|
|
|
|
|
214 |
H2("Combining the Best of Web and Curated Sources"),
|
215 |
H3("Why combine the web and highly curated sources? Isn't the web-only data enough?"),
|
216 |
P("Table 1: TxT360 combines both the web data and highly-curated sources, which none of the existing datasets have covered. The following table shows TxT360 and other well-known datasets on the coverage and size of data sources."),
|
217 |
table_div,
|
218 |
P("Table 2: Statistics of TxT360. The basic statistics of TxT360 are presented."),
|
219 |
table_div1,
|
|
|
|
|
|
|
220 |
id="inner-text",
|
221 |
)
|
222 |
)
|
|
|
209 |
table_html = dataset_sources.to_html(index=False, border=0)
|
210 |
table_div1 = Div(NotStr(table_html), style="margin: 40px;")
|
211 |
|
212 |
+
quality_text = P("""The quality and size of a pre-training dataset play a crucial role in the performance of large language models (LLMs).
|
213 |
+
The community has introduced a variety of datasets for this purpose, including purely web-based datasets like RefinedWeb{citation_obj.display_citation("refinedweb")}, RedPajama-Data-V2{citation_obj.display_citation("redpajama-v2")}, DCLM{citation_obj.display_citation("dclm")}, and FineWeb{citation_obj.display_citation("fineweb")},
|
214 |
+
as well as comprehensive datasets derived from multiple highly-curated data sources such as The Pile{citation_obj.display_citation("thepile")}, RedPajama-Data-V1{citation_obj.display_citation("redpajama-v1")}, and Dolma{citation_obj.display_citation("dolma")}.
|
215 |
+
It is commonly known that web-based datasets provide a vast quantity of data, while highly-curated multi-source datasets consistently deliver high quality and diversity,
|
216 |
+
both critical for effective LLM pre-training.""")
|
217 |
+
|
218 |
+
quality_text2 = P("However, despite the advancements in both types of data, each type of dataset has its limitations. For instance, the processing scripts for the web dataset, RefinedWeb, known for its high quality, are not public, and only about 10% of the entire dataset has been disclosed. Conversely, the web component of existing highly-curated multi-source datasets is relatively small compared to purely web-based datasets, limiting their coverage and diversity compared to the scale of information from the internet.")
|
219 |
+
quality_text3 = P("By integrating the extensive reach of web data with the exceptional quality of curated sources, TxT360 is crafted to meet and surpass the rigorous standards required for state-of-the-art LLM pre-training.")
|
220 |
+
|
221 |
+
data_processing_image_desc = P("Figure 1: Data processing pipeline. All the steps are adopted for processing web data while the yellow blocks are adopted for processing curated sources.")
|
222 |
+
data_processing_overview = P("We enforce a fully transparent data processing pipeline when producing TxT360, designed to handle both web and curated datasets with precision and clarity. This transparent pipeline presents a unified framework for processing both data types, making it convenient and adaptive for users to revise and fine-tune the pipeline. ")
|
223 |
+
data_processing_overview2 = P("For web datasets, the pipeline focuses on extracting meaningful, high-quality text from raw web content, which is inherently noisy and varied. The process includes sophisticated filtering and deduplication techniques to clean the data and remove any redundancies or irrelevant information. On the other hand, curated datasets, which are already more structured and reliable, are processed with selective steps to maintain their integrity while integrating them seamlessly into the larger dataset.")
|
224 |
+
data_processing_overview3 = P("We will open-source the scripts for the whole pipeline, allowing the community to review, replicate, and build upon our processes.")
|
225 |
def overview():
|
226 |
return Div(Section(
|
227 |
+
H2("Our General Appoach to Data Processing"),
|
228 |
+
data_processing_overview,
|
229 |
+
data_processing_overview2,
|
230 |
+
data_processing_overview3,
|
231 |
+
data_processing_image_desc,
|
232 |
H2("Combining the Best of Web and Curated Sources"),
|
233 |
H3("Why combine the web and highly curated sources? Isn't the web-only data enough?"),
|
234 |
P("Table 1: TxT360 combines both the web data and highly-curated sources, which none of the existing datasets have covered. The following table shows TxT360 and other well-known datasets on the coverage and size of data sources."),
|
235 |
table_div,
|
236 |
P("Table 2: Statistics of TxT360. The basic statistics of TxT360 are presented."),
|
237 |
table_div1,
|
238 |
+
quality_text,
|
239 |
+
quality_text2,
|
240 |
+
quality_text3,
|
241 |
id="inner-text",
|
242 |
)
|
243 |
)
|