{
"paper_id": "I11-1036",
"header": {
"generated_with": "S2ORC 1.0.0",
"date_generated": "2023-01-19T07:31:28.380018Z"
},
"title": "CODACT: Towards Identifying Orthographic Variants in Dialectal Arabic",
"authors": [
{
"first": "Pradeep",
"middle": [],
"last": "Dasigi",
"suffix": "",
"affiliation": {
"laboratory": "",
"institution": "Columbia University",
"location": {}
},
"email": ""
},
{
"first": "Mona",
"middle": [],
"last": "Diab",
"suffix": "",
"affiliation": {},
"email": "mdiab@ccls.columbia.edu"
}
],
"year": "",
"venue": null,
"identifiers": {},
"abstract": "Dialectal Arabic (DA) is the spoken vernacular for over 300M people worldwide. DA is emerging as the form of Arabic written in online communication: chats, emails, blogs, etc. However, most existing NLP tools for Arabic are designed for processing Modern Standard Arabic, a variety that is more formal and scripted. Apart from the genre variation that is a hindrance for any language processing, even in English, DA has no orthographic standard, compared to MSA that has a standard orthography and script. Accordingly, a word may be written in many possible inconsistent spellings rendering the processing of DA very challenging. To solve this problem, such inconsistencies have to be normalized. This work is the first step towards addressing this problem, as we attempt to identify spelling variants in a given textual document. We present an unsupervised clustering approach that addresses the problem of identifying orthographic variants in DA. We employ different similarity measures that exploit string similarity and contextual semantic similarity. To our knowledge this is the first attempt at solving the problem for DA. Our approaches are tested on data in two dialects of Arabic-Egyptian and Levantine. Our system achieves the highest Entropy of 0.19 for Egyptian (corresponding to 68% cluster precision) and Levantine (corresponding to 64% cluster precision) respectively. This constitutes a significant reduction in entropy (from 0.47 for Egyptian and 0.51 for Levantine) and improvement in cluster precision (from 29% for both) from the baseline.",
"pdf_parse": {
"paper_id": "I11-1036",
"_pdf_hash": "",
"abstract": [
{
"text": "Dialectal Arabic (DA) is the spoken vernacular for over 300M people worldwide. DA is emerging as the form of Arabic written in online communication: chats, emails, blogs, etc. However, most existing NLP tools for Arabic are designed for processing Modern Standard Arabic, a variety that is more formal and scripted. Apart from the genre variation that is a hindrance for any language processing, even in English, DA has no orthographic standard, compared to MSA that has a standard orthography and script. Accordingly, a word may be written in many possible inconsistent spellings rendering the processing of DA very challenging. To solve this problem, such inconsistencies have to be normalized. This work is the first step towards addressing this problem, as we attempt to identify spelling variants in a given textual document. We present an unsupervised clustering approach that addresses the problem of identifying orthographic variants in DA. We employ different similarity measures that exploit string similarity and contextual semantic similarity. To our knowledge this is the first attempt at solving the problem for DA. Our approaches are tested on data in two dialects of Arabic-Egyptian and Levantine. Our system achieves the highest Entropy of 0.19 for Egyptian (corresponding to 68% cluster precision) and Levantine (corresponding to 64% cluster precision) respectively. This constitutes a significant reduction in entropy (from 0.47 for Egyptian and 0.51 for Levantine) and improvement in cluster precision (from 29% for both) from the baseline.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Abstract",
"sec_num": null
}
],
"body_text": [
{
"text": "Arabic is the native tongue of over 300M people world wide. The Arabic language exhibits a relatively unique linguistic phenomena known as diglossia (Ferguson, 1959) where two forms of the language live side by side: a standard formal form known as Modern Standard Arabic (MSA) and an informal spoken form, the vernaculars used in everyday communication referred to as Dialectal Arabic (DA). MSA is the only language of education and is used in formal settings and Broadcast news. The only written standard is in MSA using the Arabic script. Technically there are no native speakers of MSA. On the other hand, DA is the mother tongue for all native speakers of Arabic however it is not traditionally a written form of the language and it differs significantly enough from MSA on all levels of linguistic representation that results in huge inconsistencies in orthography. This was not a problem a decade ago from an NLP perspective since all the resources were in MSA. Now with the proliferation of online media and informal genres, DA is ubiquitous online. Users of DA online write in different scripts (Arabic, Romanizations interspersed with digits), they also sometimes write phonemically. Similar to other languages (not unique to DA) in these informal genres, we observe rampant speech effects such as elongations and the use of emoticons within the text which compounds the problem further for processing DA. If NLP tools want to process real Arabic as spoken by its people, they need to address DA seriously. This paper presents an initial attempt at addressing the pervasive inconsistencies in DA orthography in informal media.",
"cite_spans": [
{
"start": 149,
"end": 165,
"text": "(Ferguson, 1959)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "We cast the problem of lack of DA orthographic standards as an identification of spelling variants problem using unsupervised clustering techniques. We evaluate our results against a gold corrected set of data in two dialects: Egyptian (EGY) and Levantine (LEV) . We focus our current efforts on identifying the orthographic variants in Arabic script though our work is extendible to the Romanizations as well. Such an identification is a necessary step for normalizing the variation which is useful for addressing the sparseness problem for DA. We contend that there are patterns in the variations that could be captured and processed. Also it is worth pointing out that this problem encompasses the spelling mistakes problem but it goes beyond it to address legitimate orthographic variants. Hence we attempt an approach that is generic enough to cover both scopes.",
"cite_spans": [
{
"start": 246,
"end": 261,
"text": "Levantine (LEV)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "This paper is organized as follows: in Section 2 we show some of the variations between MSA and DA on different levels of linguistic representation; Section 3 discusses some related work; in section 4 we outline our approach and experimental conditions; in Section 5 we describe the data against which we evaluate our approach; we discuss the results and evaluation in Section 6; in Section 7 we discuss errors and performance of the system and approach proposed; finally, in Section 8 we conclude with some final remarks and a look at some future directions.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Introduction",
"sec_num": "1"
},
{
"text": "Most of the research effort, to date in creating tools and resources for Arabic has focused on MSA. In recent years we have seen a concentrated effort on making Arabic processing tools on par with English processing tools (Habash and Rambow., 2005; Diab et al., 2007; Kulick, 2010; Green and Manning., 2010) . Researchers interested in handling realistic Arabic text have come to the realization that DA needs to be addressed. Applying state of the art MSA processing tools directly to DA yields very low performance proving the significant difference between the two varieties. For instance applying MSA tokenizers to DA yields a performance of 88% which is completely unacceptable as an initial processing tool performance. It is worth noting that state of the art MSA tokenization is at 99.2% (Diab et al., 2007) . This low performance on DA can be explained by the genre differences (MSA tools are trained on newswire genres) but compared to English, we do not observe such a huge discrepancy between tokenizers trained on newswire when applied to informal genres. The significant drop in performance can be safely relegated to the inherent differences between the two varieties of Arabic. MSA differs from DA on the phonological, morphological, lexical, syntactic, semantic and pragmatic levels. The degree of variation depends on the specific dialect of Arabic. For instance, phonologically MSA would pronounce the word for dress as fustAn and spell it while the same word in LEV is pronounced as fusTAn 1 with an emphatic T and could possibly be written phonemically in LEV as . Morphologically, DA exhibits simpler inflectional morphology than MSA overall however cliticization is more nuanced rendering tokenization a more complex problem in DA than in MSA. For example, DA has lost all explicit marking of grammatical case and dual marking on verbal predicates. The MSA phrase AlmwZ-fAn AklA, , meaning 'the employees dual nominative ate dual', becomes AlmwZfyn AlAtnyn AkAlw, , 'the two employees plural no case ate plural'. Hence we note the loss of dual inflection marking and nominative case marking. On the other hand, clitization is more complex in DA as follows: EGY mAHkytl-hAlhw$, , 'she did not recount it to him' is expressed in three words in MSA as lm tHkyhA lh",
"cite_spans": [
{
"start": 222,
"end": 248,
"text": "(Habash and Rambow., 2005;",
"ref_id": null
},
{
"start": 249,
"end": 267,
"text": "Diab et al., 2007;",
"ref_id": "BIBREF5"
},
{
"start": 268,
"end": 281,
"text": "Kulick, 2010;",
"ref_id": "BIBREF7"
},
{
"start": 282,
"end": 307,
"text": "Green and Manning., 2010)",
"ref_id": null
},
{
"start": 796,
"end": 815,
"text": "(Diab et al., 2007)",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "DA vs. MSA Phenomena",
"sec_num": "2"
},
{
"text": ". The lexical, syntactic, semantic and pragmatic variations abound between MSA and DA. The phonological and morphological differences lend themselves directly to the orthographic variation problem exhibited with DA. Writers of DA use a myriad of scripts to encode DA. All of which are used inconsistently even within the writings of the same author in the same post/article/blog. The most frequent scripts used are Arabic and Romanization. We note that people have use also Hebrew and Cyrillic scripts to write Arabic as well. For Arabic script we see inconsistencies in characters that exhibit regional variations such as the qAf sound . This letter is pronounced as a glottal stop\u00b4in EGY and LEV but as a g sound in the Gulf states and q sound in Tunisia, in most cases. Speakers and writers pertaining to these different dialects could render it in the orthography as it appears in a word such as he said qAl as\u00c1l [EGY] , gAl [Gulf] or qAl [Tunisian] . Moreover, we observe more severe variants in the Romanized script for the same word where the writers can render the EGY as 2Al, AAl, or qAl. For the purposes of this paper we will focus our discussion on identifying the orthographic variants only in the Arabic script leaving the handling of orthographic variants in Romanization for future work. 2",
"cite_spans": [
{
"start": 917,
"end": 922,
"text": "[EGY]",
"ref_id": null
},
{
"start": 929,
"end": 935,
"text": "[Gulf]",
"ref_id": null
},
{
"start": 943,
"end": 953,
"text": "[Tunisian]",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "DA vs. MSA Phenomena",
"sec_num": "2"
},
{
"text": "Most of the recent work in the area of orthographic variant detection and spelling correction has been towards resolving inconsistencies in spellings of Named Entities(NE). Huang et al. 2008use NE spelling variant detection to improve the performance of Machine Translation (MT) and Information Extraction (IE) systems. (Habash and Metsky, 2008) cluster Urdu phrases mapping to the same English phrase to automatically learn morphological variation rules. Since Urdu is a morphologically rich language, such variations result in many OOV words. They use the these rules learned, as a part of MT system to replace OOV Urdu words with in-vocabulary words online. Raghavan and Allan (2005) use the edit distance metric along with generative models trained from Automatic Speech Recognition (ASR) output to cluster queries to improve performance of Information Retrieval (IR) systems. Bhagat and Hovy 2007attempted to generate all possible spelling variations of a person's name. One method is supervised and uses CMU speech dictionary to train a phonetic model. Another is to cluster a large set of names that are known to sound similar using Soundex (Knuth, 1973) . Although some earlier work related to spelling variations (Golding and Roth, 1999) dealt with the generalized problem, most of the recent work is confined to NEs.This is because of the relevance of the problem to NLP applications such as MT, IE, IR and ASR. Accordingly we note that the problem we try to solve is more generic since the lack of orthographic standard in DA affects the spellings of all kinds of words.",
"cite_spans": [
{
"start": 320,
"end": 345,
"text": "(Habash and Metsky, 2008)",
"ref_id": "BIBREF2"
},
{
"start": 1148,
"end": 1161,
"text": "(Knuth, 1973)",
"ref_id": "BIBREF3"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Related Work",
"sec_num": "3"
},
{
"text": "Our goal is to identify orthographic variations in textual DA. We build a system, CODACT, that aims at identifying and eventually normalizing such DA orthographic variants. We use techniques noted in the spelling correction literature. Our approach is mainly unsupervised. We view the problem as a clustering problem where our goal is to identify if two strings are similar, and hence cluster together. To that end we explore three basic similarity measures: (a) String based Similarity as direct Levenshtein Edit Distance; (b) String based Similarity Biased Edit Distance; and ( c) Contextual String Similarity. We model the strings of interest in a vector space. We build a matrix for the string types of interest. We induce the clusters from the matrix by grouping the strings in the row entries together based on the similarity of their respective vectors in the matrix. We use Cosine Similarity between vectors and we use the implementation of the CLUTO Repeated Bisection (RB) algorithm with cosine similarity being the measure of similarity between vectors. (Zhao and Karypis, 2001) . CLUTO is very suitable for clustering high dimensional datasets. CLUTO's repeated bisection partition method is used for clustering. In this method, for obtaining a k -way clustering, k-1 repeated bisections are made. Each partition is made to the input dataset such that the clustering criterion function is optimized.",
"cite_spans": [
{
"start": 1065,
"end": 1089,
"text": "(Zhao and Karypis, 2001)",
"ref_id": null
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Approach",
"sec_num": "4"
},
{
"text": "The row entries for the matrix are referred to as the focal string types of interest. We vary the dimensions as follows: (a) for N focal words, we have the same N focal words in the matrix dimensions, yielding an NxN matrix; or (b) the dimensions are all the string types in the corpus of interest yielding an NxM matrix. The cells of the matrix are populated based on one of the different similarity measures or a combination of them after normalization. We describe the different similarity measures next.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Approach",
"sec_num": "4"
},
{
"text": "Strings that vary from each other minimally are likely to be orthographic variants of one another. Following this intuition, strings are grouped based on their string edit distance. We explore the basic known Levenshtein Edit Distance measure as in (Levenshtein, 1966 ) (LEDM). Moreover we extend the LEDM to account for known phonological variations on the character level. We refer to this as the Biased Edit Distance Metric (BEDM). BEDM has the same exact formulation as LEDM as a metric however it is more relaxed in that it treats letters that are considered similar as if they are the same, i.e. they are not substitutions of each other. The intuition behind adding such a bias is the fact that Arabic letters may have different pronunciations depending on the context. For example, the letter might have a sound equivalent to the any of letters , , and . This is ignored by LEDM, and they are treated as different letters therefore incurring the substitution penalty. When BEDM is applied, any two letters that have the same sound are treated as a match. For example, (1) fstAn, 'dress', , and the possible variants (2) fstAn, (3) fSTAn, (4) fsTAn, and (5) fjTAn would have the following calculations: (1) and (2) would be a perfect match, i.e. a distance of 0 according to both LEDM and BEDM; (1) and (3) would be penalized for substituting S, T for s,t in (1), therefore a distance of 0.4 according to LEDM, however for BEDM s and S are considered similar to each other and so are t and T, then the distance of (1) and 3is 0. Similarily for (1) and (4) according to LEDM the distance is 0.2, but for BEDM the distance is 0. For (1) and (5) the LEDM will be 0.4 and for BEDM it will be 0.2. Hence, BEDM is a more nuanced and relaxed form of LEDM. The list of similar letters is taken from scholar seeded studies of phonological variations across different DA. The list is rendered in Table 1 . We refer to this list as sound change rules (SCR). 3 The SCR are not always symmetric, for example a v can be replaced with a S but not vice versa.",
"cite_spans": [
{
"start": 249,
"end": 267,
"text": "(Levenshtein, 1966",
"ref_id": "BIBREF8"
},
{
"start": 1953,
"end": 1954,
"text": "3",
"ref_id": null
}
],
"ref_spans": [
{
"start": 1892,
"end": 1899,
"text": "Table 1",
"ref_id": "TABREF0"
}
],
"eq_spans": [],
"section": "String Based Similarity Metrics",
"sec_num": "4.1"
},
{
"text": "We explore another relatedness measure of contextual string similarity (CSS). The intuition is that if two strings are variants of each other as they are semantically similar, they are bound to appear with similar contexts. Accordingly we model this notion via representing strings with their context co-occurrence vectors. In this framework , we represent the co-occurrence frequency of the focal string and dimensional string in all the sentences in the corpus within a window of 3 tokens. The observations are aggregated and used in the cell. If the focal string and the dimensional string never co-occur, then the cell value is set to 0. Contextual Similarity Metric (CSS) between two words is defined as a cosine similarity between their context vectors. ",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Contextual String Similarity",
"sec_num": "4.2"
},
{
"text": "A {, <, >, ', &, }, w, y, | ' A,{, }, <, >, |, y, &,w } A, y, &, ', {, <, >, |,w & A, y, }, ', {, <, >, |,w | A, y, ', {, <, >, } { A, y, &, ', }, <, >, | < A, ', {, >, |, } > A, ', {, <, |, } t T, v v s, t, S j q, y, $ H h, E d * , D * d, z, Z z * , Z, d s $, S, v $ s, v S s D Z, d, z, * T S, Z, t Z T, D, z, d, * E H g E, x q ', A, }, k, j k q h p, A p h, t w &, A, Y y }, A, Y Y y, A",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Letter Similar Sounding Letters",
"sec_num": null
},
{
"text": "We experimented with each of these measures in isolation and in combination. In the case of combination, we normalized the values of metrics. Table 2 illustrates the values contained in the cells of the constructed matrices in the different conditions.",
"cite_spans": [],
"ref_spans": [
{
"start": 142,
"end": 149,
"text": "Table 2",
"ref_id": "TABREF2"
}
],
"eq_spans": [],
"section": "Experimental Conditions",
"sec_num": "4.3"
},
{
"text": "We have two different matrix dimension sizes depending on how extensive the feature space is. The first case is NxN, meaning that the set of words corresponding to both the rows and columns of the matrix are the focal words. The second case has all the unique words in the corpus representing the columns, making it NxM. This yields 6 isolated conditions and 8 combined conditions. \u2022 LEDM-NxM where the similarity measure is a LEDM and the matrix size is NxM",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Experimental Conditions",
"sec_num": "4.3"
},
{
"text": "\u2022 BEDM-NxM where the similarity measure is a BEDM and the matrix size is NxM",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metric",
"sec_num": null
},
{
"text": "\u2022 CSS-NxM where the similarity measure is a CSS and the matrix size is NxM",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metric",
"sec_num": null
},
{
"text": "\u2022 LEDM-BEDM-NxN where the similarity measure is a combination of LEDM and BEDM and the matrix size is NxN",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metric",
"sec_num": null
},
{
"text": "\u2022 LEDM-CSS-NxN where the similarity measure is a combination of LEDM and CSS and the matrix size is NxN",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metric",
"sec_num": null
},
{
"text": "\u2022 BEDM-CSS-NxN where the similarity measure is a combination of CSS and BEDM and the matrix size is NxN",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metric",
"sec_num": null
},
{
"text": "\u2022 LEDM-BEDM-CSS-NxN where the similarity measure is a combination of LEDM, BEDM, and CSS the matrix size is NxN",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metric",
"sec_num": null
},
{
"text": "\u2022 LEDM-BEDM-NxM where the similarity measure is a combination of LEDM and BEDM and the matrix size is NxM",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metric",
"sec_num": null
},
{
"text": "\u2022 LEDM-CSS-NxM where the similarity measure is a combination of LEDM and CSS and the matrix size is NxM",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metric",
"sec_num": null
},
{
"text": "\u2022 BEDM-CSS-NxM where the similarity measure is a combination of CSS and BEDM and the matrix size is NxM",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metric",
"sec_num": null
},
{
"text": "\u2022 LEDM-BEDM-CSS-NxM where the similarity measure is a combination of LEDM, BEDM, and CSS the matrix size is NxM",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Metric",
"sec_num": null
},
{
"text": "In order to measure the performance of this approach we need data that has the variants identified. We created such data by asking native speakers of DA to normalize variants into a standard conventionalized form in Arabic script. We targeted two dialects of Arabic: Egyptian (EGY) and Levantine (LEV). For EGY we specifically focused on Cairene Egyptian. For Levantine, we had a collection of Palestinian, Jordanian, Lebanese and Syrian Dialectal data. Most of the data is considered Syrian in our LEV collection however. Both data sets for both DA are derived from the web (Diab et al., 2010) . The data is part of a larger collection we refer to in this paper as COM-MENTDA collection. COMMENTDA comprises 3M token strings for EGY and 3M token strings for LEV. For EGY we had 2 annotators and an adjudicator, and for LEV we had 4 annotators and an adjudicator. The annotators were instructed to identify tokens that are considered incorrect orthographically according to a specific convention that we devised known as CODA (Conventionalized Orthography for Dialectal Arabic) after being trained on CODA (Habash et al., 2011) . The annotators we asked to identify three different classes of variation from the CODA convention: (a) change in spelling of a string which included dealing with speech effects such as elongations, (b) introduction of spaces or splitting a string into multiple strings, and (c ) deletion of spaces or merging strings. Many corrections included simultaneously both a spelling change and a split or merge of a string as well. It is worth noting that roughly 24% of the EGY data had changes of different types on the token level corresponding to 37% of changes to the types for EGY. For LEV, only 11% of the tokens were changed corresponding to 23% of the types that were changed. This suggests that the EGY data had a lot more variability. It was actually noted that a lot of the EGY data was not consistently EGY but rather from other DA compared to LEV that was considered relatively homogeneous. The last row in the table shows the number of cases where annotators agreed with each other on the correction. It also shows their percentages to the number of common annotations as shown by the row above it.",
"cite_spans": [
{
"start": 575,
"end": 594,
"text": "(Diab et al., 2010)",
"ref_id": "BIBREF5"
},
{
"start": 1106,
"end": 1127,
"text": "(Habash et al., 2011)",
"ref_id": "BIBREF5"
}
],
"ref_spans": [],
"eq_spans": [],
"section": "Evaluation Data",
"sec_num": "5"
},
{
"text": "We created the gold data clusters of variants by grouping all the strings that are mapped to the same corrected CODA form. This data consisted of 290 clusters of strings in LEV with an average of 2.6 orthographic variants per cluster, and 312 string clusters in EGY with an average of 2.5 variant per cluster. All our experiments are conducted on surface forms of the strings with no preprocessing.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Evaluation Data",
"sec_num": "5"
},
{
"text": "In order to derive statistics to build our matrices, we use two data sets: COMMENTDA and an augmented data set (RCorpora) which is double the size of COMMENTDA for each dialect. RCorpora comprises 6M strings for EGY, and 6M strings for LEV. In the NxN matrix conditions, the size of the corpora used to derive the statistics only affects the conditions involving CSS. In the NxM conditions, the corpora sizes affect the number of matrix dimensions as well as the cell values for the CSS conditions. Application of seven metrics to four class-data size combinations gives 25 distinct runs per dialect since in NxN case, augmenting the data does not change the metrics LEDM, BEDM, and LEDM+BEDM. It has to be noted that for NxM experiments this is not the case. Table 4 gives the various statistics on the different data sizes of unique string types.",
"cite_spans": [],
"ref_spans": [
{
"start": 760,
"end": 767,
"text": "Table 4",
"ref_id": null
}
],
"eq_spans": [],
"section": "Evaluation",
"sec_num": "6"
},
{
"text": "Each clustering output is compared with the gold-standard clusters using Purity and Entropy measures (Zhao and Karypis, 2001) . Every word in a given cluster in the output belongs to one or more gold clusters. These gold clusters are referred to as relevant gold clusters of the given output cluster. Purity or precision of a cluster is the fraction of its words in its relevant gold clusters. Entropy gives the measure of ambiguity in the clustering output. The larger the number of word in a relevant gold cluster, the higher the entropy value. In addition to these measures, the value of recall is also calculated. This equals the fraction of words in the relevant gold clusters that are in the given cluster. Together these three measures give a complete assessment of the quality of the output clusters. As a baseline for comparison, where strings are randomly assigned to clusters assuming the gold number of clusters per dialect. Table 5 shows the results of all the experiments described.",
"cite_spans": [
{
"start": 101,
"end": 125,
"text": "(Zhao and Karypis, 2001)",
"ref_id": null
}
],
"ref_spans": [
{
"start": 937,
"end": 944,
"text": "Table 5",
"ref_id": "TABREF7"
}
],
"eq_spans": [],
"section": "Evaluation",
"sec_num": "6"
},
{
"text": "COMMENTDA COMMENTDA+RCorpora (C+R) EGY LEV EGY LEV NxN 729 717 729 717 NxM 205088 237598 433697 410206 Table 4 : Data Size Variations in the two dimensions conditions",
"cite_spans": [],
"ref_spans": [
{
"start": 47,
"end": 119,
"text": "LEV NxN 729 717 729 717 NxM 205088 237598 433697 410206 Table 4",
"ref_id": "TABREF0"
}
],
"eq_spans": [],
"section": "Evaluation",
"sec_num": "6"
},
{
"text": "All the systems outperform the random baseline.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "7"
},
{
"text": "The best results are presented in bold in Table 5 .",
"cite_spans": [],
"ref_spans": [
{
"start": 42,
"end": 49,
"text": "Table 5",
"ref_id": "TABREF7"
}
],
"eq_spans": [],
"section": "Discussion",
"sec_num": "7"
},
{
"text": "Phonological Bias From the results, it can be seen that of the individual metrics, BEDM consistently performs better than LEDM. This shows that introducing a phonological bias while matching letters does have a significant positive effect in identifying spelling variants. We believe that this effect will be more pronounced if the SCR are more tailored to the specific dialect under study. We note that CSS has the worst performance of the three individual metrics in all cases. This may be attributed to the level of processing of the data. The data is dealt with specifically on the surface level and Arabic being a very rich morphological language results in a very sparse distribution of forms. This can be mitigated by using even larger corpora. Typically in such studies that rely on distributional similarity an order of magnitude larger than what we employed is exploited. We relegate this to future work.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "7"
},
{
"text": "Combined Metrics Combination of LEDM and BEDM showed better precision, recall, and entropy than each of them in isolation in every case for both dialects. Although adding CSS has shows improvement in LEDM, mainly in EGY, it actually worsened the performance of BEDM. This is more evident when CSS+LEDM+BEDM is compared with LEDM+BEDM. Almost all the quality measures show that CSS metric adds noise.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "7"
},
{
"text": "NxM vs. NxN Using a bigger class of words as a feature set for vector similarity significantly improves the performance of the system. This holds for all the metrics in both the dialects.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "7"
},
{
"text": "Data Augmentation Although there are slight improvements due to increase in data, on the whole this does not seem to affect the performance of the system significantly. We only see some effect in the CSS measures which is expected.",
"cite_spans": [],
"ref_spans": [],
"eq_spans": [],
"section": "Discussion",
"sec_num": "7"
},
{
"text": "Cluster type Members Gold "
},
"TABREF2": {
"type_str": "table",
"num": null,
"html": null,
"text": "Matrix Cell Values",
"content": "
\u2022 LEDM-NxN where the similarity measure is a LEDM and the matrix size is NxN |
\u2022 BEDM-NxN where the similarity measure is a BEDM and the matrix size is NxN |
\u2022 CSS-NxN where the similarity measure is a CSS and the matrix size is NxN |
"
},
"TABREF3": {
"type_str": "table",
"num": null,
"html": null,
"text": "",
"content": ""
},
"TABREF4": {
"type_str": "table",
"num": null,
"html": null,
"text": "",
"content": ""
},
"TABREF7": {
"type_str": "table",
"num": null,
"html": null,
"text": "Results",
"content": ""
}
}
}
}