diff --git "a/txt/2202.13056.txt" "b/txt/2202.13056.txt" deleted file mode 100644--- "a/txt/2202.13056.txt" +++ /dev/null @@ -1,1427 +0,0 @@ -1 -Automated Identification of Toxic Code Reviews Using ToxiCR -JAYDEB SARKER, ASIF KAMAL TURZO, MING DONG, AMIANGSHU BOSU, Wayne State -University, USA -Toxic conversations during software development interactions may have serious repercussions on a Free and Open -Source Software (FOSS) development project. For example, victims of toxic conversations may become afraid -to express themselves, therefore get demotivated, and may eventually leave the project. Automated filtering of -toxic conversations may help a FOSS community to maintain healthy interactions among its members. However, -off-the-shelf toxicity detectors perform poorly on Software Engineering (SE) dataset, such as one curated from -code review comments. To encounter this challenge, we present ToxiCR, a supervised learning-based toxicity -identification tool for code review interactions. ToxiCR includes a choice to select one of the ten supervised learning -algorithms, an option to select text vectorization techniques, eight preprocessing steps, and a large scale labeled -dataset of 19,651 code review comments. Two out of those eight preprocessing steps are SE domain specific. With -our rigorous evaluation of the models with various combinations of preprocessing steps and vectorization techniques, -we have identified the best combination for our dataset that boosts 95.8% accuracy and 88.9% 𝐹11score. ToxiCR -significantly outperforms existing toxicity detectors on our dataset. We have released our dataset, pretrained -models, evaluation results, and source code publicly available at: https://github.com/WSU-SEAL/ToxiCR). -CCS Concepts: •Software and its engineering →Collaboration in software development ;Integrated and visual -development environments ;•Computing methodologies →Supervised learning . -Additional Key Words and Phrases: toxicity, code review, sentiment analysis, natural language processing, tool -development -ACM Reference Format: -Jaydeb Sarker, Asif Kamal Turzo, Ming Dong, Amiangshu Bosu. 2023. Automated Identification of Toxic -Code Reviews Using ToxiCR. ACM Trans. Softw. Eng. Methodol. 32, 1, Article 1 (January 2023), 33 pages. -https://doi.org/10.1145/3583562 -1 INTRODUCTION -Communications among the members of many Free and Open Source Software(FOSS) communities -include manifestations of toxic behaviours [ 12,32,46,62,66,81]. These toxic communications may have -decreased the productivity of those communities by wasting valuable work hours [ 16,73]. FOSS developers -being frustrated over peers with ‘prickly’ personalities [ 17,34] may contemplate leaving a community for -good [11,26]. Moreover, as most FOSS communities rely on contributions from volunteers, attracting and -retaining prospective joiners is crucial for the growth and survival of FOSS projects [ 72]. However, toxic -interactions with existing members may pose barriers against successful onboarding of newcomers [ 48,83]. -Therefore, it is crucial for FOSS communities to proactively identify and regulate toxic communications. -Author’s address: Jaydeb Sarker, Asif Kamal Turzo, Ming Dong, Amiangshu Bosu, jaydebsarker,asifkamal,mdong,amiangshu. -bosu, Wayne State University, Detroit, Michigan, USA. -Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee -provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and -the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be -honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, -requires prior specific permission and/or a fee. Request permissions from permissions@acm.org. -©2023 Copyright held by the owner/author(s). Publication rights licensed to ACM. -1049-331X/2023/1-ART1 $15.00 -https://doi.org/10.1145/3583562 -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.arXiv:2202.13056v3 [cs.SE] 8 Feb 20231:2•Sarker, et al. -Large-scale FOSS communities, such as Mozilla, OpenStack, Debian, and GNU manage hundreds of -projects and generate large volumes of text-based communications among their contributors. Therefore, -it is highly time-consuming and infeasible for the project administrators to identify and timely intervene -with ongoing toxic communications. Although, many FOSS communities have codes of conduct, those are -rarely enforced due to time constraints [ 11]. As a result, toxic interactions can be easily found within -the communication archives of many well-known FOSS projects. As an anonymous FOSS developer -wrote after leaving a toxic community, “ ..it’s time to do a deep dive into the mailing list archives or -chat logs. ... Searching for terms that degrade women (chick, babe, girl, bitch, cunt), homophobic slurs -used as negative feedback (“that’s so gay”), and ableist terms (dumb, retarded, lame), may allow you to -get a sense of how aware (or not aware) the community is about the impact of their language choice on -minorities. ” [11]. Therefore, it is crucial to develop an automated tool to identify toxic communication of -FOSS communities. -Toxic text classification is a Natural Language Processing (NLP) task to automatically classify a text -as ‘toxic’ or ‘non-toxic’. There are several state-of-the-art tools to identify toxic contents in blogs and -tweets [7,9,39,54]. However, off-the-shelf toxicity detectors do not work well on Software Engineering -(SE) communications [ 77], since several characteristics of such communications (e.g., code reviews and -bug interactions) are different from those of blogs and tweets. For example, compared to code review -comments, tweets are shorter and are limited to a maximum length. Tweets rarely include SE domain -specific technical jargon, URLs, or code snippets [ 6,77]. Moreover, due to different meanings of some -words (e.g, ‘kill’, ‘dead’, and ‘dumb’) in the SE context, SE communications with such words are often -incorrectly classified as ‘toxic’ by off-the-shelf toxicity detectors [73, 77]. -To encounter this challenge, Raman etal. developed a toxicity detector tool (referred as the ‘STRUDEL -tool’ hereinafter) for the SE domain [ 73]. However, as the STRUDEL tool was trained and evaluated -with only 611 SE texts. Recent studies have found that it performed poorly on new samples [ 59,70]. To -further investigate these concerns, Sarker et al. conducted a benchmark study of the STRUDEL tool and -four other off-the-shelf toxicity detectors using two large scale SE datasets [ 77]. To develop their datasets, -they empirically developed a rubric to determine which SE texts should be placed in the ‘toxic’ group -during their manual labeling. Using that rubric, they manually labeled a dataset of 6,533 code review -comments and 4,140 Gitter messages [ 77]. The results of their analyses suggest that none of the existing -tools are reliable in identifying toxic texts from SE communications, since all the five tools’ performances -significantly degraded on their SE datasets. However, they also found noticeable performance boosts (i.e., -accuracy improved from 83% to 92% and F-score improved from 40% to 87%) after retraining two of the -existing off-the-shelf models (i.e., DPCNN [ 94] and BERT with FastAI [ 54]) using their datasets. Being -motivated by these results, we hypothesize that a SE domain specific toxicity detector can boost even -better performances, since off-the-shelf toxicity detectors do not use SE domain specific preprocessing -steps, such as preprocessing of code snippets included within texts. On this hypothesis, this paper presents -ToxiCR, a SE domain specific toxicity detector. ToxiCR is trained and evaluated using a manually labeled -dataset of 19,651 code review comments selected from four popular FOSS communities (i.e., Android, -Chromium OS, OpenStack and LibreOffice). We selected code review comments, since a code review -usually represents a direct interaction between two persons (i.e., the author and a reviewer). Therefore, a -toxic code review comment has the potential to be taken as a personal attack and may hinder future -collaboration between the participants. ToxiCR is written in Python using the Scikit-learn [ 67] and -TensorFlow [ 3]. It provides an option to train models using one of the ten supervised machine learning -algorithms including five classical and ensemble-based, four deep neural network-based, and a Bidirectional -Encoder Representations from Transformer (BERT) based ones. It also includes eight preprocessing -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:3 -steps with two being SE domain specific and an option to choose from five different text vectorization -techniques. -We empirically evaluated various optional preprocessing combinations for each of the ten algorithms -to identify the best performing combination. During our 10-fold cross-validations evaluations, the best -performing model of ToxiCR significantly outperforms existing toxicity detectors on the code review -dataset with an accuracy of 95.8% and F-score of 88.9%. -The primary contributions of this paper are: -∙ToxiCR: An SE domain specific toxicity detector. ToxiCR is publicly available on Github at: -https://github.com/WSU-SEAL/ToxiCR. -∙An empirical evaluation of ten machine learning algorithms to identify toxic SE communications. -∙Implementations of eight preprocessing steps including two SE domain specific ones that can be -added to model training pipelines. -∙An empirical evaluation of three optional preprocessing steps in improving the performances of -toxicity classification models. -∙Empirical identification of the best possible combinations for all the ten algorithms. -Paper organization: The remainder of this paper is organized as following. Section 2 provides a brief -background and discusses prior related works. Section 3 discusses the concepts utilized in designing -ToxiCR. Section 4 details the design of ToxiCR. Section 5 details the results of our empirical evaluation. -Section 6 discusses the lessons learned based on this study. Section 7 discusses threats to validity of our -findings. Finally, Section 8 provides a future direction based on this work and concludes this paper. -2 BACKGROUND -This section defines toxic communications, provides a brief overview of prior works on toxicity in FOSS -communities and describes state-of-the-art toxicity detectors. -2.1 What constitutes a toxic communication? -Toxicity is a complex phenomenon to construct as it is highly subjective than other text classification -problems (i.g., online abuse, spam) [ 53]. Whether a communication should be considered as ‘toxic’ also -depends on a multitude of factors, such as communication medium, location, culture, and relationship -between the participants. In this research, we focus specially on written online communications. According -to the Google Jigsaw AI team, a text from an online communication can be marked as toxic if it contains -disrespectful or rude comments that make a participant to leave the discussion forum [ 7]. On the other -hand, the Pew Research Center marks a text as toxic if it contains threat, offensive call, or sexually -expletive words [ 28]. Anderson etal.’s definition of toxic communication also includes insulting language -or mockery [ 10]. Adinolf and Turkay studied toxic communication in online communities and their views -of toxic communications include harassment, bullying, griefing (i.e, constantly making other players -annoyed), and trolling [ 5]. To understand, how persons from various demographics perceive toxicity, -Kumaretal. conducted a survey with 17,280 participants inside the USA. To their surprise, their results -indicate that the notion toxicity cannot be attributed to any single demographic factor [ 53]. According to -Miller et al., various antisocial behaviors fit inside the Toxicity umbrella such as hate speech, trolling, -flaming, and cyberbullying [ 59]. While some of the SE studies have investigated antisocial behaviors -among SE communities using the ‘toxicity’ construct [ 59,73,77], other studies have used various other -lens such as incivility [ 33], pushback [ 29], and destructive criticism [ 40]. Table 1 provides a brief overview -of the studied constructs and their definitions. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:4•Sarker, et al. -SE study Construct Definition -Sarker et -al. [77]Toxicity “includes any of the following: i) offensive name calling, ii) insults, -iii) threats, iv) personal attacks, v) flirtations, vi) reference to -sexual activities, and vii) swearing or cursing. ” -Miller et -al. [59]Toxicity “an umbrella term for various antisocial behaviors including -trolling, flaming, hate speech, harassment, arrogance, entitlement, -and cyberbullying ”. -Ferreira et -al. [33]Incivility “features of discussion that convey an unnecessarily disrespectful -tone toward the discussion forum, its participants, or its topics ” -Gunawardena -et al. [40]Destructive -criticismnegative feedback which is nonspecific and is delivered in a harsh or -sarcastic tone, includes threats, or attributes poor task performance -to flaws of the individual. -Egelman et -al. [29]Pushback “the perception of unnecessary interpersonal conflict in code review -while a reviewer is blocking a change request ” -Table 1. Anti-social constructs investigated in prior SE studies -2.2 Toxic communications in FOSS communities -Several prior studies have identified toxic communications in FOSS communities [ 21,66,73,77,81]. Squire -and Gazda found occurrences of expletives and insults in publicly available IRC and mailing-list archives of -top FOSS communities, such as Apache, Debian, Django, Fedora, KDE, and Joomla [ 81]. More alarmingly, -they identified sexist ‘maternal insults’ being used by many developers. Recent studies have also reported -toxic communications among issue discussions on Github [73], and during code reviews [33, 40, 66, 77]. -Although toxic communications are rare in FOSS communities [ 77], toxic interactions can have severe -consequences [ 21]. Carillo etal. termed Toxic communications as a ‘poison’ that impacts mental health -of FOSS developers [ 21], and may contribute to stress and burnouts [ 21,73]. When the level of toxicity -increases in a FOSS community, the community may disintegrate as developers may no longer wish -to be associated with that community [ 21]. Moreover, toxic communications hamper onboarding of -prospective joiners, as a newcomer may get turned off by the signs of a toxic culture prevalent in a -FOSS community [ 48,83]. Milleretal. conducted a qualitative study to better understand toxicity in -the context of FOSS development [ 59]. They created a sample of 100 Github issues representing various -types of toxic interactions such as insults, arrogance, trolling, entitlement, and unprofessional behaviour. -Their analyses also suggest that toxicity in FOSS communities differ from those observed on other online -platforms such as Reddit or Wikipedia [59]. -Ferreira et al. [ 33] investigated incivility during code review discussions based on a qualitative analysis -of 1,545 emails from Linux Kernel Mailing Lists and found that the most common forms of incivility -among those forums are frustration, name-calling, and importance. Egelman et al. studied the negative -experiences during code review, which they referred as “pushback”, a scenario when a reviewer is blocking a -change request due to unnecessary conflict [ 29]. Qiu et al. further investigated such “pushback” phenomena -to automatically identify interpersonal conflicts [ 70]. Gunawardena et al. investigated negative code review -feedbacks based on a survey of 93 software developers, and they found that destructive criticism can be a -threat to gender diversity in the software industry as women are less motivated to continue when they -receive negative comments or destructive criticisms [40]. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:5 -2.3 State of the art toxicity detectors -To combat abusive online contents, Google’s Jigsaw AI team developed the Perspective API (PPA), -which is publicly available [ 7]. PPA is one of the general purpose state-of-the-art toxicity detectors. For a -given text, PPA generates the probability (0 to 1) of that text being toxic. As researchers are working to -identify adversarial examples to deceive the PPA [ 44], the Jigsaw team periodically updates it to eliminate -identified limitations. The Jigsaw team also published a guideline [ 8] to manually identify toxic contents -and used that guideline to curate a crowd-sourced labeled dataset of toxic online contents [ 2]. This -dataset has been used to train several deep neural network based toxicity detectors [ 22,30,37,39,82,86]. -Recently, Bhat et al.proposed ToxiScope , a supervised learning-based classifier to identify toxic on -workplace communications [ 14]. However, ToxiScope’s best model achieved a low F1-Score (i.e., =0.77) -during their evaluation -One of the major challenges in developing toxicity detectors is character-level obfuscations, where one -or more characters of a toxic word are intentionally misplaced (e.g. fcuk), or repeated (e.g., shiiit), or -replaced (e.g., s*ck) to avoid detection. To address this challenge, researchers have used character-level -encoders instead of word-level encoders to train neural networks [ 54,60,63]. Although, character-level -encoding based models can handle such character level obfuscations, they come with significant increments -of computation times [ 54]. Several studies have also found racial and gender bias among contemporary -toxicity detectors, as some trigger words (i.e., ‘gay’, ‘black’) are more likely to be associated with false -positives (i.e, a nontoxic text marked as toxic) [75, 85, 89]. -However, off-the-shelf toxicity detectors suffer significant performance degradation on SE datasets [ 77]. -Such degradation is not surprising, since prior studies found off-the-shelf natural language processing -(NLP) tools also performing poorly on SE datasets [ 6,50,56,64]. Raman etal. created the STRUDEL -tool, an SE domain specific toxicity detector [ 73], by leveraging the PPA tool and a customized version -of Stanford’s Politeness Detector [ 25]. Sarker etal. investigated the performance of the STRUDEL tool -and four other off-the-shelf toxicity detectors on two SE datasets [ 77]. In their benchmark, none of the -tools achieved reliable performance to justify practical applications on SE datasets. However, they also -achieved encouraging performance boosts, when they retrained two of the tools (i.e., DPCNN [ 94] and -BERT with FastAI [54]) using their SE datasets. -3 RESEARCH CONTEXT -To better understand our tool design, this section provides a brief overview of the machine learning (ML) -algorithms integrated in ToxiCR and five word vectorization techniques for NLP tasks. -3.1 Supervised machine learning algorithms -For ToxiCR we selected ten supervised ML algorithms from the ones that have been commonly used for -text classification tasks. Our selection includes three classical, two ensemble methods based, four deep -neural network (DNN) based, and a Bidirectional Encoder Representations from Transformer (BERT) -based algorithms. Following subsections provide a brief overview of the selected algorithms. -3.1.1 Classical ML algorithms: We have selected the following three classical algorithms, which have been -previously used for classification of SE texts [6, 20, 55, 84, 84]. -(1)Decision Tree (DT) : In this algorithm, the dataset is continuously split according to a certain -parameter. DT has two entities, namely decision nodes and leaves. The leaves are the decisions or the -final outcomes. And the decision nodes are where the data is split into two or more sub-nodes [ 71]. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:6•Sarker, et al. -(2)Logistic Regression (LR): LR creates a mathematical model to predict the probability for one of -the two possible outcomes and is commonly used for binary classification tasks [13]. -(3)Support-Vector Machine (SVM): After mapping the input vectors into a high dimensional non-linear -feature space, SVM tries to identify the best hyperplane to partition the data into n-classes, where -n is the number of possible outcomes [24]. -3.1.2 Ensemble methods: Ensemble methods create multiple models and then combine them to produce -improved results. We have selected the following two ensemble methods based algorithms, based on prior -SE studies [6, 55, 84]. -(1)Random Forest (RF): RF is an ensemble based method that combines the results produced by -multiple decision trees [ 42]. RF creates independent decision trees and combines them in parallel -using on the ‘bagging’ approach [18]. -(2)Gradient-Boosted Decision Trees (GBT): Similar to RF, GBT is also an ensemble based method -using decision trees [ 36]. However, GBT creates decision trees sequentially, so that each new tree can -correct the errors of the previous one and combines the results using the ‘boosting’ approach [80]. -3.1.3 Deep neural networks: In recent years, DNN based models have shown significant performance -gains over both classical and ensemble based models in text classification tasks [ 52,94]. In this research, -we have selected four state-of-the-art DNN based algorithms. -(1)Long Short Term Memory (LSTM): A Recurrent Neural Network (RNN) processes inputs sequen- -tially, remembers the past, and makes decisions based on what it has learnt from the past [ 74]. -However, traditional RNNs may perform poorly on long-sequence of inputs, such as those seen -in text classification tasks due to ‘the vanishing gradient problem’. This problem occurs, when a -RNN’s weights are not updated effectively due to exponentially decreasing gradients. To overcome -this limitation, Hochreiter and Schmidhuber proposed LSTM, a new type of RNN architecture, -that overcomes the challenges posed by long term dependencies using a gradient-based learning -algorithm [ 43]. LSTM consists four units: i) input gate, which decides what information to add from -current step, ii) forget gate, which decides what is to keep from prior steps, iii) output gate, which -determines the next hidden state, and iv) memory cell, stores information from previous steps. -(2)Bidirectional LSTM (BiLSTM): A BiLSTM is composed of a forward LSTM and a backward -LSTM to model the input sequences more accurately than an unidirectional LSTM [ 23,38]. In -this architecture, the forward LSTM takes input sequences in the forward direction to model -information from the past, while the backward LSTM takes input sequences in the reverse direction -to model information from the future [ 45]. BiLSTM has been shown to be performing better than -the unidirectional LSTM in several text classification tasks, as it can identify language contexts -better than LSTM [38]. -(3)Gated Recurrent Unit (GRU): Similar to LSTM, GRU belongs to the RNN family of algorithms. -However, GRU aims to handle ‘the vanishing gradient problem’ using a different approach than -LSTM. GRU has a much simpler architecture with only two units, update gate and reset gate. The -reset gate decides what information should be forgot for next pass and update gate determines -which information should pass to next step. Unlike LSTM, GRU does not require any memory cell, -and therefore needs shorter training time than LSTM [31]. -(4)Deep Pyramid CNN (DPCNN): Convolutional neural networks (CNN) are a specialized type of -neural networks that utilizes a mathematical operation called convolution in at least one of their -layers. CNNs are most commonly used for image classification tasks. Johnson and Zhang proposed a -special type of CNN architecture, named deep pyramid convolutional neural network (DPCNN) for -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:7 -text classification tasks [ 49]. Although DPCNN achieves faster training time by utilizing word-level -CNNs to represent input texts, it does not sacrifice accuracy over character-level CNNs due to its -carefully designed deep but low-complexity network architecture. -3.1.4 Transformer model: In recent years, Transformer based models have been used for sequence -to sequence modeling such as neural machine translations. For a sequence to sequence modeling, a -Transformer architecture includes two primary parts: i) the encoder , which takes the input and generates -the higher dimensional vector representation, ii) the decoder , which generates the the output sequence -from the abstract vector from the encoder. For classification tasks, the output of encoders are used -for training. Transformers solve the ‘vanishing gradient problem’ on long text inputs using the ‘self -attention mechanism’, a technique to identify the important features from different positions of an input -sequence [87]. -In this study, we select Bidirectional Encoder Representations from Transformers, which is commonly -known as BERT [ 27]. BERT based models have achieved remarkable performances in various NLP -tasks, such as question answering, sentiment classification, and text summarization [ 4,27]. BERT’s -transformer layers use multi-headed attention instead of recurrent units (e.g, LSTM, GRU) to model the -contextualized representation of each word in an input. -3.2 Word vectorization -To train an NLP model, input texts need to be converted into a vector of features that machine learning -models can work on. Bag-of-Words (BOW) is one of the most basic representation techniques, that -turns an arbitrary text into fixed-length vector by counting how many times each word appears. As -BOW representations do not account for grammar and word order, ML models trained using BOW -representations fail to identify relationships between words. On the other hand, word embedding techniques -convert words to n-dimensional vector forms in such a way that words having similar meanings have -vectors close to each other in the n-dimensional space. Word embedding techniques can be further divided -into two categories: i) context-free embedding, which creates the same representation of a word regardless -of the context where it occurs; ii) contextualized word embeddings aim at capturing word semantics -in different contexts to address the issue of polysemous (i.e., words with multiple meanings) and the -context-dependent nature of words. For this research, we have experimented with five word vectorization -techniques: one BOW based, three context-free, and one contextualized. Following subsections provide a -brief overview of those techniques. -3.2.1 Tf-Idf: TF-IDF is a BOW based vectorization technique that evaluates how relevant a word is to a -document in a collection of documents. TF-IDF score for a word is computed by multiplying two metrics: -how many times a word appears in a document ( 𝑇𝑓), and the inverse document frequency of the word -across a set of documents ( 𝐼𝑑 𝑓). Following equations show the computation steps for Tf-Idf scores. -𝑇𝑓(︀ -𝑤, 𝑑)︀ -=𝑓(︀ -𝑤, 𝑑)︀ -𝑡𝜖𝑑𝑓(︀ -𝑡, 𝑑)︀ -(1) -Where, 𝑓(︀ -𝑡, 𝑑)︀ -is the frequency of the word ( 𝑤) in the document ( 𝑑), and -𝑡𝜖𝑑𝑓(︀ -𝑡, 𝑑)︀ -represents the total -number of words in 𝑑. Inverse document frequency (Idf) measures the importance of a term across all -documents. -𝐼𝑑 𝑓(︀ -𝑤)︀ -=𝑙𝑜𝑔𝑒(︀ -𝑁𝑤𝑁)︀ -(2) -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:8•Sarker, et al. -Fig. 1. A simplified overview of ToxiCR showing key pipeline -Here, 𝑁is the total number of documents and 𝑤𝑁represents the number of documents having w. Finally, -we computed TfIdf score of a word as: -𝑇𝑓𝐼𝑑 𝑓(︀ -𝑤, 𝑑)︀ -=𝑇𝑓(︀ -𝑤, 𝑑)︀ -*𝐼𝑑 𝑓(︀ -𝑤)︀ -(3) -3.2.2 Word2vec: In 2013, Mikolaev etal. [58] proposed Word2vec, a context free word embedding -technique. It is based on two neural network models named Continuous Bag-of-Words (CBOW) and -Skip-gram. CBOW predicts a target word based on its context, while skip-gram uses the current word -to predict its surrounding context. During the training, word2vec takes a large corpus of text as input -and generates a vector space, where each word in the corpus is assigned a unique vector and words with -similar meaning are located close to one another. -3.2.3 GloVe: Proposed by Pennington et al. [ 68], Global Vectors for Word Representation (GloVe) is an -unsupervised algorithm to create context free word embedding. Unlike word2vec, GloVe generates vector -space from global co-occurrence of words. -3.2.4 fastText: Developed by the Facebook AI team, fastText is a simple and efficient method to generate -context-free word embeddings [ 15]. While Word2vec and GloVe cannot provide embedding for out of -vocabulary words, fastText overcomes this limitation by taking into account morphological characteristics -of individual words. A word’s vector in fastText based embedding is built from vectors of substrings of -characters contained in it. Therefore, fasttext performs better than Word2vec or GloVe in NLP tasks, if a -corpus contains unknown or rare words [15]. -3.2.5 BERT:. Unlike context-free embeddings (e.g., word2vec, GloVe, and fastText), where each word -has a fixed representation regardless of the context within which the word appears, a contextualized -embedding produces word representations that are dynamically informed by the words around them. In -this study, we use BERT [27]. Similar to fastText, BERT can also handle out of vocabulary words. -4 TOOL DESIGN -Figure 1 shows the architecture of ToxiCR. It takes a text ( i.e., code review comment) as input and -applies a series of mandatory preprocessing steps. Then, it applies a series of optional preprocessing based -on selected configurations. Preprocessed texts are then fed into one of the selected vectorizers to extract -features. Finally, output vectors are used to train and validate our supervised learning-based models. The -following subsections detail the research steps to design ToxiCR. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:9 -4.1 Conceptualization of Toxicity -As we have mentioned in Section 2.1, what constitutes a ‘toxic communication’ depends on various -contextual factors. In this study, we specifically focus on popular FOSS projects such as Android, -Chromium OS, LibreOffice, and OpenStack, where participants represent diverse culture, education, -ethnicity, age, religion, gender, and political views. As participants are expected and even recommended -to maintain a high level of professionalism during their interactions with other members of those -communities [ 35,65,69], we adopt the following expansive definition of toxic contents for this context.1: -“An SE conversation will be considered as toxic, if it includes any of the following: i) offensive -name calling, ii) insults, iii) threats, iv) personal attacks, v) flirtations, vi) reference to sexual -activities, and vii) swearing or cursing.” -Our conceptualization of toxicity closely aligns with another recent work by Bhat etel. that focuses -on professional workplace communication [ 14]. According to their definition, a toxic behaviour includes -any of the following: sarcasm, stereotyping, rude statements, mocking conversations, profanity, bullying, -harassment, discrimination and violence. -4.2 Training Dataset Creation -As of May 2021, there are three publicly available labeled datasets of toxic communications from the -SE domain. Raman et al.’s dataset created for the STRUDEL tool [ 73] includes only 611 texts. In our -recent benchmark study (referred as ‘the benchmark study’ hereinafter), we created two datasets, i) a -dataset of 6,533 code review comments selected from three popular FOSS projects (referred as ‘code -review dataset 1’ hereinafter), i.e., Android, Chromium OS, and LibreOffice; ii) a dataset of 4,140 Gitter -messages selected from the Gitter Ethereum channel (referred as ‘gitter dataset’ hereinafter) [ 77]. We -followed the exact same process used in the benchmark study to select and label additional 13,038 code -review comments selected from the OpenStack projects. In the following, we briefly describe our four-step -process, which is detailed in our prior publication [77]. -4.2.1 Data Mining. In the benchmark, we wrote a Python script to mine the Gerrit [61] managed code -review repositories of three popular FOSS projects, i.e., Android, Chromium OS, and LibreOffice. Our -script leverages Gerrit’s REST API to mine and store all publicly available code reviews in a MySQL -dataset. We use the same script to mine ≈2.1million code review comments belonging to 670,996 code -reviews from the OpenStack projects’ code review repository hosted at https://review.opendev.org/. We -followed an approach similar to Paul etal. [66] to identify potential bot accounts based on keywords -(e.g., ‘bot’, ‘auto’, ‘build’, ‘auto’, ‘travis’, ‘CI’, ‘jenkins’, and ‘clang’). If our manual manual validations -of comments authored by a potential bot account confirmed it to be a bot, we excluded all comments -posted by that account. -4.2.2 Stratified sampling of code review comments. Since toxic communications are rare [ 77] during code -reviews, a randomly selected dataset of code review comments will be highly imbalanced with less than -1% toxic instances. To overcome this challenge, we adopted a stratified sampling strategy as suggested -by Särndal etal. [79]. We used Google’s Perspective API (PPA) [ 7] to compute the toxicity score for -each review comment. If the PPA score is more than 0.5, then the review comment is more likely to be -toxic. Among the 2.1 million code review comments, we found 4,118 comments with PPA scores greater -than 0.5. In addition to those 4,118 review comments, we selected 9,000 code review comments with -PPA scores less than 0.5. We selected code review comments with PPA scores less than 0.5 in a well -1We introduced this definition in our prior study [ 77]. We are repeating this definition to assist better comprehension of -this paper’s context -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:10•Sarker, et al. -Table 2. An overview of the three SE domain specific toxicity datasets used in this study -Dataset # total texts # toxic # non-toxic -Code review 1 6,533 1,310 5,223 -Code review 2 13,118 2,447 10,591 -Gitter dataset 4,140 1,468 2,672 -Code review (combined) 19,651 3,757 15,819 -distributed manner. We split the texts into five categories (i.e, score: 0-0.1, 0.11-0.2, and so on) and took -the same amount (1,800 texts) from each category. For example, we took 1,800 samples that have a score -between 0.3 to 0.4. -4.2.3 Manual Labeling. During the benchmark study [ 77], we developed a manual labeling rubric fitting -our definition as well as study context. Our initial rubric was based on the guidelines published by -the Conversation AI Team (CAT) [ 8]. With these guidelines as our starting point, two of the authors -independently went through 1,000 texts to adopt the rules to better fit our context. Then, we had a -discussion session to merge and create a unified set of rules. Table 3 represents our final rubric that has -been used for manual labeling during both the benchmark study and this study. -Although we have used the guideline from the CAT as our starting point, our final rubric differs from -the CAT rubric in two key aspects to better fit our target SE context. First, our rubric is targeted -towards professional communities with contrast to the CAT rubric, which is targeted towards general -online communications. Therefore, profanities and swearing to express a positive attitude may not be -considered as toxic by the CAT rubric. For example, “That’s fucking amazing! thanks for sharing.” is -given as an example of ‘Not Toxic, or Hard to say’ by the CAT rubric. On the contrary, any sentence -with profanities or swearing is considered ‘toxic’ according to our rubric, since such a sentence does not -constitute a healthy interaction. Our characterization of profanities also aligns with the recent SE studies -on toxicity [ 59] and incivility [ 33]. Second, the CAT rubric is for labeling on a four point-scale (i.e., ‘Very -Toxic’, ‘Toxic’, ‘Slightly Toxic or Hard to Say’, and ‘Non toxic’) [ 1]. On the contrary, our labeling rubric -is much simpler on a binary scale (‘Toxic’ and ‘Non-toxic’), since development of four point rubric as -well as classifier is significantly more challenging. We consider development of a four point rubric as a -potential future direction. -Using this rubric, two of the authors independently labeled the 13,118 texts as either ‘toxic’ or ‘non- -toxic’. After the independent manual labeling, we compared the labels from the two raters to identify -conflicts. The two raters had agreements on 12,608 (96.1%) texts during this process and achieved a -Cohen’s Kappa ( 𝜅) score of 0.92 (i.e., an almost perfect agreement)2. We had meetings to discuss the -conflicting labels and assign agreed upon labels for those cases. At the end of conflict resolution, we found -2,447 (18.76%) texts labeled as ‘toxic’ among the 13,118 texts. We refer to this dataset as ‘code review -dataset 2’ hereinafter. Table 2 provides an overview of the three dataset used in this study. -4.2.4 Dataset aggregation. Since the reliability of a supervised learning-based model increases with the -size of its training dataset, we decided to merge the two code review dataset into a single dataset (referred -as ‘combined code review dataset’ hereinafter). We believe such merging is not problematic, due to the -following reasons. -(1) Both of the datasets are labeled using the same rubrics and following the same protocol. -2Kappa ( 𝜅) values are commonly interpreted as follows: values ≤0 as indicating ‘no agreement’ and 0.01 – 0.20 as ‘none to -slight’, 0.21 – 0.40 as ‘fair’, 0.41 – 0.60 as ‘moderate’, 0.61–0.80 as ‘substantial’, and 0.81–1.00 as ‘almost perfect agreement’. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:11 -Table 3. Rubric to label the SE text as toxic or non-toxic, adjusted from [77] -# Rule Rationale Example* -Rule 1: If a text includes profane -or curse words it would be -marked as ‘toxic’.Profanities are the most common -sources of online toxicities.fuck! Consider it -done!. -Rule 2: If a text includes an acronym, -that generally refers to exple- -tive or swearing, it would be -marked as ‘toxic’.Sometimes people use acronyms of -profanities, which are equally toxic -as their expanded form.WTF are you doing! -Rule 3: Insulting remarks regarding -another person or entities -would be marked as ‘toxic’.Insulting another developer may -create a toxic environment and -should not be encouraged....shut up, smarty- -pants. -Rule 4: Attacking a person’s identity -(e.g., race, religion, national- -ity, gender or sexual orien- -tation) would be marked as -‘toxic’.Identity attacks are considered -toxic among all categories of on- -line conversations.Stupid fucking super- -stitious Christians. -Rule 5: Aggressivebehaviororthreat- -ening another person or a -community would be marked -as ‘toxic’.Aggregations or threats may stir -hostility between two developers -and force the recipients to leave -the community.yeah, but I’d really -give a lot for an oppor- -tunity to punch them -in the face. -Rule 6: Both implicit or explicit Ref- -erences to sexual activities -would be marked as ‘toxic’.Implicit or explicit references to -sexual activities may make some -developers, particularly females, -uncomfortable and make them -leave a conversation.This code makes me -so horny. It’s beauti- -ful. -Rule 7: Flirtations would be marked -as ‘toxic’.Flirtations may also make a de- -veloper uncomfortable and make -a recipient avoid the other person -during future collaborationsI really miss you my -girl. -Rule 8: If a demeaning word (e.g., -‘dumb’, ‘stupid’, ‘idiot’, ‘ig- -norant’) refers to either the -writer him/herself or his/her -work, the sentence would not -be marked as ‘toxic’, if it does -not fit any of the first seven -rules.It is common in SE community to -use those word for expressing their -own mistakes. In those cases, the -use of those toxic words to them- -selves or their does not make toxic -meaning. While such texts are un- -professional [ 59], those do not de- -grade future communication or col- -laboration.I’m a fool and didn’t -get the point of the -deincrement. It makes -sense now. -Rule 9: A sentence, that does not fit -rules 1 through 8, would be -marked as ‘non-toxic’.General non-toxic comments. I think ResourceWith- -Props should be there -instead of GenericRe- -source. -*Examples are provided verbatim from the datasets, to accurately represent the context. We did not -censor any text, except omitting the reference to a person’s name. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:12•Sarker, et al. -Table 4. Examples of text preprocessing steps implemented in ToxiCR -Step Original Post Preprocessing -URL-rem ah crap. Not sure how I missed that. -http://goo.gl/5NFKcDah crap. Not sure how I missed that. -Cntr-exp this line shouldn’t end with a period this line should not end with a period -Sym-rem Missing: Partial-Bug: #1541928 Missing Partial Bug 1541928 -Rep-elm haha...looooooooser! haha.. loser! -Adv-ptrn oh right, sh*t oh right, shit -Kwrd-rem†Thesestaticvalues should be put at the -topThese values should be put at the top -Id-split†idp = self._create_dummy_idp -(add_clean_up = False)idp = self. create dummy idp(add clean -up= False) -†– an optional SE domain specific pre-processing step -(2) We used the same set of raters for manual labeling. -(3) Both of the dataset are picked from the same type of repository (i.e., Gerrit based code reviews). -The merged code review dataset includes total 19,651 code review comments, where 3,757 comments -(19.1%) are labeled as ‘toxic’. -4.3 Data preprocessing -Code review comments are different from news, articles, books, or even spoken language. For example, -review comments often contain word contractions, URLs, and code snippets. Therefore, we implemented -eight data preprocessing steps. Five of those steps are mandatory, since those aim to remove unnecessary -or redundant features. The remaining three steps are optional and their impacts on toxic code review -detection are empirically evaluated in our experiments. Two out of the three optional pre-processing steps -are SE domain specific. Table 4 shows examples of texts before and after preprocessing. -4.3.1 Mandatory preprocessing. ToxiCR implements the following five mandatory pre-processing steps. -∙URL removal (URL-rem): A code review comment may include an URL (e.g., reference to docu- -mentation or a StackOverflow post). Although URLs are irrelevant for a toxicity classifier, they can -increase the number of features for supervised classifiers. We used a regular expression matcher to -identify and remove all URLs from our datasets. -∙Contraction expansion (Cntr-exp): Contractions, which are shortened form of one or two words, are -common among code review texts. For example, some common words are: doesn’t →does not, we’re -→we are. By creating two different lexicons of the same term, contractions increase the number -of unique lexicons and add redundant features. We replaced the commonly used 153 contractions, -each with its expanded version. -∙Symbol removal (Sym-rem): Since special symbols (e.g., &, #, and ˆ ) are irrelevant for toxicity -classification tasks, we use a regular expression matcher to identify and remove special symbols. -∙Repetition elimination (Rep-elm): A person may repeat some of the characters to misspell a toxic -word to evade detection from a dictionary based toxicity detectors. For example, in the sentence -“You’re duumbbbb!”, ‘dumb’ is misspelled through character repetitions. We have created a pattern -based matcher to identify such misspelled cases and replace each with its correctly spelled form. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:13 -∙Adversarial pattern identification (Adv-ptrn): A person may misspell profane words by replacing -some characters with a symbol (e.g., ‘f*ck’ and ‘b!tch’) or use an acronym for a slang (e.g., ‘stfu’). -To identify such cases, we have developed a profanity preprocessor, which includes pattern matchers -to identify various forms of the 85 commonly used profane words. Our preprocessor replaces each -identified case with its correctly spelled form. -4.3.2 Optional preprocessing. ToxiCR includes options to apply following three optional preprocessing -steps. -∙Identifier splitting (Id-split): In this preprocessing, we use a regular expression matcher to split -identifiers written in both camelCase and under_score forms. For example, this step will replace -‘isCrap’ with ‘is Crap’ and replace ‘is_shitty’ with ‘is shitty’. This preprocessing may help to identify -example code segments with profane words. -∙Programming Keywords Removal (Kwrd-rem): Codereviewtextsoftenincludeprogramminglanguage -specific keywords (e.g., ‘while’, ‘case’, ‘if’, ‘catch’, and ‘except’). These keywords are SE domain -specific jargon and are not useful for toxicity prediction. We have created a list of 90 programming -keywords used in the popular programming languages (e.g., C++, Java, Python, C#, PHP, -JavaScript, and Go). This step searches and removes occurrences of those programming keywords -from a text. -∙Count profane words (profane-count): Since the occurrence of profane words is suggestive of a -toxic text, we think the number of profane words in a text may be an excellent feature for a -toxicity classifier. We have created a list of 85 profane words, and this step counts the occurrences -of these words in a text. While the remaining seven pre-processing steps modify an input text -pre-vectorization, this step adds an additional dimension to the post vectored output of a text. -4.4 Word Vectorizers -ToxiCR includes option to use five different word vectorizers. However, due to the limitations of the -algorithms, each of the vectorizers can work with only one group of algorithms. In our implementation, -Tfidf works only with the classical and ensemble (CLE) methods, Word2vec, GloVe, and fastText work -with the deep neural network based algorithms, and BERT model includes its pre-trained vectorizer. For -vectorizers, we chose the following implementations. -(1)TfIdf:We select the TfidfVectorizer from the scikit-learn library. To prevent overfitting, we -discard words not belonging to at least 20 documents in the corpus. -(2)Word2vec: We select the pre-trained word2vec model available at: https://code.google.com/archive/ -p/word2vec/. This model was trained with a Google News dataset of 100 billion words and contains -300-dimensional vectors for 3 million words and phrases. -(3)GloVe:Among the publicly available, pretrained GloVe models (https://github.com/stanfordnlp/ -GloVe), we select the common crawl model. This model was trained using web crawl data of 820 -billion tokens and contains 300 dimensional vectors for 2.2 million words and phrases. -(4)fastText: From the pretrained fastText models (https://fasttext.cc/docs/en/english-vectors.html), -we select the common crawl model. This model was trained using the same dataset as our selected -our GloVe model and contains 300 dimensional vectors for 2 million words. -(5)BERT:We select a variant of BERT model published as ‘BERT_en_uncased’. This model was -pre-trained on a dataset of 2.5 billion words from the Wikipedia and 800 million words from the -Bookcorpus [95]. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:14•Sarker, et al. -4.5 Architecture of the ML Models -This section discusses the architecture of the ML models implemented in ToxiCR. -4.5.1 Classical and ensemble methods. We have used the scikit-learn [ 67] implementations of the CLE -classifiers. -∙Decision Tree (DT): We have used the DecisionTreeClassifier class with default parameters. -∙Logistic Regression (LR): We have used the LogisticRegression class with default parameters. -∙Support-Vector Machine (SVM): Among the various SVM implementations offered by scikit-learn, -we have selected the LinearSVC class with default parameters. -∙Random Forest (RF): We have used the RandomForestClassifier class from scikit-learn ensembles. -To prevent overfitting, we set the minimum number of samples to split at to 5. For the other -parameters, we accepted the default values. -∙Gradient-Boosted Decision Trees (GBT): We have used the GradientBoostingClassifier class -from the scikit-learn library. We set n_iter_no_change =5, which stops the training early if the -last 5 iterations did not achieve any improvement in accuracy. We accepted the default values for -the other parameters. -4.5.2 Deep Neural Networks Model. We used the version 2.5.0 of the TensorFlow library [ 3] for training -the four deep neural network models (i.e., LSTM, BiLSTM, GRU, and DPCNN). -Common parameters of the four models are: -∙We setmax_features = 5000 (i.e., maximum number of features to use) to reduce the memory -overhead as well as to prevent model overfitting. -∙Maximum length of input is set to 500, which means our models can take texts with at most 500 -words as inputs. Any input over this length would be truncated to 500 words. -∙As all the three pre-trained word embedding models use 300 dimensional vectors to represent words -and phrases, we have set embedding size to 300. -∙The embedding layer takes input embedding matrix as inputs. Each of word ( 𝑤𝑖) from a text is -mapped (embedded) to a vector ( 𝑣𝑖) using one of the three context-free vectorizers (i.e., fastText, -GloVe, and word2vec). For a text 𝑇, its embedding matrix will have a dimension of 300𝑋𝑛, where -𝑛is the total number of words in that text. -∙Since we are developing binary classifiers, we have selected binary_crossentropy loss function for -model training. -∙We have selected the Adam optimizer (Adaptive Moment Estimation) [ 51] to update the weights -of the network during the training time. The initial learning_rate is set to 0.001. -∙During the training, we set accuracy (𝐴) as the evaluation metric. -The four deep neural models of ToxiCR are primarily based on three layers as described briefly in the -following. Architecture diagrams of the models are included in our replication package [76]. -∙Input Embedding Layer: After preprocessing of code review texts, those are converted to input -matrix. Embedded layer maps input matrix to a fixed dimension input embedding matrix. We -used three pre-trained embeddings which help the model to capture the low level semantics using -position based texts. -∙Hidden State Layer: This layer takes the position wise embedding matrix and helps to capture the -high level semantics of words in code review texts. The configuration of this layer depends on the -choice of the algorithm. ToxiCR includes one CNN (i.e., DPCNN) and three RNN (i.e., LSTM, -BiLSTM, GRU) based hidden layers. In the following, we describe the key properties of these four -types of layers. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:15 -–DPCNN blocks: Following the implementation of DPCNN [ 49], we set 7 convolution blocks with -Conv1Dlayer after the input embedding layer. We also set the other parameters of DPCNN model -following [ 49]. Outputs from each of the CNN blocks is passed to a GlobalMaxPooling1D layer to -capture the most important features from the inputs. A dense layer is set with 256 units which is -activated with a linear activation function. -–LSTM blocks: From the Keras library, we use LSTMunit to capture the hidden sequence from -input embedding vector. LSTM unit generates the high dimensional semantic representation -vector. To reshape the output dimension, we use flatten and dense layer after LSTM unit. -–BiLSTM blocks: For text classification tasks, BiLSTM works better than LSTM for capturing the -semantics of long sequence of text. Our model uses 50 units of Bidirectional LSTM units from -the Keras library to generate the hidden sequence of input embedding matrix. To downsample -the high dimension hidden vector from BiLSTM units, we set a GlobalMaxPool1D layer. This -layer downsamples the hidden vector from BiLSTM layer by taking the maximum value of each -dimension and thus captures the most important features for each vector. -–GRU blocks: We use bidirectional GRUs with 80 units to generate the hidden sequence of input -embedding vector. To keep the most important features from GRU units, we set a concatenation -ofGlobalAveragePooling1D andGlobalMaxPooling1D layers.GlobalAveragePooling1D calcu- -lates the average of entire sequence of each vector and GlobalMaxPooling1D finds the maximum -value of entire sequence. -∙Classifier Layer: The output vector of hidden state layer project to the output layer with a dense -layer and a sigmoid activation function. This layer generates the probability of the input vector -from the range 0 to 1. We chose a sigmoid activation function because it provides the probability -of a vector within 0 to 1 range. -4.5.3 Transformer models. Among the several pre-trained BERT models3we have used bert_en_uncased , -which is also known as the 𝐵𝐸𝑅𝑇_𝑏𝑎𝑠𝑒model. We downloaded the models from the tensorflow_hub , -which consists trained machine learning models ready for fine tuning. -Our BERT model architecture is as following: -∙Input layer: takes the preprocessed input text from our SE dataset. To fit into BERT pretrained en- -coder,wepreprocesseachtextusingmatchingpreprocessingmodel(i.e. bert_en_uncased_preprocess -4). -∙BERT encoder: From each preprocessed text, this layer produces BERT embedding vectors with -higher level semantic representations. -∙Dropout Layer: To prevent overfitting as well as eliminate unnecessary features, outputs from the -BERT encoder layer is passed to a dropout layer with a probability of 0.1 to drop an input. -∙Classifier Layer: Outputs from the dropout layer is passed to a two-unit dense layer, which transforms -the outputs into two-dimensional vectors. From these vectors, a one-unit dense layer with linear -activation function generates the probabilities of each text being toxic. Unlike deep neural network’s -output layer, we have found that linear activation function provides better accuracy than non-linear -ones (e.g, 𝑟𝑒𝑙𝑢,𝑠𝑖𝑔𝑚𝑜𝑖𝑑) for the BERT-based models. -∙Parameters: Similar to the deep neural network models, we use binary_crossentropy as the loss -function and Binary Accuracy as the evaluation metric during training. -3https://github.com/google-research/bert -4https://tfhub.dev/tensorflow/bert_en_uncased_preprocess/3 -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:16•Sarker, et al. -Table 5. An overview of the hyper parameters for our deep neural networks and transformers -Hyper-Parameters Deep neural networks (i.e., DPCNN, -LSTM, BiLSTM, and GRU)Transformer (BERT) -Activation sigmoid linear -Loss function binary crossentropy binary crossentropy -Optimizer adam Adamw -Learning rate 0.001 3e-5 -Early stopping monitor val_loss val_loss -Epochs 40 15 -Batch size 128 256 -∙Optimizer: We set the optimizer as Adamw[57] which improved the generalization performance of -‘adam’ optimizer. Adamwminimizes the prediction loss and does regularization by decaying weight. -Following the recommendation of Devlin etal. [27], we set the initial learning rate to 3𝑒−5. -4.6 Model Training and Validation -Following subsections detail our model training and validation approaches. -4.6.1 Classical and ensembles. We evaluated all the models using 10-fold cross validations, where the -dataset was randomly split into 10 groups and each of the ten groups was used as test dataset once, while -the remaining nine groups were used to train the model. We used stratified split to ensure similar ratios -of the classes between the test and training sets. -4.6.2 DNN and Transformers. We have customized several hyper-parameters of the DNN models to train -our models. Table 5 provides an overview of those customized hyper-parameters. A DNN model can be -overfitted due to over training. To encounter that, we have configured our training parameters to find the -best fit model that is not overfitted. During training, we split our dataset into three sets according to -8:1:1 ratio. These three sets are used for training, validation, and testing respectively during our 10-fold -cross validations to evaluate our DNN and transformer models. For training, we have set maximum 40 -epochs5for the DNN models and maximum 15 epochs for the BERT model. During each epoch, a model -is trained using 80% samples, is validated using 10% samples, and the remaining 10% is used to measure -the performance of the trained model. To prevent overfitting, we have used an EarlyStopping function -from the Keras library, which monitors minimum val loss . If the performance of a model on the validation -dataset starts to degrade (e.g. loss begins to increase or accuracy begins to drop), then the training -process is stopped. -4.7 Tool interface -We have designed ToxiCR to support standalone evaluation as well as being used as a library for toxic -text identification. We have also included pre-trained models to save model training time. Listing 1 shows -a sample code to predict the toxicity of texts using our pretrained BERT model. -We have also included a command line based interface for model evaluation, retraining, and fine tuning -hyperparameters. Figure 2 shows the usage help message of ToxiCR. Users can customize execution with -eight optional parameters, which are as follows: -5the number times that a learning algorithm will work through the entire training dataset -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:17 -Fig. 2. The command line interface of ToxiCR showing various customization options -1from ToxiCR import ToxiCR -2 -3clf=ToxiCR(ALGO="BERT" , count_profanity=False , remove_keywords=True , -4 split_identifier=False , -5 embedding="bert" , load_pretrained=True) -6 -7clf . init_predictor () -8sentences=[" this is crap" , "thank you for the information" , -9 " shi ∗tty code" ] -10 -11results=clf . get_toxicity_class(sentences) -Listing 1. Example usage of ToxiCR to classify toxic texts -∙Algorithm Selection: Users can select one of the ten included algorithms by using the –algo ALGO -option. -∙Number of Repetitions: Users can specify the number of times to repeat the 10-fold cross-validations -in evaluation mode using –repeat n option. Default value is 5. -∙Embedding: ToxiCR includes five different vectorization techniques: tfidf,word2vec ,glove, -fasttext , andbert.tfidfis configured to be used only with the CLE models. word2vec ,glove, -andfastext can be used only with the DNN models. Finally, bertcan be used only with the -transformer model. Users can customize this selection using the –embed EMBED option. -∙Identifier splitting: Using the –splitoption, users can select to apply the optional preprocessing -step to split identifiers written in camelCases or under_scores. -∙Programming keywords: Using the –keyword option, users can select to apply the optional prepro- -cessing step to remove programming keywords. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:18��Sarker, et al. -∙Profanity: The –profanity optional preprocessing step allows to add the number of profane words -in a text as an additional feature. -∙Missclassification diagnosis: The –retrooption is useful for error diagnosis. If this option is selected, -ToxiCR will write all misclassified texts in a spreadsheet to enable manual analyses. -∙Execution mode: ToxiCR can be executed in three different modes. The evalmode will run 10-fold -cross validations to evaluate the performance of an algorithm with the selected options. In the eval -mode, ToxiCR writes results of each run and model training time in a spreadsheet. The retrain -mode will train a classifier with the full dataset. This option is useful for saving models in a file to be -used in the future. Finally, the tuningmode allows to explore various algorithm hyperparameters -to identify the optimum set. -5 EVALUATION -We empirically evaluated the ten algorithms included in ToxiCR to identify the best possible configuration -to identify toxic texts from our datasets. Following subsections detail our experimental configurations -and the results of our evaluations. -5.1 Experimental Configuration -To evaluate the performance of our models, we use precision, recall, f-score, accuracy for both toxic (class -1) and non-toxic (class 0) classes. We computed the following evaluation metrics. -∙Precision ( 𝑃):For a class, precision is the percentage of identified cases that truly belongs to that -class. -∙Recall ( 𝑅):For a class, recall is the ratio of correctly predicted cases and total number of cases. -∙F1-score ( 𝐹1):F1-score is the harmonic mean of precision and recall. -∙Accuracy ( 𝐴):Accuracy is the percentage of cases that a model predicted correctly. -In our evaluations, we consider F1-score for the toxic class (i.e., 𝐹11) as the most important metric to -evaluate these models, since: i) identification of toxic texts is our primary objective, and ii) our datasets -are imbalanced with more than 80% non-toxic texts. -To estimate the performance of the models more accurately, we repeated 10-fold cross validations -five times and computed the means of all metrics over those 5 *10 =50 runs. We use Python’s Random -module, which is a pseudo-random number generator, to create stratified 10-fold partitions, preserving -the ratio between the two classes across all partitions. If initialized with the same seed number, Random -would generate the exact same sequence of pseudo-random numbers. At the start of each algorithm’s -evaluation, we initialized the Randomgenerator using the same seed to ensure the exact same sequence -of training/testing partitions for all algorithms. As the model performances are normally distributed, -we use paired sample t-tests to check if observed performance differences between two algorithms are -statistically significant ( 𝑝 < 0.05). We use the ‘paired sample t-test’, since our experimental setup -guarantees cross-validation runs of two different algorithms would get the same sequences of train/test -partitions. We have included the results of the statistical tests in the replication package [76]. -We conducted all evaluations on an Ubuntu 20.04 LTS workstation with Intel i7-9700 CPU, 32GB -RAM, and an NVIDIA Titan RTX GPU with 24 GB memory. For python configuration, we created an -Anaconda environment with Python 3.8.0, and tensorflow /tensorflow-gpu 2.5.0. -5.2 Baseline Algorithms -To establish baseline performances, we computed the performances of four existing toxicity detectors -(Table 6) on our dataset. We briefly describe the four tools in the following. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:19 -Table6. Performancesofthefourcontemporarytoxicdetectorstoestablishabaselineperformance.Forourclassifications, -we consider toxic texts as the ‘class 1’ and non-toxic texts as the ‘class 0’. -ModelsNon-toxic ToxicAccuracy𝑃0𝑅0𝐹10 𝑃1𝑅1𝐹11 -Perspective API [7] (off-the-shelf) 0.920.790.850.450.700.55 0.78 -Strudel Tool (off-the-shelf) [73] 0.930.760.830.430.770.55 0.76 -Strudel (retrain) [78] 0.970.960.970.850.860.85 0.94 -DPCNN (retrain) [77] 0.940.950.940.810.760.78 0.91 -(1)Perspective API [ 7] (off-the-shelf): To prevent the online community from abusive content, Jigsaw -and Google’s Counter Abuse Technology team developed Perspective API [ 7]. Algorithms and -datasetstotrainthesemodelsarenotpubliclyavailable.PerspectiveAPIcangeneratetheprobability -score of a text being toxic, servere_toxic, insult, profanity, threat, identity_attack, and sexually -explicit. The score for each category is from 0 to 1 where the probability of a text belonging to that -category increases with the score. For our two class classification, we considered a text as toxic if its -Perspective API score for the toxicity category is higher than 0.5. -(2)STRUDEL tool [ 73] (off-the-shelf): The STRUDEL tool is an ensemble based on two existing tools: -Perspective API and Stanford politeness detector and BoW vector obtained from preprocessed -text. Its classification pipeline obtains toxicity score of a text using the Perspective API, computes -politeness score using the Stanford politeness detector tool [ 25], and computes BoW vector using -TfIdf. For SE specificity, its TfIdf vectorizer excludes words that occur more frequently in the SE -domain than in a non-SE domain. Although, STRUDEL tool also computes several other features -such as sentiment score, subjectivity score, polarity score, number of LIWC anger words, and the -number of emoticons in a text, none of these features contributed to improved performances during -its evaluation [ 73]. Hence, the best performing ensemble from STRUDEL uses only the Perspective -API score, Stanford politeness score, and TfIdf vector. The off-the-shelf version is trained on a -manually labeled dataset of 654 Github issues. -(3)STRUDEL tool [ 78] (retrain): Due to several technical challenges, we were unable to retrain -the STRUDEL tool using the source code provided in its repository [ 73]. Therefore, we wrote a -simplified re-implementation based on the description included in the paper and our understanding -of the current source code. Upon contacting, the primary author of the tool acknowledged our -implementation as correct. Our implementation is publicly available inside the WSU-SEAL directory -of the publicly available repository: https://github.com/WSU-SEAL/toxicity-detector. Our pull -request with this implementation has been also merged to the original repository. For computing -baseline performance, we conducted a stratified 10-fold cross validation using our code review -dataset. -(4)DPCNN [ 77] (retrain): We cross-validated a DPCNN model [ 49], using our code review dataset. -We include this model in our baseline, since it provided the best retrained performance during our -benchmark study [77]. -Table 6 shows the performances of the four baseline models. Unsurprisingly, the two retrained models -provide better performances than the off-the-shelf ones. Overall, the retrained Strudel tool provides -the best scores among the four tools on all seven metrics. Therefore, we consider this model as the key -baseline to improve on. The best toxicity detector among the ones participating in the 2020 SemEval -challenge achieved 0.92 𝐹1score on the Jigsaw dataset [ 91]. As the baseline models listed in Table 6 are -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:20•Sarker, et al. -Table 7. Mean performances of the ten selected algorithms based on 10-fold cross validations. For each group, shaded -background indicate significant improvements over the others from the same group -Group Models VectorizerNon-toxic ToxicAccuracy ( 𝐴)𝑃0 𝑅0 𝐹10 𝑃1 𝑅1 𝐹11 -CLEDT tfidf 0.9540.9630.9590.8410.8060.823 0.933 -GBT tfidf 0.9260.9850.9550.9160.6720.775 0.925 -LR tfidf 0.9180.9830.9490.9000.6330.743 0.916 -RF tfidf 0.9560.9820.9690.9160.8100.859 0.949 -SVM tfidf 0.9290.9790.9540.8880.6880.775 0.923 -DNN1DPCNN word2vec 0.9620.9660.9640.8700.8410.849 0.942 -DPCNN GloVe 0.9630.9660.9640.8710.8420.851 0.943 -DPCNN fasttext 0.9640.9670.9650.8700.8450.852 0.944 -DNN2LSTM word2vec 0.9290.9780.9530.8660.6980.778 0.922 -LSTM GloVe 0.9440.9710.9570.8640.7580.806 0.930 -LSTM fasttext 0.9360.9740.9540.8530.7180.778 0.925 -DNN3BiLSTM word2vec 0.9650.9740.9690.8870.8510.868 0.950 -BiLSTM GloVe 0.9650.9760.9680.8950.8280.859 0.948 -BiLSTM fasttext 0.9660.9740.9700.8880.8540.871 0.951 -DNN4GRU word2vec 0.9640.9760.9700.8940.8470.870 0.951 -GRU GloVe 0.9650.9770.9710.9010.8510.875 0.953 -GRU fasttext 0.9650.9740.9690.8880.8520.869 0.951 -Transformer BERT en uncased 0.9710.9760.9730.9010.8760.887 0.957 -evaluated on a different dataset, it may not be fair to compare these models against the ones trained on -Jigsaw dataset. However, the best baseline model’s 𝐹1score is 7 (i.e., 0.92 -0.85 ) points lower than the -ones from a non-SE domain. This result suggests that with existing technology, it may be possible to train -SE domain specific toxicity detectors with better performances than the best baseline listed in Table 6. -Finding 1: Retrained models provide considerably better performances than the off-the-shelf ones, -with the retrained STRUDEL tool providing the best performances. Still the 𝐹11score from the best -baseline model lags 7 points behind the 𝐹11score of state-of-the-art models trained and evaluated on -the Jigsaw dataset during 2020 SemEval challenge [91]. -5.3 How do the algorithms perform without optional preprocessing? -Following subsections detail the performances of the three groups of algorithm described in the Section 4.5. -5.3.1 Classical and Ensemble (CLE) algorithms. The top five rows of the Table 7 (i.e., CLE group) show -the performances of the five CLE models. Among, those five algorithms, RF achieves significantly higher -𝑃0(0.956), 𝐹10(0.969), 𝑅1(0.81), 𝐹11(0.859) and accuracy (0.949) than the four other algorithms -from this group. The RF model also significantly outperforms (One-sample t-test) the key baseline (i.e., -retrained STRUDEL) in terms of the two key metrics, accuracy ( 𝐴) and 𝐹11. Although, STRUDEL -retrain achieves better recall ( 𝑅1), our RF based model achieves better precision ( 𝑃1). -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:21 -5.3.2 Deep Neural Networks (DNN). We evaluated each of the four DNN algorithms using three different -pre-trained word embedding techniques (i.e., word2vec, GloVe, and fastText) to identify the best per- -forming embedding combinations. Rows 6 to 17 (i.e., groups: DNN1, DNN2, DNN3, and DNN4) of the -Table 7 show the performances of the four DNN algorithms using three different embeddings. For each -group, statistically significant improvements (paired-sample t-tests) over the other two configurations -are highlighted using shaded background. Our results suggest that choice of embedding does influence -performances of the DNN algorithms. However, such variations are minor. -For DPCNN, only 𝑅1score is significantly better with fastText than it is with GloVe or word2vec. The -other scores do not vary significantly among the three embeddings. Based on these results, we recommend -fastText for DPCNN in ToxiCR. For LSTM and GRU, GloVe boosts significantly better 𝐹11scores than -those based on fastText or word2vec. Since 𝐹11is one of the key measures to evaluate our models, we -recommend the GloVe for both LSTM and GRU in ToxiCR. Glove also boosts the highest accuracy -for both LSTM (although not statistically significant) and GRU. For BiLSTM, since fastText provides -significantly higher 𝑃0,𝑅1, and 𝐹11scores than those based on GloVe or word2vec. We recommend -fastText for BiLSTM in ToxiCR. These results also suggest that three out of the four selected DNN -algorithms (i.e., except LSTM) significantly outperform (one-sample t-test) the key baseline (i.e., retrained -STRUDEL) in terms of both accuracy and 𝐹11-score. -5.3.3 Transformer. The bottom row of Table 7 shows the performance of our BERT based model. This -model achieves the highest mean accuracy (0.957) and 𝐹11(0.887) among all the 18 models listed in -Table 7. This model also outperforms the baseline STRUDEL retrain on all the seven metrics. -Finding 2: From the CLE group, RF provides the best performances. From the DNN group GRU with -glove provides the best performances. Among the 18 models from the six groups, BERT achieves the -best performances. Overall, ten out of the 18 models also outperform the baseline STRUDEL retrain -model. -5.4 Do optional preprocessing steps improve performance? -For each of the ten selected algorithms, we evaluated whether the optional preprocessing steps (especially -SE domain specific ones) improve performances. Since ToxiCR includes three optional preprocessing (i.e., -identifier splitting ( id-split), keyword removal ( kwrd-remove ), and counting profane words ( profane-count ), -we ran each algorithm with 23=8different combinations. For the DNN models, we did not evaluate all -three embeddings in this step, as that would require evaluating 3*8= 24 possible combinations for each one. -Rather we used only the best performing embedding identified in the previous step (i.e., Section 5.3.2). -To select the best optional preprocessing configuration from the eight possible configurations, we use -mean accuracy and mean 𝐹11scores based on five time 10-fold cross validations. We also used pair -sampled t-tests to check whether any improvement over its base configuration’s, as listed in the Table 7 -(i.e., no optional preprocessing selected), is statistical significant ( paired sample t-test, 𝑝 <0.05). -Table 8 shows the best performing configurations for all algorithms and the mean scores for those -configurations. Checkmarks ( ✓) in the preprocessing columns for an algorithm indicate that the best -configuration for that algorithm does use that pre-processing. To save space, we report the performances of -only the best combination for each algorithm. Detailed results are available in our replication package [ 76]. -These results suggest that optional pre-processing steps do improve the performances of the models. -Notably, CLE models gained higher improvements than the other two groups. RF’s accuracy improved -from 0.949 to 0.955 and 𝐹11improved from 0.859 to 0.879 with the profane-count preprocessing. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:22•Sarker, et al. -Table 8. Best performing configurations of each model with optional preprocessing steps. Shaded background indicates significant improvements -over its base configuration (i.e., no optional preprocessing). For each column, bold font indicates the highest value for that measure. †– indicates -an optional SE domain specific pre-processing step. -Group Algo VectorizerPreprocessing Non-toxic Toxic𝐴profane- -countkwrd- -removeid- -split𝑃0 𝑅0 𝐹10 𝑃1 𝑅1 𝐹11 -CLEDT tfidf ✓ ✓ -0.9600.9680.9640.862 0.8300.845 0.942 -GBT tfidf ✓ ✓ -0.9380.9810.9590.901 0.7290.806 0.932 -LR tfidf ✓ ✓ -0.9320.9810.9560.898 0.6980.785 0.927 -RF tfidf ✓ - -0.9640.9810.9720.917 0.8450.879 0.955 -SVM tfidf ✓ ✓ -0.9390.9770.9580.886 0.7360.804 0.931 -DNNDPCNN fasttext ✓ - -0.9640.9730.9680.889 0.8460.863 0.948 -LSTM glve ✓ ✓ ✓0.9440.9740.9590.878 0.7560.810 0.932 -BiLSTM fasttext ✓ - ✓0.9660.9750.9710.892 0.8580.875 0.953 -BiGRU glove ✓ - ✓0.9660.9760.9710.897 0.8560.876 0.954 -Transormer BERT bert - ✓ -0.9700.9780.9740.907 0.8740.889 0.958 -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:23 -Table 9. Performance of ToxiCR on Gitter dataset -Mode Models VectorizerNon-toxic ToxicAccuracy𝑃 𝑅 𝐹1 𝑃 𝑅 𝐹1 -Cross-validation -(retrain)RF TfIdf 0.8510.9450.8970.8790.6990.779 0.859 -BERT BERT-en-uncased 0.9310.9090.9190.8430.8770.856 0.898 -Cross-prediction -(off-the-shelf)RF TfIdf 0.8570.9770.9140.9450.7040.807 0.881 -BERT BERT-en-uncased 0.8970.9490.9230.8970.8020.847 0.897 -During these evaluations, other CLE models also achieved between 0.02 to 0.04 performance boosts in -our key measures (i.e., 𝐴and𝐹11). Improvements from optional preprocessing also depend on algorithm -choices. While the profane-count preprocessing improved performances of all the CLE models, kwrd-remove -improved all except RF. On the other hand, id-splitimproved none of the CLE models. -All the DNN models also improved performances with the profane-count preprocessing. Contrasting -the CLE models, id-splitwas useful for three out of the four DNNs. kwrd-remove preprocessing improved -only LSTM models. Noticeably, gains from optional preprocessing for the DNN models were less than -0.01 over the base configurations’ and statistically insignificant (paired-sample t-test, 𝑝 >0.05) for most -of the cases. Finally, although we noticed slight performance improvement (i.e., in 𝐴and𝐹11) of the -BERT model with kwrd-remove , the differences are not statistically significant. Overall, at the end of our -extensive evaluation, we found the best performing combination was a BERT model with kwrd-remove -optional preprocessing. The best combination provides 0.889 𝐹11score and 0.958 accuracy. The best -performing model also significantly outperforms (one sample t-test, 𝑝 < 0.05) the baseline model (i.e, -STRUDEL retrain in Table 6) in all the seven performance measures. -Finding3: Eight out of the ten models (i.e., except SVM and DPCNN) achieved significant performance -gains through SE domain preprocessing such as programming keyword removal and identifier splitting. -Although keyword removal may be useful for all the four classes of algorithms, identifier splitting is -useful only for three DNN models. Our best model is based on BERT, which significantly outperforms -the STRUDEL retrain model on all seven measures. -5.5 How do the models perform on another dataset? -To evaluate the generality of our models, we have used the Gitter dataset of 4,140 messages from our -benchmark study [ 77]. In this step, we conducted two types of evaluations. First, we ran 10-fold cross -validations of the top CLE model (i.e., RF) and the BERT model using the Gitter dataset. Second, we -evaluated cross dataset prediction performance (i.e., off-the-shelf) by using the code review dataset for -training and the Gitter dataset for testing. -The top two rows of the Table 9 shows the results of 10-fold cross-validations for the the two models. -We found that the BERT model provides the best accuracy (0.898) and the best 𝐹11(0.856). On the -Gitter dataset, all the seven performance measures achieved by the BERT model are lower than those -on the code review dataset. It may be due to the smaller size of the Gitter dataset (4,140 texts) than -the code review dataset (19,651 texts). The bottom two rows of the Table 9 shows the results of our -cross-predictions (i.e., off-the-shelf). Our BERT model achieved similar performances in terms of 𝐴and -𝐹11in both modes. However, the RF model performed better on the Gitter dataset in cross-prediction -mode (i.e., off-the-shelf) than in cross-validation mode. This result further supports our hypothesis that -the performance drops of our models on the Gitter dataset may be due to smaller sized training data. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:24•Sarker, et al. -Table 10. Confusion Matrix for our best performing model (i.e., BERT) for the combined code review dataset -Predicted -ToxicNon-toxic -ActualToxic 3259 483 -Non-toxic 373 15,446 -Finding 4: Although, our best performing model provides higher precision off-the-shelf on the Gitter -dataset than that from the retrained model, the later achieves better recall. Regardless, our BERT -model achieves similar accuracy and 𝐹11during both off-the-shelf usage and retraining. -5.6 What are the distributions of misclassifications from the best performing model? -The best-performing model (i.e., BERT) misclassified only 856 texts out of the 19,651 texts from our -dataset. There are 373 false positives and 483 false negatives. Table 10 shows the confusion matrix of the -BERT model. To understand the reasons behind misclassifications, we adopted an open coding approach -where two of the authors independently inspected each misclassified text to identify general scenarios. -Next, they had a discussion session, where they developed an agreed upon higher level categorization -scheme of five groups. With this scheme, those two authors independently labeled each misclassified text -into one of those five groups. Finally, they compared their labels and resolved conflicts through mutual -discussions. -False negative False positive -0.0%10.0%20.0%30.0%40.0%MissclassificationsError categories -Confunding contexts -Bad acronym -Self deprecation -SE domain specific words -General error -Fig. 3. Distribution of the misclassifications from the BERT model -Figure 3 shows distributions of the five categories of misclassifcations from ToxiCR grouped by False -Positives (FP) and False Negatives (FN). Following subsections detail those error categories. -5.6.1 General erros (GE). General errors are due to failures of the classifier to identify the pragmatic -meaning of various texts. These errors represent 45% of the false positives and 46% of the false negatives. -Many GE false positives are due to words or phrases that more frequently occur in toxic contexts and -vice versa. For example, “If we do, should we just get rid of the HBoundType?” and“Done. I think they -came from a messed up rebase.” are two false positive cases, due to the phrases ‘get rid of’ and ‘messed -up’ that have occurred more frequently in toxic contexts. -GE errors also occurred due to infrequent words. For example, “"Oh, look. The stupidity that makes -me rant so has already taken root. I suspect it’s not too late to fix this, and fixing this rates as a mitzvah -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:25 -in my book." – is incorrectly predicted as non-toxic as a very few texts in our dataset include the word -‘stupidity’. Another such instance was “this is another instance of uneducated programmers calling any -kind of polymorphism overloading, please translate it to override.” , due to to word ‘uneducated’. As we did -not have many instances of identify attacks in our dataset, most of those were also incorrectly classified. -For example, “most australian dummy var name ever!" was predicted as non-toxic by our classifier. -5.6.2 SE domain specific words (SE):. Words that have different meanings in the SE domain than its’ -meaning in the general domain (die, dead, kill, junk, and bug) [ 77] were responsible for 40% false positives -and 43% false negatives. For example, the text “you probably wanted ‘die‘ here. eerror is not fatal.” , -is incorrectly predicted as toxic due to the presence of the words ‘die’ and ‘fatal’. On the other hand, -although the word ‘junk’ is used to harshly criticize a code in the sentence “I don’t actually need all this -junk...”, this sentence was predicted as non-toxic as most of the code review comments from our dataset -do not use ‘junk’ in such a way. -5.6.3 Self deprecation (SD):. Usage of self-deprecating texts to express humility is common during code -reviews [59,77]. We found that 13% of 373 false positives and 11% of 493 false negatives were due to the -presence of self deprecating phrases. For example, “ Missing entry in kerneldoc above... (stupid me) ” is -labeled as ‘non-toxic’ in our dataset but is predicted as ‘toxic’ by our model. Although, our model did -classify many of the SD texts expressing humbleness correctly, those texts also led to some false negatives. -For example, although “ Huh? Am I stupid? How’s that equivalent? ” was misclassified as non-toxic, it fits -‘toxic’ according to our rubric due to its aggressive tone. -5.6.4 Bad acronym (BA). In few cases, developers have used acronyms with with alternate toxic expansion. -For example, the webkit framework used the acronym ‘WTF’ -‘Web Template Framework’6, for a -namespace. Around 2% of our false positive cases were comments referring to the ‘WTF’ namespace from -Webkit. -5.6.5 Confounding contexts (CC). Some of the texts in our dataset represent confounding contexts and -were challenging even for the human raters to make a decision. Such cases represent 0.26% false positives -and 1.04% false negatives. For example, “This is a bit ugly, but this is what was asked so I added a null -ptr check for |inspector_agent_|. Let me know what you think.” is a false positive case from our dataset. -We had labeled it as non-toxic, since the word ‘ugly’ is applied to critique code written by the author -of this text. On the other hand, “I just know the network stack is full of _bh poop. Do you ever get -called from irq context? Sorry, I didn’t mean to make you thrash.” is labeled as toxic due to thrashing -another person’s code with the word ‘poop’. However, the reviewer also said sorry in the next sentence. -During labeling, we considered it as toxic, since the reviewer could have critiqued the code in a nicer way. -Probably due to the presence of mixed contexts, our classifier incorrectly predicted it as ‘non-toxic’. -Finding 5: Almost 85% of the misclassifications are due to either our model’s failure to accurately -comprehend the pragmatic meaning of a text (i.e., GE) or words having SE domain specific synonyms. -6 IMPLICATIONS -Based on our design and evaluation of ToxiCR, we have identified following lessons. -Lesson 1: Development of a reliable toxicity detector for the SE domain is feasible. Despite of creating -an ensemble of multiple NLP models (i.e., Perspective API, Sentiment score, Politeness score, Subjectivity, -and Polarity) and various categories of features (i.e., BoW, number of anger words, and emoticons), the -6https://stackoverflow.com/questions/834179/wtf-does-wtf-represent-in-the-webkit-code-base -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:26•Sarker, et al. -STRUDEL tool achieved only 0.57 F-score during their evaluation. Moreover, a recent study by Miller -etal. found false positive rates as high as 98% [ 59]. On the other hand, the best model from the ‘2020 -Semeval Multilingual Offensive Language Identification in Social Media task’ achieved a 𝐹11score of -92.04% [92]. Therefore, the question remains, whether we can build a SE domain specific toxicity detector -that achieves similar performances (i.e., 𝐹1=0.92) as the ones from non-SE domains. -In designing ToxiCR, we adopted a different approach, i.e., focusing on text preprocessing and leveraging -state-of-the-art NLP algorithms rather than creating ensembles to improve performances. Our extensive -evaluation with a large scale SE dataset has identified a model that has 95.8% accuracy and boosts 88.9% -𝐹11score in identifying toxic texts. This model’s performances are within 3% of the best one from a -non-SE domain. This result also suggests that with a carefully labeled large scale dataset, we can train -an SE domain specific toxicity detector that achieves performances that are close to those of toxicity -detectors from non-SE domains. -Lesson 2: Performance from Random Forest’s optimum configuration may be adequate if GPU is not -available. -While a deep learning-based model (i.e., BERT) achieved the best performances during our evaluations, -that model is computationally expensive. Even with a high end GPU such as Titan RTX, our BERT -model required on average 1,614 seconds for training. We found that RandomForest based models trained -on a Core-i7 CPU took only 64 seconds on average. -During a classification task, RF generates the decision using majority voting from all sub trees. RF is -suitable for high dimensional noisy data like the ones found in text classification tasks [ 47]. With carefully -selected preprocessing steps to better understand contexts (e.g., profanity count) RF may perform well -for binary toxicity classification tasks. In our model, after adding profane count features, RF achieved an -average accuracy of 95.5% and 𝐹11- score of 87.9%, which are within 1% of those achieved by BERT. -Therefore, if computation cost is an issue, a RandomForest based model may be adequate for many -practical applications. However, as our RF model uses a context-free vectorizer, it may perform poorly -on texts, where prediction depends on surrounding contexts. Therefore, for a practical application, a user -must take that limitation into account. -Lesson 3: Preprocessing steps do improve performances. -We have implemented five mandatory and three optional preprocessing steps in ToxiCR. The mandatory -preprocessing steps do improve performances of our models. For example, a DPCNN model without these -preprocessing achieved 91% accuracy and 78% 𝐹11(Table 6). On the other hand, a model based on the -same algorithm achieved 94.4% accuracy and 84.5% 𝐹11with these preprocessing steps. Therefore, we -recommend using both domain specific and general preprocessing steps. -Two of our pre-processing steps are SE domain specific (i.e., Identifier Splitting, Programming Keywords -removal). Our empirical evaluation of those steps (Section 5.4) suggest that eight out of the ten models -(i.e., except SVM and DPCNN) achieved significant performance improvements through these steps. -Although, none of the models showed significant degradation through these steps, significant gains were -dependent on algorithm selection, with CLE algorithms gaining only from keyword removal and identifier -splitting improving only the DNN ones. -Lesson 4: Performance boosts from the optional preprocessing steps are algorithm dependent. -The three optional preprocessing steps also improved performances of the classifiers. However, per- -formance gains through the these steps were algorithm dependent. The profane-count preprocessing -had the highest influence as nine out of the ten models gained performance with this step. On the other -id-split was the least useful one with only three DNN models gaining minor gains with this step. CLE -algorithms gained the most with ≈1% boost in terms of accuracies and 1 -3% in terms of 𝐹11scores. On -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:27 -the other hand, DNN algorithms had relatively minor gains (i.e., less than 1%) in both accuracies and -𝐹11scores. Since DNN models utilize embedding vectors to identify semantic representation of texts, -those are less dependent on these optional preprocessing steps. -Lesson 5: Accurate identification of self-deprecating texts remains a challenge. -Almost 11% (out of 856 misclassified texts) of the errors from our best performing model were -due to self-deprecating texts. Challenges in identifying such texts have been also acknowledged by -prior toxicity detectors [ 41,88,93]. Due to the abundance of self-deprecating texts among code review -interactions [ 59,77], we believe that this can be an area to improve on for future SE domain specific -toxicity detectors. -Lesson 6: Achieving even higher performance is feasible. Since 85% of errors are due to failures of -our models to accurately comprehend the contexts of words, we believe achieving further improved -performance is feasible. Since supervised models learn better from larger training datasets, a larger -dataset (e.g., Jigsaw dataset includes 160K samples), may enable even higher performances. Second, NLP -is a rapidly progressing area with state-of-the-art techniques changing almost every year. Although, we -have not evaluated the most recent generation of models, such as GPT-3 [ 19] and XLNet [ 90] in this -study, those may help achieving better performances, as they are better at identifying contexts. -7 THREATS TO VALIDITY -In the following, we discuss the four common types of threats to validity for this study. -7.1 Internal validity -The first threat to validity for this study is our selection of data sources which come from four FOSS -projects. While these projects represent four different domains, many domains are not represented in -our dataset. Moreover, our projects represent some of the top FOSS projects with organized governance. -Therefore, several categories of highly offensive texts may be underrepresented in our datasets. -The notion of toxicity also depends on multitude of different factors such as culture, ethnicity, country -of origin, language, and relationship between the participants. We did not account for any such factors -during our dataset labeling. -7.2 Construct validity -Our stratified sampling strategy was based on toxicity scores obtained from the perspective API. Although, -we manually verified all the texts classified as ‘toxic’ by the PPA, we randomly selected only (5,5107 -+ 9,000 =14,510) texts that had PPA scores of less than 0.5. Among those 14,510 texts, we identified -only 638 toxic ones (4.4%). If both the PPA and our random selections missed some categories of toxic -comments, instances of such texts may be missing in our datasets. Since our dataset is relatively large -(i.e., 19,651 ), we believe that this threat is negligible. -According to our definition, Toxicity is a large umbrella that includes various anti-social behaviour such -as offensive names, profanity, insults, threats, personal attacks, flirtations, and sexual references. Though -our rubric is based on the Conversational AI team, we have modified it to fit a diverse and multicultural -professional workplace such as an OSS project. As the notion of toxicity is a context dependent complex -phenomena, our definition may not fit many organizations, especially the homogeneous ones. -Researcher bias during our manual labeling process could also cause mislabeled instances. To eliminate -such biases, we focused on developing a rubric first. With the agreed upon rubric, two of the authors -7Code review 1 dataset -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:28•Sarker, et al. -independently labeled each text and achieved ‘almost perfect’ ( 𝜅=0.92) inter-rater agreement. Therefore, -we do not anticipate any significant threat arising from our manual labeling. -We did not change most of the hyperparameters for the CLE algorithms and accepted the default -parameters. Therefore, some of the CLE models may have achieved better performances on our datasets -through parameter tuning. To address this threat, we used the GridSearchCV function from the scikit- -learn library with the top two CLE models (i.e., RandomForest andDecisionTree ) to identify the best -parameter combinations. Our implementation explored six parameters with total 5,040 combinations for -RandomForest and five parameters with 360 combinations for DecisionTree. Our results suggest that -most of the default values are identical to those from the best performing combinations identified through -GridSearchCV . We also reevaluated RF and DT with the GridSearchCV suggested values, but did not -find any statistically significant (paired sample t-tests, 𝑝 >0.05) improvements over our already trained -models. -For the DNN algorithms, we did not conduct extensive hyperparameter search due to computational -costs. However, parameter values were selected based on the best practice reported in the deep learning -literature. Moreover, to identify the best DNN models we used validation sets and used EarlyStopping . -Still we may not have been able to achieve the best possible performances from the DNN models during -our evaluations. -7.3 External validity -Although, we have not used any project or code review specific pre-processing, our dataset may not -adequately represent texts from other projects or other software development interactions such as issue -discussions, commit messages, or question /answers on StackExchange. Therefore, our pretrained models -may have degraded performances on other contexts. However, our models can be easily retrained using a -different labeled datasets from other projects or other types of interactions. To facilitate such retraining, -we have made both the source code and instructions to retrain the models publicly available [76]. -7.4 Conclusion validity -To evaluate the performances our models, we have standard metrics such as accuracy, precision, recall, -and F-scores. For the algorithm implementations, we have extensively used state-of-the-art libraries such -as scikit-learn [ 67] and TensorFlow [ 3]. We also used 10-fold cross-validations to evaluate the performances -of each model. Therefore, we do not anticipate any threats to validity arising from the set of metrics, -supporting library selection, and evaluation of the algorithms. -8 CONCLUSION AND FUTURE DIRECTIONS -This paper presents design and evaluation of ToxiCR, a supervised learning-based classifier to identify -toxic code review comments. ToxiCR includes a choice to select one of the ten supervised learning -algorithms, an option to select text vectorization techniques, five mandatory and three optional processing -steps, and a large-scale labeled dataset of 19,651 code review comments. With our rigorous evaluation -of the models with various combinations of preprocessing steps and vectorization techniques, we have -identified the best combination that boosts 95.8% accuracy and 88.9% 𝐹11score. We have released our -dataset, pretrained models, and source code publicly available on Github [ 76]. We anticipate this tool -being helpful in combating toxicity among FOSS communities. As a future direction, we aim to conduct -empirical studies to investigate how toxic interactions impact code review processes and their outcomes -among various FOSS projects. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:29 -REFERENCES -[1]2018. Annotation instructions for Toxicity with sub-attributes. https://github.com/conversationai/conversationai. -github.io/blob/main/crowdsourcing_annotation_schemes/toxicity_with_subattributes.md -[2]2018. Toxic Comment Classification Challenge. https://www.kaggle.com/c/jigsaw-toxic-comment-classification- -challenge -[3]Martín Abadi, Paul Barham, Jianmin Chen, Zhifeng Chen, Andy Davis, Jeffrey Dean, Matthieu Devin, Sanjay -Ghemawat, Geoffrey Irving, Michael Isard, Manjunath Kudlur, Josh Levenberg, Rajat Monga, Sherry Moore, Derek G. -Murray, Benoit Steiner, Paul Tucker, Vijay Vasudevan, Pete Warden, Martin Wicke, Yuan Yu, and Xiaoqiang Zheng. -2016. TensorFlow: A System for Large-Scale Machine Learning. In 12th USENIX Symposium on Operating Systems -Design and Implementation (OSDI 16) . USENIX Association, Savannah, GA, 265–283. -[4]Ashutosh Adhikari, Achyudh Ram, Raphael Tang, and Jimmy Lin. 2019. Docbert: Bert for document classification. -arXiv preprint arXiv:1904.08398 (2019). -[5]Sonam Adinolf and Selen Turkay. 2018. Toxic behaviors in Esports games: player perceptions and coping strategies. In -Proceedings of the 2018 Annual Symposium on Computer-Human Interaction in Play Companion Extended Abstracts . -365–372. -[6]Toufique Ahmed, Amiangshu Bosu, Anindya Iqbal, and Shahram Rahimi. 2017. SentiCR: a customized sentiment -analysis tool for code review interactions. In 2017 32nd IEEE/ACM International Conference on Automated Software -Engineering (ASE) . IEEE, 106–111. -[7]Conversation AI. [n.d.]. What if technology could help improve conversations online? https://www.perspectiveapi.com/ -[8]Conversation AI. 2018. Annotation instructions for Toxicity with sub-attributes. -https://github.com/conversationai/conversationai.github.io/blob/master /crowdsourc- -ing_annotation_schemes/toxicity_with_subattributes.md. -[9]Basemah Alshemali and Jugal Kalita. 2020. Improving the reliability of deep neural networks in NLP: A review. -Knowledge-Based Systems 191 (2020), 105210. -[10]Ashley A Anderson, Sara K Yeo, Dominique Brossard, Dietram A Scheufele, and Michael A Xenos. 2018. Toxic talk: -How online incivility can undermine perceptions of media. International Journal of Public Opinion Research 30, 1 -(2018), 156–168. -[11]Anonymous. 2014. Leaving Toxic Open Source Communities. https://modelviewculture.com/pieces/leaving-toxic- -open-source-communities -[12] Hayden Barnes. 2020. Toxicity in Open Source. https://boxofcables.dev/toxicity-in-linux-and-open-source/ -[13]Joseph Berkson. 1944. Application of the logistic function to bio-assay. Journal of the American statistical association -39, 227 (1944), 357–365. -[14]Meghana Moorthy Bhat, Saghar Hosseini, Ahmed Hassan, Paul Bennett, and Weisheng Li. 2021. Say ‘YES’to Positivity: -Detecting Toxic Language in Workplace Communications. In Findings of the Association for Computational Linguistics: -EMNLP 2021 . 2017–2029. -[15]Piotr Bojanowski, Edouard Grave, Armand Joulin, and Tomas Mikolov. 2017. Enriching word vectors with subword -information. Transactions of the Association for Computational Linguistics 5 (2017), 135–146. -[16] Amiangshu Bosu and Jeffrey C Carver. 2013. Impact of peer code review on peer impression formation: A survey. In -2013 ACM/IEEE International Symposium on Empirical Software Engineering and Measurement . IEEE, 133–142. -[17]Amiangshu Bosu, Anindya Iqbal, Rifat Shahriyar, and Partha Chakroborty. 2019. Understanding the Motivations, -Challenges and Needs of Blockchain Software Developers: A Survey. Empirical Software Engineering 24, 4 (2019), -2636–2673. -[18] Leo Breiman. 1996. Bagging predictors. Machine learning 24, 2 (1996), 123–140. -[19]Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, -Pranav Shyam, Girish Sastry, Amanda Askell, et al .2020. Language models are few-shot learners. Advances in neural -information processing systems 33 (2020), 1877–1901. -[20]Fabio Calefato, Filippo Lanubile, and Nicole Novielli. 2017. EmoTxt: a toolkit for emotion recognition from text. In -2017 seventh international conference on Affective Computing and Intelligent Interaction Workshops and Demos -(ACIIW) . IEEE, 79–80. -[21]Kevin Daniel André Carillo, Josianne Marsan, and Bogdan Negoita. 2016. Towards Developing a Theory of Toxicity in -the Context of Free/Open Source Software & Peer Production Communities. SIGOPEN 2016 (2016). -[22]Hao Chen, Susan McKeever, and Sarah Jane Delany. 2019. The Use of Deep Learning Distributed Representations in -the Identification of Abusive Text. In Proceedings of the International AAAI Conference on Web and Social Media , -Vol. 13. 125–133. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:30•Sarker, et al. -[23]Savelie Cornegruta, Robert Bakewell, Samuel Withey, and Giovanni Montana. 2016. Modelling Radiological Language -with Bidirectional Long Short-Term Memory Networks. EMNLP 2016 (2016), 17. -[24] Corinna Cortes and Vladimir Vapnik. 1995. Support-vector networks. Machine learning 20, 3 (1995), 273–297. -[25]Cristian Danescu-Niculescu-Mizil, Moritz Sudhof, Dan Jurafsky, Jure Leskovec, and Christopher Potts. 2013. A -Computational Approach to Politeness with Application to Social Factors. In 51st Annual Meeting of the Association -for Computational Linguistics . ACL, 250–259. -[26]R Van Wendel De Joode. 2004. Managing conflicts in open source communities. Electronic Markets 14, 2 (2004), -104–113. -[27]Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. BERT: Pre-training of Deep Bidirectional -Transformers for Language Understanding. (June 2019), 4171–4186. https://doi.org/10.18653/v1/N19-1423 -[28] Maeve Duggan. 2017. Online harassment 2017. (2017). -[29]Carolyn D Egelman, Emerson Murphy-Hill, Elizabeth Kammer, Margaret Morrow Hodges, Collin Green, Ciera Jaspan, -and James Lin. 2020. Predicting developers’ negative feelings about code review. In 2020 IEEE/ACM 42nd International -Conference on Software Engineering (ICSE) . IEEE, 174–185. -[30]Ahmed Elnaggar, Bernhard Waltl, Ingo Glaser, Jörg Landthaler, Elena Scepankova, and Florian Matthes. 2018. Stop -Illegal Comments: A Multi-Task Deep Learning Approach. In Proceedings of the 2018 Artificial Intelligence and Cloud -Computing Conference . 41–47. -[31]Nelly Elsayed, Anthony S Maida, and Magdy Bayoumi. 2019. Deep Gated Recurrent and Convolutional Network -Hybrid Model for Univariate Time Series Classification. International Journal of Advanced Computer Science and -Applications 10, 5 (2019). -[32] Samir Faci. 2020. The Toxicity Of Open Source. https://www.esamir.com/20/12/23/the-toxicity-of-open-source/ -[33]Isabella Ferreira, Jinghui Cheng, and Bram Adams. 2021. The" Shut the f** k up" Phenomenon: Characterizing -Incivility in Open Source Code Review Discussions. Proceedings of the ACM on Human-Computer Interaction 5, -CSCW2 (2021), 1–35. -[34]Anna Filippova and Hichang Cho. 2016. The effects and antecedents of conflict in free and open source software -development. In Proceedings of the 19th ACM Conference on Computer-Supported Cooperative Work & Social -Computing . 705–716. -[35]LibreOffice: The Document Foundation. [n.d.]. Code of Conduct. https://www.documentfoundation.org/foundation/ -code-of-conduct/. -[36]Jerome H Friedman. 2001. Greedy function approximation: a gradient boosting machine. Annals of statistics (2001), -1189–1232. -[37]Spiros V Georgakopoulos, Sotiris K Tasoulis, Aristidis G Vrahatis, and Vassilis P Plagianakos. 2018. Convolutional -neural networks for toxic comment classification. In Proceedings of the 10th Hellenic Conference on Artificial Intelligence . -1–6. -[38]Alex Graves and Jürgen Schmidhuber. 2005. Framewise phoneme classification with bidirectional LSTM and other -neural network architectures. Neural networks 18, 5-6 (2005), 602–610. -[39]Isuru Gunasekara and Isar Nejadgholi. 2018. A review of standard text classification practices for multi-label toxicity -identification of online content. In Proceedings of the 2nd workshop on abusive language online (ALW2) . 21–25. -[40] Sanuri Dananja Gunawardena, Peter Devine, Isabelle Beaumont, Lola Garden, Emerson Rex Murphy-Hill, and Kelly -Blincoe. 2022. Destructive Criticism in Software Code Review Impacts Inclusion. (2022). -[41] Laura Hanu and Unitary team. 2020. Detoxify. Github. https://github.com/unitaryai/detoxify. -[42]Tin Kam Ho. 1995. Random decision forests. In Proceedings of 3rd international conference on document analysis and -recognition , Vol. 1. IEEE, 278–282. -[43]Sepp Hochreiter and Jürgen Schmidhuber. 1997. Long short-term memory. Neural computation 9, 8 (1997), 1735–1780. -[44]Hossein Hosseini, Sreeram Kannan, Baosen Zhang, and Radha Poovendran. 2017. Deceiving google’s perspective api -built for detecting toxic comments. arXiv preprint arXiv:1702.08138 (2017). -[45]Zhiheng Huang, Wei Xu, and Kai Yu. 2015. Bidirectional LSTM-CRF models for sequence tagging. arXiv preprint -arXiv:1508.01991 (2015). -[46]N. Imtiaz, J. Middleton, J. Chakraborty, N. Robson, G. Bai, and E. Murphy-Hill. 2019. Investigating the Effects -of Gender Bias on GitHub. In 2019 IEEE/ACM 41st International Conference on Software Engineering (ICSE) . -700–711. -[47]Md Zahidul Islam, Jixue Liu, Jiuyong Li, Lin Liu, and Wei Kang. 2019. A semantics aware random forest for text -classification. In Proceedings of the 28th ACM international conference on information and knowledge management . -1061–1070. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:31 -[48]Carlos Jensen, Scott King, and Victor Kuechler. 2011. Joining free/open source software communities: An analysis of -newbies’ first interactions on project mailing lists. In 2011 44th Hawaii international conference on system sciences . -IEEE, 1–10. -[49]Rie Johnson and Tong Zhang. 2017. Deep pyramid convolutional neural networks for text categorization. In Proceedings -of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) . 562–570. -[50]Robbert Jongeling, Proshanta Sarkar, Subhajit Datta, and Alexander Serebrenik. 2017. On negative results when using -sentiment analysis tools for software engineering research. Empirical Software Engineering 22, 5 (2017), 2543–2584. -[51]Diederik P. Kingma and Jimmy Ba. 2015. Adam: A Method for Stochastic Optimization. (2015). http://dblp.uni- -trier.de/db/conf/iclr/iclr2015.html#KingmaB14 -[52]Kamran Kowsari, Donald E Brown, Mojtaba Heidarysafa, Kiana Jafari Meimandi, Matthew S Gerber, and Laura E -Barnes. 2017. Hdltex: Hierarchical deep learning for text classification. In 2017 16th IEEE international conference on -machine learning and applications (ICMLA) . IEEE, 364–371. -[53]Deepak Kumar, Patrick Gage Kelley, Sunny Consolvo, Joshua Mason, Elie Bursztein, Zakir Durumeric, Kurt Thomas, -and Michael Bailey. 2021. Designing toxic content classification for a diversity of perspectives. In Seventeenth Symposium -on Usable Privacy and Security (SOUPS 2021) . 299–318. -[54]Keita Kurita, Anna Belova, and Antonios Anastasopoulos. 2019. Towards Robust Toxic Content Classification. arXiv -preprint arXiv:1912.06872 (2019). -[55]Zijad Kurtanović and Walid Maalej. 2018. On user rationale in software engineering. Requirements Engineering 23, 3 -(2018), 357–379. -[56]Bin Lin, Fiorella Zampetti, Gabriele Bavota, Massimiliano Di Penta, Michele Lanza, and Rocco Oliveto. 2018. Sentiment -analysis for software engineering: How far can we go?. In Proceedings of the 40th international conference on software -engineering . 94–104. -[57]Ilya Loshchilov and Frank Hutter. 2018. Decoupled Weight Decay Regularization. International Conference on Learning -Representations (2018). -[58]Tomas Mikolov, Ilya Sutskever, Kai Chen, Greg S Corrado, and Jeff Dean. 2013. Distributed representations of words -and phrases and their compositionality. In Advances in neural information processing systems . 3111–3119. -[59]Courtney Miller, Sophie Cohen, Daniel Klug, Bodgan Vasilescu, and Christian Kästner. 2022. “Did You Miss My -Comment or What?” Understanding Toxicity in Open Source Discussions. In International Conference on Software -Engineering, ICSE, ACM (2022) (ICSE) . IEEE, ACM. -[60]Pushkar Mishra, Helen Yannakoudakis, and Ekaterina Shutova. 2018. Neural Character-based Composition Models for -Abuse Detection. EMNLP 2018 (2018), 1. -[61]Murtuza Mukadam, Christian Bird, and Peter C Rigby. 2013. Gerrit software code review data from android. In 2013 -10th Working Conference on Mining Software Repositories (MSR) . IEEE, 45–48. -[62]Dawn Nafus, James Leach, and Bernhard Krieger. 2006. Gender: Integrated report of findings. FLOSSPOLS, Deliverable -D16 (2006). -[63]Chikashi Nobata, Joel Tetreault, Achint Thomas, Yashar Mehdad, and Yi Chang. 2016. Abusive language detection in -online user content. In Proceedings of the 25th international conference on world wide web . 145–153. -[64]Nicole Novielli, Daniela Girardi, and Filippo Lanubile. 2018. A benchmark study on sentiment analysis for software -engineering research. In 2018 IEEE/ACM 15th International Conference on Mining Software Repositories (MSR) . -IEEE, 364–375. -[65] OpenStack. [n.d.]. OpenStack Code of Conduct. https://wiki.openstack.org/wiki/Conduct. -[66]Rajshakhar Paul, Amiangshu Bosu, and Kazi Zakia Sultana. 2019. Expressions of Sentiments during Code Reviews: -Male vs. Female. In Proceedings of the 26th IEEE International Conference on Software Analysis, Evolution and -Reengineering (Hangzhou, China) (SANER ‘19) . IEEE. -[67]Fabian Pedregosa, Gaël Varoquaux, Alexandre Gramfort, Vincent Michel, Bertrand Thirion, Olivier Grisel, Mathieu -Blondel, Peter Prettenhofer, Ron Weiss, Vincent Dubourg, et al .2011. Scikit-learn: Machine learning in Python. the -Journal of machine Learning research 12 (2011), 2825–2830. -[68]Jeffrey Pennington, Richard Socher, and Christopher D Manning. 2014. Glove: Global vectors for word representation. -InProceedings of the 2014 conference on empirical methods in natural language processing (EMNLP) . 1532–1543. -[69] Android Open Source Project. [n.d.]. Code of Conduct. https://source.android.com/setup/cofc. -[70]Huilian Sophie Qiu, Bogdan Vasilescu, Christian Kästner, Carolyn Denomme Egelman, Ciera Nicole Christopher -Jaspan, and Emerson Rex Murphy-Hill. 2022. Detecting Interpersonal Conflict in Issues and Code Review: Cross -Pollinating Open-and Closed-Source Approaches. (2022). -[71] J. Ross Quinlan. 1986. Induction of decision trees. Machine learning 1, 1 (1986), 81–106. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.1:32•Sarker, et al. -[72]Israr Qureshi and Yulin Fang. 2011. Socialization in open source software projects: A growth mixture modeling -approach. Organizational Research Methods 14, 1 (2011), 208–238. -[73]Naveen Raman, Minxuan Cao, Yulia Tsvetkov, Christian Kästner, and Bogdan Vasilescu. 2020. Stress and Burnout in -Open Source: Toward Finding, Understanding, and Mitigating Unhealthy Interactions. In International Conference on -Software Engineering, New Ideas and Emerging Results (ICSE) . ACM, TBD. -[74]David E Rumelhart, Geoffrey E Hinton, and Ronald J Williams. 1986. Learning representations by back-propagating -errors.nature323, 6088 (1986), 533–536. -[75]Maarten Sap, Dallas Card, Saadia Gabriel, Yejin Choi, and Noah A Smith. 2019. The risk of racial bias in hate speech -detection. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics . 1668–1678. -[76]Jaydeb Sarkar, Asif Turzo, Ming Dong, and Amiangshu Bosu. 2022. ToxiCR: Replication package. Github. -https://github.com/WSU-SEAL/ToxiCR. -[77]Jaydeb Sarker, Asif Kamal Turzo, and Amiangshu Bosu. 2020. A benchmark study of the contemporary toxicity -detectors on software engineering interactions. In 2020 27th Asia-Pacific Software Engineering Conference (APSEC) -(Singapore). 218–227. https://doi.org/10.1109/APSEC51365.2020.00030 -[78]Jaydeb Sarker, Asif Kamal Turzo, Ming Dong, and Amiangshu Bosu. 2022. WSU SEAL implementation of the -STRUDEL Toxicity detector. https://github.com/WSU-SEAL/toxicity-detector/tree/master/WSU_SEAL. -[79]Carl-Erik Särndal, Bengt Swensson, and Jan Wretman. 2003. Model assisted survey sampling . Springer Science & -Business Media. -[80]Robert E Schapire. 2003. The boosting approach to machine learning: An overview. Nonlinear estimation and -classification (2003), 149–171. -[81]Megan Squire and Rebecca Gazda. 2015. FLOSS as a Source for Profanity and Insults: Collecting the Data. In 2015 -48th Hawaii International Conference on System Sciences . IEEE, 5290–5298. -[82]Saurabh Srivastava, Prerna Khurana, and Vartika Tewari. 2018. Identifying aggression and toxicity in comments -using capsule network. In Proceedings of the First Workshop on Trolling, Aggression and Cyberbullying (TRAC-2018) . -98–105. -[83]Igor Steinmacher and Marco Aurélio Gerosa. 2014. How to support newcomers onboarding to open source software -projects. In IFIP International Conference on Open Source Systems . Springer, 199–201. -[84]Pannavat Terdchanakul, Hideaki Hata, Passakorn Phannachitta, and Kenichi Matsumoto. 2017. Bug or not? bug -report classification using n-gram idf. In 2017 IEEE international conference on software maintenance and evolution -(ICSME) . IEEE, 534–538. -[85]Ameya Vaidya, Feng Mai, and Yue Ning. 2020. Empirical analysis of multi-task learning for reducing identity bias in -toxic comment detection. Proceedings of the International AAAI Conference on Web and Social Media 14 (2020), -683–693. -[86]Betty van Aken, Julian Risch, Ralf Krestel, and Alexander Löser. 2018. Challenges for Toxic Comment Classification: -An In-Depth Error Analysis. Proceedings of the 2nd Workshop on Abusive Language Online (ALW2) (2018), 33–42. -[87]Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Łukasz Kaiser, and Illia -Polosukhin. 2017. Attention is all you need. Advances in neural information processing systems 30 (2017). -[88]Susan Wang and Zita Marinho. 2020. Nova-Wang at SemEval-2020 Task 12: OffensEmblert: An Ensemble ofOffensive -Language Classifiers. In Proceedings of the Fourteenth Workshop on Semantic Evaluation . 1587–1597. -[89]Mengzhou Xia, Anjalie Field, and Yulia Tsvetkov. 2020. Demoting Racial Bias in Hate Speech Detection. Proceedings -of the Eighth International Workshop on Natural Language Processing for Social Media (2020), 7–14. -[90]Zhilin Yang, Zihang Dai, Yiming Yang, Jaime Carbonell, Russ R Salakhutdinov, and Quoc V Le. 2019. Xlnet: -Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems -32 (2019). -[91]Sara Zaheri, Jeff Leath, and David Stroud. 2020. Toxic Comment Classification. SMU Data Science Review 3, 1 (2020), -13. -[92]Marcos Zampieri, Preslav Nakov, Sara Rosenthal, Pepa Atanasova, Georgi Karadzhov, Hamdy Mubarak, Leon Derczyn- -ski, Zeses Pitenis, and Çağrı Çöltekin. 2020. SemEval-2020 Task 12: Multilingual Offensive Language Identification in -Social Media (OffensEval 2020). In Proceedings of the Fourteenth Workshop on Semantic Evaluation . 1425–1447. -[93]Justine Zhang, Jonathan P Chang, Cristian Danescu-Niculescu-Mizil, Lucas Dixon, Yiqing Hua, Nithum Tahin, and -Dario Taraborelli. 2018. Conversations Gone Awry: Detecting Early Signs of Conversational Failure.. In Proceedings of -the 56th Annual Meeting of the Association for Computational Linguistics. , Vol. 1. -[94]Xiang Zhang, Junbo Zhao, and Yann LeCun. 2015. Character-level convolutional networks for text classification. -Advances in neural information processing systems 28 (2015), 649–657. -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023.Automated Identification of Toxic Code Reviews Using ToxiCR •1:33 -[95]Yukun Zhu, Ryan Kiros, Rich Zemel, Ruslan Salakhutdinov, Raquel Urtasun, Antonio Torralba, and Sanja Fidler. 2015. -Aligning Books and Movies: Towards Story-Like Visual Explanations by Watching Movies and Reading Books. In The -IEEE International Conference on Computer Vision (ICCV) . -ACM Trans. Softw. Eng. Methodol., Vol. 32, No. 1, Article 1. Publication date: January 2023. \ No newline at end of file