--- license: cc-by-4.0 --- # 1. Project Structure ## 1.a. Goals and Values ### Project Overview BigCode is an open scientific collaboration working on the responsible development and use of large language models for code, aiming to empower the machine learning and open source communities through open governance. Code LLMs enable the completion and synthesis of code, both from other code snippets and natural language descriptions, and can be used across a wide range of domains, tasks, and programming languages. These models can, for example, assist professional and citizen developers with building new applications. One of the challenges typically faced by researchers working on code LLMs is the lack of transparency around the development of these systems. While a handful of papers on code LLMs have been published, they do not always give full insight into the development process, which hinders both external accountability and the ability of all but a few well funded research labs to meaningfully participate in shaping the technology. BigCode is a community project jointly led by Hugging Face and ServiceNow. Both organizations committed research, engineering, ethics, governance, and legal resources to ensure that the collaboration runs smoothly and makes progress towards the stated goals. ServiceNow Research and Hugging Face have made their respective compute clusters available for large-scale training of the BigCode models, and Hugging Face hosts the datasets, models, and related applications from the community to make it easy for everyone to access and use. An open-invitation was extended to the global AI research community to join forces on the development of state-of-the-art code LLMs, with a focus on research topics such as: * Constructing a representative evaluation suite for code LLMs, covering a diverse set of tasks and programming languages * Developing new methods for faster training and inference of LLMs * The legal, ethics, and governance aspects of code LLMs The BigCode project is conducted in the spirit of open science. Datasets, models, and experiments are developed through collaboration and released with permissive licenses back to the community. All technical governance takes place within working groups and task forces across the community. As code LLMs are developed with data from the open-source community, we believe open governance can help to ensure that these models are benefiting the larger AI community. We developed tools to give code creators agency over whether their source code is included in the training data, and to find approaches that give attribution to developers when models output near-copies of the training data contained in The Stack. ### Technical and Performance Goals The overarching technical goal envisioned before the project was announced was to train and release a 12-billion parameter model that matches Codex as described in [this research paper](https://arxiv.org/abs/2107.03374). This model from OpenAI is not released and is only available as API service under a cushman-001, although it’s not entirely clear if this model matches the one described in the paper. It has also been [suggested](https://thakkarparth007.github.io/copilot-explorer/posts/copilot-internals#other-random-tidbits) that this model is used in Github CoPilot. Our original plan was to compare model performance on HumanEval and APPS, but along the way, we recognized the need for creating an extensive evaluation suite for Code LLMs. The project ended up breaking the challenge into development phases, starting with the collection of permissively licensed repositories from Github. This effort resulted in the creation of The Stack, a source code dataset that marked the first milestone for the project. This initial phase was carried out over several months prior to the official launch of BigCode. Two cycles of model development were conducted by the BigCode community. The first cycle took place in November-December 2022, and culminated with the release of SantaCoder, a 1.1B parameter model trained on the Java, JavaScript, and Python code from The Stack. In the next cycle, which was held from January to April 2023, the community scaled-up their efforts and trained 15.5B parameter models on 1T tokens from The Stack. The resulting StarCoder models either match or surpass the code-cushman-001 model on a variety of coding benchmarks. ### Social Impact Dimensions and Considerations Technical goals and considerations of social impact go hand in hand, and participate equally in _responsible_ development of code LLMs. Within the BigCode project, this means that organizational and technical choices were jointly shaped by the pursuit of the performance goals outlined above and by a best-effort approach to accounting for their impact on various categories of external stakeholders. In particular, participants of the BigCode project focused on the following three dimensions of social impact: * **Consent of data subjects**: the success of code LLMs depends on their training data, which is the product of the professional and volunteer work of software developers. Since training large models constitutes a novel use and transformation of this work, it poses new questions and challenges with respect to the wishes and rights of the developers. * **Privacy**: investigations into the behaviors of previous large-scale code LLMs outlined privacy risks when the models can be prompted to generate private information contained in its training data. Addressing these risks was the focus of several related efforts in BigCode. * **Software safety and security**: [recent work](https://arxiv.org/abs/2207.14157) has also shed light on different hazards that are unique to or exacerbated by code LLMs, including their dual use potential in facilitating malware generation or their likelihood of recommending code that includes security flaws. We found that while these considerations did sometimes result in trade-offs between the performance goals and social impact concerns, they were more often better addressed by developing new technical and organizational tools, which we describe in the rest of this document and share as an important outcome of the BigCode project so they can be leveraged by future similar endeavors. ## 1.b. Organizers and Participants ### Inception The idea for the BigCode Project came about in Utrecht during a discussion initiated by Harm de Vries (ServiceNow Research) with Thomas Wolf (Hugging Face). Inspired by the BigScience Project, Harm recognized the shared vision of Hugging Face to responsibly develop open and responsible large language models for code, and approached Thomas to explore the idea of a jointly led open-scientific collaboration with the global machine learning and open source communities. As it turns out, the visions were indeed aligned, and work got started to initiate the project. A research collaboration agreement between ServiceNow and Hugging Face created the enabling framework for the project, and set out the terms for rallying the broader scientific community at large to work towards developing, training, exploring, and releasing large foundation models for code. The scope of the collaboration covered the preparation of training data, including developing tools for downloading publicly accessible code data and for running license detectors, developing tools for filtering data sources based on approved licenses and file type, and to release this training data to the AI community. The scope also covered the training of dense transformer models that adopt the mechanism of self-attention through the Megatron-LM architecture, training of retrieval augmented code generation models, and developing tools to diagnose instabilities arising from training transformer models at scale. The collaboration would also prepare an evaluation suite with the help of the scientific community 1) to develop the tools and scripts needed to use existing program synthesis benchmarks such as HumanEval and CodexGlue, and 2) to construct and release openly available benchmarks that measure desirable capabilities of large multi-lingual code LLMs and tasks such as program synthesis, text2code, and code summarization. Key milestones identified during the initial stages were focussed on developing a community engagement plan, a first attempt at constructing an evaluation suite over multiple programming languages, investigating viable sources of data, and then training and releasing a 12B parameter model that matches Codex performance on HumanEval and APPS. ### Participants BigCode is a research collaboration and is open to participants who 1. have a professional research background and 2. are able to commit time to the project. In general, we expect applicants to be affiliated with a research organization (either in academia or industry) and work on the technical/ethical/legal aspects of LLMs for coding applications. Throughout the project the community invited guest subject matter experts to participate in certain discussions, and this resulted in a lift in the number of participants in chat channels relative to the number of researchers that had formally applied to participate in the research. BigCode has 675 participants with 629 members across the research community (including from Hugging Face and ServiceNow) from 62 countries. The top 5 countries include USA (222), India (60), UK (36), Canada (35), and Germany (30). The community communicates across a total of 48 Slack channels, including Steering Committee (3 channels), Working Groups (7 channels), Task Forces (25 channels), and General Community (13 channels). Everyone who joins the project is required to follow the [BigCode Code of Conduct](https://www.bigcode-project.org/docs/about/code_of_conduct/), understand [how we manage intellectual property](https://www.bigcode-project.org/docs/about/ip/), and are encouraged to introduce themselves, and to join any working group or task force that aligns to their own interests. If a group does not cover their interests, they are encouraged to pitch their ideas and to take a leadership role for a new working group or task force with the approval of the Steering Committee. Researchers that wish to cite StarCoder are asked to please use the following: TBD ### Project Governance The BigCode project is governed by a steering committee jointly led by Harm de Vries (ServiceNow) and Leandro von Werra (Hugging Face), and supported by a core team comprised of Raymond Li, Denis Kocetkov, and Sean Hughes from ServiceNow, and Carlos Muñoz Ferrandis, Loubna Ben Allal, and Thomas Wolf of Hugging Face. Through the course of the project, additional members were added to the core team, including Yacine Jernite, Armel Randy, Joel Lamy-Poirier. The Steering Committee is effectively responsible for organizing and managing the project (including research strategy and publication goals), and provides oversight across all working groups. Decisions that cannot be addressed at the community level would be elevated to the lead of the Working Group for facilitated discussion, with further inputs and tie-breaker decision making by the Steering Committee as a last resort. Governance for the project is open, meaning that the BigCode project encourages anyone from the community to join any working group or task force of interest, and for them to engage and contribute to work and decision making in the group. ### Timeline, Milestones, and Community Events The BigCode project was announced on [September 26, 2022](https://twitter.com/BigCodeProject/status/1574427555871875072?s=20). We shared the goal of the project to train a state of the art ~15B parameter language model for code that will be trained using[ ](https://twitter.com/ServiceNowRSRCH)the [ServiceNow Research](https://www.servicenow.com/research/) in-house GPU cluster. With an adapted version of Megatron-LM, we planned to train the large model on distributed infrastructure. On [October 6, 2022](https://youtu.be/8cUpsXIEbAo) ServiceNow and Hugging Face held a webinar with the BigCode Community to provide strategic direction for the project and research goals. On [October 27, 2022](https://twitter.com/BigCodeProject/status/1585631176353796097?s=20) we introduced The Stack, a large dataset of more than 3 TB of permissively licensed source code. [Our paper](https://arxiv.org/abs/2211.15533) described the details of the dataset collection, presented a brief dataset analysis, and showed promising results on the HumanEval benchmark. Our experimental results show that near-deduplication is an important pre-processing step for achieving competitive results on text2code benchmarks. We released all permissively licensed files for 30 common programming languages, along with a near-deduplicated version. On [November 15, 2022](https://twitter.com/BigCodeProject/status/1592569651086905344?s=20) we introduced a new tool called “Am I in The Stack” that allows developers to check whether any data from their GitHub repositories is included in The Stack. We also introduced v1 of the BigCode Opt-Out process, which gives agency back to Developers by providing a way for them to request that their data be removed from the dataset. On [November 23, 2022](https://twitter.com/LoubnaBenAllal1/status/1595457541592346634?s=20) Loubna Ben Allal shared details of our initial approach for how we planned to tackle the de-identification to remove personally identifiable information (PII) from The Stack. On [November 29, 2022](https://twitter.com/BigCodeProject/status/1597589730425974786?s=20) we shared the Weights and Biases dashboards for our first models so that the broader community could follow along. On [December 1, 2022](https://twitter.com/BigCodeProject/status/1598345535190179843?s=20) we released The Stack v1.1, with 358 programming languages included, and more than double the data, going from 3Tb to 6.4TB, with the help of the legal tech community that identified 193 viable permissive source code license types. Before releasing v1.1 we also removed the first batch of repositories based on opt-out requests. On [December 2, 2022](https://twitter.com/BigCodeProject/status/1598734387247550481?s=20) we held an in-person meetup alongside NeurIPS 2022 in New Orleans with more than 75 members of the BigCode community, where we were able to make the connections to foster greater awareness and understanding of the BigCode project. On [December 9, 2022](https://twitter.com/BigCodeProject/status/1601133018714112000?s=20), a member of the BigCode community held a similar meetup at EMNLP 2022 in Abhu Dhabi, another opportunity to raise awareness of BigCode and to discuss our project with the NLP research community. On [December 12, 2022](https://twitter.com/BigCodeProject/status/1602372753008386049?s=20) we sent out another message to raise awareness of “Am I in The Stack” and to inform developers about the option to opt-out from the dataset. On [December 14, 2022](https://youtu.be/Kh8yXfJJfU4) Hugging Face and ServiceNow held a second webinar with the BigCode Community to review progress and provide an update on plans for ongoing research towards the 15B parameter model. On [December 22, 2022](https://twitter.com/BigCodeProject/status/1605958778330849281?s=20) we released [SantaCoder,](https://huggingface.co/bigcode/santacoder) a 1.1B multilingual large language model for code that outperforms much larger open-source models on both left-to-right generation, and infilling. The SantaCoder models are licensed under an open & responsible AI model license (CodeML [OpenRAIL-M v0.1](https://huggingface.co/spaces/bigcode/license)). These are AI-specific licenses enabling free use and distribution of the model while setting specific use restrictions (e.g. malware generation). We published a [detailed technical report ](https://arxiv.org/abs/2301.03988)that included details of all the key contributors to the development of the model. On [February 1, 2022](https://twitter.com/utopiah/status/1620722505664319488?s=20) Members of the BigCode core team were invited to meet with the European Parliament Innovation Lab. At this meeting we [shared details](https://twitter.com/utopiah/status/1620735424351322114?s=20) of the project and answered questions from members of the Lab. Engaging with policymakers and regulators is an important part of the journey to inform and educate key stakeholders from the broader AI ecosystem. On [March 20, 2022 ](https://twitter.com/BigCodeProject/status/1637874705645584384?s=20)we announced The Stack v1.2 which included The Stack Issues, The Stack Metadata, and The Stack Commits. With this release, we simplified the opt-out process and also removed opt-outs from developers where the request was received by February 2023. Along with this release, we provided access to a dataset of GitHub issues totalling 54GB, and we applied the same opt-out mechanism to these issues. The GitHub issues dataset is more conversational and could be helpful to train models to be used as coding assistants. On [April 13, 2023](https://twitter.com/harmdevries77/status/1646524056538316805?s=20) Inspired by discussions in the training working group, Harm de Vries shared an analysis of Chinchilla scaling laws on how much additional compute resources are needed to create smaller LLMs. These insights suggest we have not reached the limit of training smaller models on more tokens - an important consideration for future research. On May 4, 2023 BigCode announced StarCoder and StarCoderBase, two code LLMs trained on permissively licensed data from GitHub, including from 80+ programming languages, git commits, GitHub issues, and Jupyter notebooks. Similar to [LLaMA](https://ai.facebook.com/blog/large-language-model-llama-meta-ai/), StarCoderBase is a ~15B parameter model for 1 trillion tokens. On top of StarCoderBase a variant called StarCoder is trained for 35B additional tokens purely on Python. ### Supporting Resources and Funding Understanding the costs of a project like BigCode can help ground conversations about the trade-offs involved in the development of code LLM technology more broadly, helping understand how various private and public institutions may participate in this development and allocate resources to maximize its overall benefits. We outline the major costs in terms of computation resources, human participation, and organization. **Compute** We trained StarCoder on a Hugging Face GPU cluster with 512 A100 80GB GPUs distributed across 64 nodes. We report the carbon footprint of training StarCoderBase. Based on the total number of GPU hours that training took (320,256) and an average power usage of 280W per GPU, this adds up to 89671.68 kWh of electricity consumed during the training process. Multiplied by the carbon intensity of the energy of the us-west-2 AWS location (0.15495 kgCO2e per kWh) and the average Power Usage Effectiveness of 1.2 across AWS datacenters, this results in 16.68 tonnes of CO2eq emitted. **ServiceNow and Hugging Face employees working on BigCode** The estimated time commitment for the duration of the project for employees of the host institutions corresponds to 6 full-time employees for the duration of the project. **Estimated volunteer hours across the project** The time commitment from volunteers is harder to estimate given the large number of participants and the variety of time investments across phases and participants. At a minimum, we estimate overall time commitment from volunteers matched time commitment from employees of the host institutions. **Community events and appreciation** ServiceNow and Hugging Face organized a community meetup that coincided with NeurIPS 2022 in New Orleans, USA. The budget for the event was approximately \$6,000 from ServiceNow Research for the venue with hospitality. Hugging face also provided promotional items including stickers and tshirts at the event, and sent named contributors to the research paper complementary BigCode branded tshirts. **Crowdsourcing costs** Hugging Face funded the data annotation services from Toloka, with a total outlay of \$xxx. Since this was a research project, Toloka agreed to waive the fees for running the annotation tasks on their platform. # 2. Data and Model Governance ## 2.a. Data Governance ### Data Collection and Management Plan In the course of the BigCode project, we collected two main datasets. The primary training dataset is The Stack, which was obtained by gathering public code files, issues, and commits from GitHub through the BigQuery Service. Additionally, we collected a dataset of annotations of several kinds of private information on a subset of The Stack to support our privacy risk mitigation efforts. The legal basis for data collection under fair use and with regards to GDPR and the corresponding case law are still evolving. In this context, the data collection and data management plans were carefully crafted with support from leading experts in the open source and legal tech community that participated in the Legal, Ethics, Governance Working Group in a best-effort approach to reflect current understandings of legal requirements for data collection and management. **The Stack Dataset Access and Management** The StarCoder model was trained on The Stack v1.2, which exclusively contains 6.4TB of [permissively licensed](https://blueoakcouncil.org) data from GitHub repositories, processed from an original source dataset of 102TB. Access and management follow the following schema: * **What data can be accessed:** the 6.4TB of processed data can be accessed through the Hugging Face Hub, while the original 102TB are only accessible to the stewards of the project for the purposes of enabling the research and to support future internal and external requirements that may arise, for example to search the full dataset to recall licenses, determine code provenance, and attribution. * **What are the conditions for accessing the data:** users are able to inspect the dataset via the Dataset Card and embedded Dataset Preview, but are required to agree to the [Terms of Use](https://huggingface.co/datasets/bigcode/the-stack#terms-of-use-for-the-stack) for The Stack before being able to download it. This included the requirements to 1) abide by the terms of original source code licenses, including attribution clauses when required (The Stack provides provenance information for each data point), 2) agree to update copies of The Stack to the most recent usable version specified [here](https://huggingface.co/datasets/bigcode/the-stack/discussions/7), and 3) include the Terms of Use and require users to agree to it if a copy is to be hosted, shared, or otherwise provided. As of May 3, 2023, The Stack had been downloaded 50,200 times. * **How can a data subject request that their data be removed:** we provide an opt-out form that lets people opt out of having any code or text they put on GitHub be included in The Stack. Additionally, anyone who is concerned about specific data they have encountered in The Stack, for example relating to PII, malicious code, or code that has an incorrect license or attribution can email contact@ bigcode-project.org. * **How often is the data updated:** For as long as we are maintaining The Stack dataset, we will provide regular updates to the dataset to remove data that has been flagged since the last version. This includes data that has been opted out, and data that was flagged as malicious code since the previous release. The current plan is to update the dataset every 3 months, although the schedule may change based on the volume of requests received. If we are not in a position to continue maintaining the dataset, we plan to stop distributing it in its current format and update its terms of use to limit its range of applications further. At the time of the data processing for the StarCoder model training, 44 people had opted out of The Stack and associated repositories were removed. **PII Dataset Access and Management** In order to support our efforts to mitigate the risk that the model may leak private information, we selected 12,000 samples of code from The Stack and annotated them to detect PII using crowd-sourcing. The resulting dataset was used to train a PII detection model that we used to detect and then mask PII (Names, Emails, IP addresses, Keys, Passwords) from our StarCoder training dataset. * **What data can be accessed:** the data will be hosted as a gated dataset on the Hugging Face Hub. The dataset will be made available to researchers on a case-by-case basis for research projects that require access, in addition to the original team who developed the dataset. * **What are the conditions for accessing the data:** researchers who want to access the dataset need to request access and be approved by the maintainers as well as agree with the datasets Terms of Use * **How can a data subject request that their data be removed:** as a derived dataset of The Stack, the PII dataset will be updated to reflect data that has been opted out from the source dataset. * **How often is the data updated:** similarly, following The Stack terms of use, the PII Dataset will be updated as often as the Stack if some of the files it contains have been opted out. ### Technical Desiderata for the Training Dataset Curation choices the the training datasets were shaped both by the social impact and performance goals of the project. A critical success factor for BigCode is the volume, variety, and validity of training data to support fundamental and applied scientific research within the AI community. The dataset curation working group invested considerable time to clean up data by combining heuristic filtering and manual inspection. **Volume** Generally speaking, more training data leads to better performance, as the LLM has more examples to learn from, however the larger the dataset, the more computational power, time and costs will be needed. There were several discussions within the BigCode community about whether to up-sample or down-sample certain programming languages, as the amount of compute budget allocated to a data source in a given language can significantly affect the model's performance in that language. However, we realized that the largest amount of available data comes from the popular programming languages and would, therefore, benefit a larger group of end-users. After deduplication, we found that several high-resource programming languages, such as C, C++, C\#, Java, Javascript, Python, and PHP, had a similar amount of data ranging from 44--87 GB. This further reinforced our belief that we did not need to drastically re-weigh the existing data distribution. Thus, in this work, we followed the natural distribution of data during training and sampled data sources proportionally to their volume. **Variety** The more diverse the training data, the better the LLM can understand and generate code in different contexts. We ultimately selected 86 of the 358~programming languages in The Stack. Selection criteria targeted all programming languages with more than 500MB of data, as well as languages that were ranked in the top 50 in programming language popularity. In addition, we included dialects of already selected programming languages (e.g., Racket and Scheme for Lisp), but excluded configuration languages (Nix, Puppet, etc.) and languages that are no longer actively supported (ActionScript). We also included data formats like JSON and YAML but limited its data volume. The full list of selected programming languages can be found in the paper. Out of the languages present in MultiPL-E, only D and Swift were not included in the training set. For D, language misclassification of the files led to a very small number of files in The Stack and Swift was omitted due to a human error. We implemented various filters to help with data cleanup, including an XML filter, an Alpha filter to remove files with less that 25% alphabetic characters, a custom HTML filter to target excessive HTML boilerplate and links, as well as JSON and YAML filters to remove most of the data heavy files. **Validity** Invalid or incorrect training data can lead to biases and errors in the model’s predictions, and it is important to ensure the model is trained on data that is representative of the real-world applications and scenarios it is intended to support. The Stack is a highly curated and highly valuable open access contribution to the community. We performed a visual inspection to ensure that we only retain data of high quality. To achieve this, we randomly selected 30,000 files from The Stack for each programming language, categorized them by extension, and kept a maximum of 1,000 files for each extension. We then reached out to the BigCode community for assistance with data inspection. Eighteen community annotators evaluated 300 programming language extensions. After inspection, we excluded 36 extensions and eliminated the long-line filter for 27 extensions. The complete outcomes of the data inspection, including annotator remarks, can be found in [this Google sheet](https://docs.google.com/spreadsheets/d/1Lk-pTk_rXI__fCgixr7ZWSi8wR09Zzd2j_G90J80r00/edit?usp=sharing). **Malicious code** An additional factor in selecting files from the source data was their impact on the dual use potential of the trained model, and in particular its ability to help users generate malicious code. On the HuggingFace platform, where the Stack is hosted, a malicious code detection tool identified 654 files as unsafe. After an initial conversation between participants about the pros and cos on leaving these files in the training dataset, a consensus was reached that the possible improvements in LLM applications to software security applications did not warrant increasing the likelihood that the model may generate malicious code by including known examples. With help from the BigCode community, we removed these files ahead of the release of The Stack v1.2. ### Consent of Data Subjects **Between implicit and explicit consent** One of the goals of BigCode is to give developers agency over their source code and let them decide whether or not it can be used to develop and evaluate LLMs. Software developers typically rely on licenses to express how they want their work to be re-used; in particular, developers who choose Open Source licenses often do so because they want their code to be broadly re-used. This motivated us to start by selecting data from repositories that met the following criteria: * The repository has a license attached - most repositories on GitHub actually do not have a license, which means that the developer retains all their rights to its content. As regards consent, this also means that we have no positive signal that the developer is OK with their code being broadly re-used * The license is an open source license - as mentioned above, open source, while chosen for very different reasons by different people, typically indicates a willingness to have one’s work reused or adapted * The license does not have an attribution clause - attribution is a difficult technical problem for code LLMs. Since we cannot guarantee that the model will be used in a way that attributes its generations to specific training data in a way that satisfies the intent of the licensor, we chose to only keep licenses without an attribution clause Selecting repositories based on licenses is only the first step, however, as many of these licenses were chosen before the recent developments in code LLMs. Thus, we complement this initial approach by also giving repository owners the ability to **opt out** of having their repositories included in The Stack. We see this approach as a meaningful step forward in improving the agency of data subject in the development of code LLMs, and we present both the tools we developed to support it and its known limitations in the rest of this section. **Technical tools to support opt-out** We developed a tool called [Am I in The Stack](https://hf.co/spaces/bigcode/in-the-stack) to help developers inspect The Stack dataset and for them to see whether any of their repositories have been included and that might be used for training LLMs. If that is the case, we show them a custom link that allows them to easily send a request through GitHub, in two clicks if they are already logged in. We chose to mirror the original platform governance by letting the repository owner decide whether code in a repository is included or not in the dataset. This also allows us to validate requests automatically, since the [GitHub username](https://www.bigcode-project.org/docs/about/the-stack/#what-data-can-i-request-be-removed-from-the-stack) must match the one used to submit the request. Validated requests and associated code pointers are stored so that the code does not appear in future versions of The Stack. **Community feedback on the approach** In the period January-March 2023, members of the BigCode project conducted community research with individuals at specific organizations whose data is used in The Stack, namely [The Alan Turing Institute](https://turing.ac.uk) and [The Turing Way](https://the-turing-way.netlify.app/) as well as two open, international workshops [Open Data Day 2023](https://opendataday.org/events/2023/#designing-for-data-rights-in-the-ai-production-pipeline) and [Mozilla Festival 2023](https://schedule.mozillafestival.org/session/KAS9YF-1) with a session titled ‘Designing for Data Rights in the AI Production Pipeline’. These qualitative interviews and participatory co-design workshops included 50 participants primarily from North America and Europe with roles like research scientist, community manager, software engineer, and principal investigator (PI). The outcomes from the community research can be summarized as follows: when it comes to governance of LLM datasets, participants feel that it is both **better to know** AND **better to have a choice**. Most participants had neutral to positive feelings about their permissively licensed data being used to train LLMs. While all had positive impressions of the ``Am I in The Stack'' tool, no one interviewed expressed a desire to actually opt-out. The main takeaway seemed to be that participants found the most value in BigCode governance tools for their ability to raise awareness of data practices and to empower individuals and communities to take actions based on their specific needs. The co-created outputs can be viewed on this [MozFest Miro Board](https://miro.com/app/board/uXjVMeuvLR8=/?share_link_id=159151239611). Additionally, during the first stage of the opt-out process, individuals **who chose to have their data removed from the Stack** were asked to specify the reasons for wanting their code to be excluded from the dataset. The responses revealed a few recurring themes, including: * Preference for an opt-in approach instead of opt-out * Perception that it is unfair to use their code without compensation * Concerns about the current limitations of AI and the potential for model generations to be traced back to their work, resulting in potential legal liability. * Belief that their code is of poor quality and unsuitable for AI training. * Presence of PII in their code, which they do not wish to be publicly exposed. The feedback form also revealed another limitation of the opt-out process. When code is licensed permissively or under a copy-left license, it can be duplicated to another repository, making it challenging to eliminate such copies if the copyright owner chooses to opt-out. More work is necessary to create workable data control and consent mechanisms for the large-scale training data of LLMs. ### Private Information Handling One significant concern with respect to privacy was the risk that the code LLM may generate private information found in its training data, including private tokens or passwords matched with identifiers or email addresses. Additionally, while users can (and have) requested that data be removed from The Stack dataset because it contains personal data, removing specific information from trained model weights after the fact remains an open technical challenge. In order to minimize this risk, we chose to apply automated PII redaction at the pre-processing stage during training. Our first step toward automatic PII redaction consisted in creating an annotated dataset for PII in code data, as we found that neither regular expression-based approaches nor existing commercial software for PII detection met our performance requirements. In doing so, we aimed to balance the constraints of costs (fair compensation), time (the timing and time to complete the work was on the critical path for the project), and quality (to ensure that PII Detection Model training was not impacted). While traditional data annotation services using salaried employees were considered, we decided to work with crowd-workers through Toloka after reviewing several service providers and their compensation practices - and finding that most would not provide sufficient transparency and guarantees about worker compensation. We selected pay and eligible countries of crowd-workers to ensure that 1. the absolute hourly wage was always higher than the US federal minimum wage (\$7.30), and 2. the hourly wage was equivalent to the highest state minimum wage in the US in terms of purchasing power parity (\$16.50 at the time of writing). We engaged 1,399 crowd-workers across 35 countries in annotating a diverse dataset for PII in source code. Our PII detection model, trained on 22,950 secrets, achieves 90% F1 score surpassing regex-based tools, especially for secret keys. The PII annotations are available to approved individuals, and researchers and developers that are granted access are expected to uphold ethical standards and data protection measures. By making it accessible, our aim is to encourage further research and development of PII redaction technology. Finally, we are also releasing **StarCoderData**, the pre-processed version of The Stack used to train the StarCoder model, which has its PII redacted using our model. ## 2.b. Model Governance ### Model Licensing The model is released under an open and responsible AI model license agreement ([BigCode OpenRAIL-M](https://huggingface.co/spaces/bigcode/bigcode-model-license-agreement)) which enables royalty free access and flexible use and sharing of it, while setting specific use restrictions for identified critical scenarios. Most importantly, the license agreement requires stakeholders wishing to share the model or a modified version of it: (i) to place the same set of use restrictions or a similar one in their legal forms; (ii) to keep the model card and provide a similar one or one of better quality when sharing a modified version of the model (FAQ for the model license agreement available here). The BigCode OpenRAIL-M license agreement (i.e. the legal document itself) is available under a CC-BY-4.0 license. Therefore, any stakeholders can freely adopt the same license agreement for their models, or modify it for their specific AI artifacts. For more information about responsible AI licensing, please visit the RAIL Initiative webpage, [The Turing Way Handbook for ML researchers](https://the-turing-way.netlify.app/reproducible-research/licensing/licensing-ml.html) (Alan Turing Insitute), or OECD AI [content](https://oecd.ai/en/wonk/rails-licenses-trustworthy-ai) on RAILs and trustworthy AI principles. ### Attribution Tool With SantaCoder we released a tool for developers to check whether generated source code had been trained on data from The Stack, and if so, the tool would return the likely matches with full attribution. We both offer a fast membership test to check if code was part the pretraining data as well as a full-text search tool. With StarCoder we are releasing a similar tool ([StarCoder Dataset Search](https://huggingface.co/spaces/bigcode/search)) enabling users to check the origin of the model output and respect any licensing conditions (if any).