jsaizant commited on
Commit
55ecb9a
1 Parent(s): 5058414

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +202 -1
README.md CHANGED
@@ -1,3 +1,204 @@
1
  ---
2
- license: cc-by-nc-sa-4.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ annotations_creators:
3
+ - no-annotation
4
+ - machine-generated
5
+ language:
6
+ - ca
7
+ language_creators:
8
+ - found
9
+ license:
10
+ - cc-by-nc-sa-4.0
11
+ multilinguality:
12
+ - monolingual
13
+ pretty_name: CATalog
14
+ size_categories:
15
+ - 10B<n<100B
16
+ source_datasets:
17
+ - extended|mc4
18
+ - extended|oscar
19
+ - extended|cawac
20
+ tags: []
21
+ task_categories:
22
+ - fill-mask
23
+ - text-generation
24
+ task_ids:
25
+ - masked-language-modeling
26
+ - slot-filling
27
+ - language-modeling
28
  ---
29
+
30
+ ## Dataset Description
31
+
32
+ - **Homepage:** [Projecte AINA](https://projecteaina.cat/)
33
+ - **Paper:** ["A CURATEd CATalog: Rethinking the Extraction of Pretraining Corpora for Mid-Resourced Languages"]()
34
+ - **Point of Contact:** [Language Technologies Unit at Barcelona Supercomputing Center (BSC)](langtech@bsc.es)
35
+
36
+ ### Dataset Summary
37
+
38
+ CATalog is a diverse, open-source Catalan corpus for language modelling. It consists of text documents from 26
39
+ different sources, including web crawling, news, forums, digital libraries and public institutions, totaling
40
+ in 17.45 billion words.
41
+
42
+ ### Supported Tasks and Leaderboards
43
+
44
+ - Fill-Mask
45
+ - Text Generation
46
+ - `other:Language-Modelling`: The dataset is suitable for training a model in Language Modelling, predicting the next
47
+ word in a given context. Success is measured by achieving a low perplexity score, indicating the model's proficiency
48
+ in accurately predicting subsequent words. [Perplexity](https://huggingface.co/spaces/evaluate-metric/perplexity)
49
+ - `other:Masked-Language-Modelling`: The dataset is designed for training models in Masked Language Modelling. This task
50
+ involves predicting masked or hidden words within a sentence. Success is typically measured by achieving a high
51
+ performance score, such as accuracy or F1 score, on correctly predicting the masked tokens.
52
+ [F1](https://huggingface.co/spaces/evaluate-metric/f1)
53
+
54
+ ### Languages
55
+
56
+ This dataset is in Catalan (ca-ES). Coming from the web, some documents may contain other languages.
57
+
58
+ ## Dataset Structure
59
+
60
+ ### Data Instances
61
+
62
+ The dataset is provided in a CSV format, where each row corresponds to a single document and contains a document
63
+ identifier, the text, a quality score, the strategy used to evaluate the document quality, languages, and a URL of the
64
+ document, if available.
65
+
66
+ ```
67
+ document text score strategy languages url
68
+
69
+ ```
70
+
71
+ ### Data Fields
72
+
73
+ - `document`: text string containing the document identifier. Consists of the subdataset code, the part number and a
74
+ document number.
75
+ - `text`: text string from the document, with paragraphs separated by two newlines escape sequences. It is meant to be
76
+ used directly as input for language modelling.
77
+ - `score`: integer representing the document quality, ranging from 0, which represents the worst quality, to 1, the
78
+ best quality.
79
+ - `strategy`: text string describing the type of evaluation applied to obtain the document score. generic_hard uses the
80
+ heuristic evaluation from CURATE and perfect_score means that manual review was done and the highest score (1) is applied.
81
+ - `languages`: dictionary containing the document languages, with a percentage indicating the character ratio for each one.
82
+ - `url`: text string with the URL of the document, if available.
83
+
84
+ ### Data Splits
85
+
86
+ We do not provide any canonical splits for CATalog.
87
+
88
+ ## Dataset Creation
89
+
90
+ ### Curation Rationale
91
+
92
+ CATalog is mainly built on filtered, non-overlapping versions of [CommonCrawl](https://commoncrawl.org/) snapshots and a smaller set of manually scored corpora from specific sources. We use the CURATE pipeline, which combines exact deduplication, language identification, and scoring heuristics.
93
+
94
+ In the design of CATalog, we adhere to the following values:
95
+
96
+ - (1) **Scale & Flexibility**. We intend to produce datasets that have a significant impact on the training of multilingual models in the range of 7B-180B parameters. Since Catalan is a medium-resource language and data acquisition is already a challenge, binary filtering will limit us in terms of the amount of data. By providing a score, we are able to easily filter the corpus according to our corpus according to our needs.
97
+ - (2) **Neutral scoring**. As opposed to ML-based filtering, we can use simple rules and heuristics to avoid introducing further bias into the model ([Dodge et al., 2021](https://arxiv.org/abs/2104.08758); [Welbl et al., 2021](https://arxiv.org/abs/2109.07445)). We only use [FastText](https://fasttext.cc/docs/en/language-identification.html) to reject documents in other languages.
98
+
99
+ During development, we performed comparative judgment experiments to evaluate the usefulness of the scoring from the CURATE pipeline, which appears in most documents in CATalog and is intended for further filtering and analysis. We found a moderate correlation between the score and the perceived quality of the text. Our main goal was to maximize the usability of the corpus without getting into a trade-off between quantity and quality.
100
+
101
+ ### Source Data
102
+
103
+ #### Initial Data Collection and Normalization
104
+
105
+ We applied extensive data processing using our CURATE pipeline.
106
+
107
+ We first filter documents by their language content using [FastText](https://fasttext.cc/docs/en/language-identification.html). Only documents with at least 50% of characters in Catalan are kept. We then perform exact document deduplication. After this stage, we score each document with a tested set of 8 heuristic evaluators, inspired from other web filterings and from our own creation.
108
+
109
+ The following pre-existing datasets were used:
110
+ - `[OSCAR-2301](https://huggingface.co/datasets/oscar-corpus/OSCAR-2301)`
111
+ - `[OSCAR-2201](https://huggingface.co/datasets/oscar-corpus/OSCAR-2201)`
112
+ - `[CaText](https://zenodo.org/records/5483031)`
113
+ - `[MaCoCu-ca 1.0](http://hdl.handle.net/11356/1837)`
114
+ - `[caWaC](https://huggingface.co/datasets/cawac)`
115
+ - `[Colossal OSCAR 1.0](https://huggingface.co/datasets/oscar-corpus/colossal-oscar-1.0)`
116
+ - `[mC4]({https://huggingface.co/datasets/mc4)`
117
+
118
+ #### Who are the source language producers?
119
+
120
+ Apart from the pre-existing datasets, all of them coming from [CommonCrawl](https://commoncrawl.org/) dumps, the following
121
+ sources provided their data on Open Data Agreements:
122
+ - Media Groups
123
+ - `[IB3](https://ib3.org/)`
124
+ - `[Grup El Món](https://grupmon.cat/)`
125
+ - `[Vilaweb](https://www.vilaweb.cat/)`
126
+ - `[Nació Digita](https://www.naciodigital.cat/)`
127
+ - `[ACN](https://www.acn.cat/)`
128
+ - `[Racó Català](https://www.racocatala.cat/)`
129
+ - `[Aquí Berguedà](https://www.aquibergueda.cat/)`
130
+ - Academic & Book Repositories
131
+ - `[Tesis Doctorals en Xarxa](https://www.tesisenred.net/)`
132
+ - `[Wikipedia](https://ca.wikipedia.org/)`
133
+ - `[Project Gutenberg](https://www.gutenberg.org/)`
134
+ - Government Institutions
135
+ - `[Valencian Parliament](https://www.cortsvalencianes.es/)`
136
+ - `[Diari Oficial de la Generalitat Valenciana](https://dogv.gva.es/)`
137
+ - `[Butlletí Oficial de la Universitat d'Alacant](https://www.boua.ua.es/)`
138
+
139
+ ### Annotations
140
+
141
+ [N/A]
142
+
143
+ #### Annotation process
144
+
145
+ [N/A]
146
+
147
+ #### Who are the annotators?
148
+
149
+ [N/A]
150
+
151
+ ### Personal and Sensitive Information
152
+
153
+ Being mostly constructed from Common Crawl, personal and sensitive information might be present.
154
+ This must be considered before training deep learning models with CATalog, specially in the case of text-generation models.
155
+
156
+ ## Considerations for Using the Data
157
+
158
+ ### Social Impact of Dataset
159
+
160
+ CATalog promotes the Catalan language in the NLP field, enabling development of advanced applications and chatbots
161
+ tailored to Catalan speakers, while improving access to information for better community understanding. However, most
162
+ of the sources in the dataset are web-scraped, which may bring in biases and privacy issues, risking biased outcomes and
163
+ potential misuse. Additionally, it might overlook the voices of low-resource communities, amplifying existing disparities
164
+ in representation.
165
+
166
+ Given that Catalan is a mid-resourced language with representation in digital sources, this dataset
167
+ becomes crucial for building inclusive NLP applications. It addresses the language's underrepresentation, empowering
168
+ communities with improved access to information in their native language. However, careful consideration of potential
169
+ biases and privacy issues is essential to ensure responsible and equitable technology use.
170
+
171
+ ### Discussion of Biases
172
+
173
+ Web-crawled content is over-represented with standard language varieties, impacting language model performance for
174
+ minority languages. Language diversity in data is crucial to avoid bias, especially in encoding non-standard dialects,
175
+ preventing the exclusion of demographic groups. Our corpus primarily focuses on Central Catalan, but we actively include
176
+ Valencian and Balearic Catalan, along with diverse sociolects from platforms like Racó Català Fòrums, aiming for a more
177
+ representative dataset. Despite legal uncertainties in web-scraped data, we prioritize permissive licenses and privacy
178
+ protection measures, acknowledging the challenges posed by personally identifiable information (PII) within large-scale
179
+ datasets. Our ongoing efforts aim to address privacy concerns and contribute to a more inclusive linguistic dataset.
180
+
181
+ ### Other Known Limitations
182
+
183
+ [N/A]
184
+
185
+ ## Additional Information
186
+
187
+ ### Dataset Curators
188
+
189
+ [Language Technologies Unit](langtech@bsc.es) at the Barcelona Supercomputing Center (BSC).
190
+
191
+ This work was funded by the [Departament de la Vicepresidència i de Polítiques Digitals i Territori de la Generalitat de Catalunya](https://politiquesdigitals.gencat.cat/ca/inici/) within the framework of [Projecte AINA](https://politiquesdigitals.gencat.cat/ca/economia/catalonia-ai/aina).
192
+
193
+ ### Licensing Information
194
+
195
+ [Creative Commons Attribution 4.0 International](https://creativecommons.org/licenses/by/4.0/)
196
+
197
+ ### Citation Information
198
+
199
+ [N/A]
200
+
201
+ ### Contributions
202
+
203
+ [N/A]
204
+