Datasets:
File size: 15,120 Bytes
1d126da 2984711 49b2b8b f400c4f 49b2b8b f400c4f 49b2b8b f400c4f 49b2b8b |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 260 261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 |
---
dataset_info:
features:
- name: text
dtype: string
- name: embedding
sequence: float64
- name: labels
sequence: int64
- name: metadata
struct:
- name: country
dtype: string
- name: file_id
dtype: string
- name: language
dtype: string
- name: locality
dtype: string
- name: size
dtype: string
- name: type
dtype: string
- name: year
dtype: int64
splits:
- name: train
num_bytes: 124052504
num_examples: 5880
- name: test
num_bytes: 36948683
num_examples: 1470
download_size: 129951175
dataset_size: 161001187
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
license: cc-by-nc-sa-4.0
task_categories:
- text-classification
language:
- en
- es
- fr
tags:
- sustainable-development-goals
- sdgs
pretty_name: SDGi Corpus
size_categories:
- 1K<n<10K
---
# Dataset Card for SDGi Corpus
<!-- Provide a quick summary of the dataset. -->
SDGi Corpus is a curated dataset for text classification by the [United Nations Sustainable Development Goals (SDGs)](https://www.un.org/sustainabledevelopment/sustainable-development-goals/).
## Dataset Details
### Dataset Description
<!-- Provide a longer summary of what this dataset is. -->
SDG Integration Corpus (SDGi Corpus) is the most comprehensive multilingual collection of texts labelled by Sustainable
Development Goals (SDGs) to date. Designed for multi-label multilingual classification, SDGi Corpus contains over 7,000
examples in English, French and Spanish. Leveraging years of international SDG reporting on the national and subnational
levels, we hand-picked texts from Voluntary National Reviews (VNRs) and Voluntary Local Reviews (VLRs) from more than 180
countries to create an inclusive dataset that provides both focused and broad perspectives on the SDGs. The dataset comes
with a predefined train/test split.
- **Curated by:** United Nations Development Programme
- **Language(s):** English, French and Spanish
- **License:** CC BY-NC-SA 4.0
### Dataset Sources [optional]
<!-- Provide the basic links for the dataset. -->
- **Repository:** https://github.com/UNDP-Data/dsc-sdgi-corpus (benchmarks)
- **Paper:** https://ceur-ws.org/Vol-3764/paper3.pdf
- **Demo:** TBA.
## Uses
<!-- Address questions around how the dataset is intended to be used. -->
The dataset is designed primarily for text classification tasks β including binary, multiclass and multi-label classification β
in one or more of the three supported languages. The dataset includes rich metadata with provenance information and can be used for
other text mining tasks like topic modelling or quantitative text analysis with a focus on the 2030 Agenda for Sustainable Development.
### Direct Use
<!-- This section describes suitable use cases for the dataset. -->
The dataset can be directly used for training machine learning models for text classification tasks. It can also be used for topic modelling to
identify the main themes that occur in the corpus or a specific subset of it. The rich metadata provided makes it possible to conduct both a trageted or comparative
analyses along linguistic, geographic (country and/or locality) and temporal dimensions.
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the dataset will not work well for. -->
The dataset is not suitable for tasks that require information not included in the dataset, such as image analysis or audio processing.
It cannot be used for predicting future trends or patterns in the SDGs and is not linked to SDG indicator data directly.
## Dataset Structure
The dataset consists of `7350` examples, with `5880` in the training set and `1470` in the test set. Each example includes the following fields:
- `text`: `str` β the text of the example in the original language.
- `embedding`: `list[float]` β 1536-dimensional embedding from OpenAI's `text-embedding-ada-002` model.
- `labels`: `list[int]` β one or more integer labels corresponding to SDGs. About 89% of the examples have just one label.
- `metadata`: `dict` β a dictionary containing metadata information, including:
- `country`: `str` β ISO 3166-1 alpha-3 code.
- `file_id`: `str` β internal ID of the original file. Used for provenance and troubleshooting only.
- `language`: `str` β one of the three supported languages, i.e., `en` (English), `fr` (French), `es` (Spanish).
- `locality`: `str` β name of the locality within `country` for examples from VLRs, e.g., city, province or region name.
- `size`: `str` β the size group of the example in terms of tokens, i.e., `s` (small, approx. < 512 tokens), `m` (medium, approx. 512-2048 tokens), `l` (large, approx. > 2048 tokens).
- `type`: `str` β one of the two document types, i.e., `vnr` (Voluntary National Review) or `vlr` (Voluntary Local Review).
- `year`: `int` β year of the publication.
<aside class="note">
<b>Note:</b>
the embeddings were produced from texts after removing digits. Embedding raw `text` will not produce the same result.
After applying the following replacements, you should be able to obtain similar emebedding vectors:
</aside>
```python
re.sub(r'(\b\d+[\.\,]?\d*\b)', 'NUM', text)
```
The dataset comes with a predefined train/test split. The examples for the test set were not sampled at random. Instead, they were
sampled in a stratified fashion using weights proportional to the cross-entropy loss of a simple classifier fitted on the full dataset.
For details on the sampling process, refer to the paper.
## Dataset Creation
### Curation Rationale
<!-- Motivation for the creation of this dataset. -->
The dataset was created to facilitate automated analysis of large corpora with respect to the 2030 Agenda for Sustainable Development.
The dataset comprises texts from Voluntary National Reviews (VNRs) and Voluntary Local Reviews (VLRs) which are arguably the most
authoritative sources of SDG-related texts. The dataset is a collection of texts labelled by the source data producets, the curators
have not labelled any data themselves.
### Source Data
<!-- This section describes the source data (e.g. news text and headlines, social media posts, translated sentences, ...). -->
All examples were collected from one of the two sources:
- [Voluntary National Reviews (VNRs)](https://hlpf.un.org/vnrs)
- [Voluntary Local Reviews (VLRs)](https://sdgs.un.org/topics/voluntary-local-reviews)
Only Reviews in English, French and Spanish published between January 2016 and December 2023 were included.
#### Data Collection and Processing
<!-- This section describes the data collection and processing process such as data selection criteria, filtering and normalization methods, tools and libraries used, etc. -->
To create SDGi Corpus, we manually analysed each document, searching and extracting specific parts clearly linked to SDGs.
Our curation process can be summarised in 4 steps as follows:
1. Manually examine a given document to identify SDG-labelled content.
2. Extract pages containing relevant content to SDG-specific folders.
3. Edit extracted pages to redact (mask) irrelevant content before and after the relevant content.
4. For content linked to multiple SDGs, fill out a metadata sheet.
For details on the curation process, refer to the paper.
#### Who are the source data producers?
<!-- This section describes the people or systems who originally created the data. It should also include self-reported demographic or identity information for the source data creators if this information is available. -->
Voluntary National Reviews (VNRs) and Voluntary Local Reviews (VLRs) are typically produced by government agencies, national
statistical offices, and other relevant national and subnational institutions within each country. These entities are responsible
for collecting, analysing, and reporting on the progress of their respective countries towards the SDGs. In addition, international
organisations, civil society organisations, academia, and other stakeholders may also contribute to the data collection and reporting
process for VNRs and VLRs.
### Annotations
<!-- If the dataset contains annotations which are not part of the initial data collection, use this section to describe them. -->
The labels in the dataset come directly from the source documents. No label annotation has been performed to produce SDGi Corpus.
#### Annotation process
<!-- This section describes the annotation process such as annotation tools used in the process, the amount of data annotated, annotation guidelines provided to the annotators, interannotator statistics, annotation validation, etc. -->
Not applicable.
#### Who are the annotators?
<!-- This section describes the people or systems who created the annotations. -->
Not applicable.
#### Personal and Sensitive Information
<!-- State whether the dataset contains data that might be considered personal, sensitive, or private (e.g., data that reveals addresses, uniquely identifiable names or aliases, racial or ethnic origins, sexual orientations, religious beliefs, political opinions, financial or health data, etc.). If efforts were made to anonymize the data, describe the anonymization process. -->
While VNR and VLR texts are unlikely to contain any sensitive Personally Identifiable Information (PII) due to their public nature
and intented use, users should adhere to ethical standards and best practices when handling the dataset. Should sensitive PII
information be found in the dataset, you are strongly encouraged to notify the curators.
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
- **Language Bias**: The dataset includes texts in three languages, with English (71.9%) examples dominating the dataset, followed by examples in Spanish (15.9%) and French (12.2%). The performance of models trained on this dataset may be biased towards these languages and may not generalise well to texts in other languages. Multilingual classifiers should ensure consistent performance across the languages of interest.
- **Geographical Bias**: The dataset includes data from various countries. However, because VNRs and VLRs are self-reported documents, some countries have produced more reports than others and are therfore overrepresented while some others are underrepresented in the dataset. This could lead to geographical bias in the models trained on this dataset.
- **Temporal Limitations**: The dataset includes data from reports published between 2016 and 2023. Some earlier reports did not have the right structure to derive SDG labels and were not included in the dataset. As a text corpus, the dataset does not lend itself for predictive modelling to determine future trends or patterns in the SDGs.
- **Labelling Bias**: While the labels in the dataset come from the source documents directly, they may not be entirely bias-free. The biases of the authors of the source documents might be reflected in the content of the section or the labels they assigned to it.
- **Domain Bias**: VNRs and VLRs are formal public documents. Models trained on the data form these sources may not generalise well to other types of documents or contexts.
- **Sociotechnical Risks**: The use of this dataset for decision-making in policy or other areas related to the SDGs should be done with caution, considering all the potential biases and limitations of the dataset. Misinterpretation or misuse of the data could lead to unfair or ineffective decisions.
- **Corrupted texts**: A small fraction of texts in the dataset were not properly extracted from source PDFs and is corrupted. Affected examples will be removed from the dataset in the next version.
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users should be made aware of the risks, biases and limitations of the dataset.
Concerning the existence of corrupted texts, users are advised to remove them early on in the processing/training pipeline.
To identify such examples, one can look for a large share of non-alphanumeric or special characters as well as the number of
single character tokens.
## Citation
<!-- If there is a paper or blog post introducing the dataset, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
```
@inproceedings{skrynnyk2024sdgi,
author = {Mykola Skrynnyk and Gedion Disassa and Andrey Krachkov and Janine DeVera},
title = {SDGi Corpus: A Comprehensive Multilingual Dataset for Text Classification by Sustainable Development Goals},
booktitle = {Proceedings of the 2nd Symposium on NLP for Social Good},
year = {2024},
editor = {Procheta Sen and Tulika Saha and Danushka Bollegala},
volume = {3764},
series = {CEUR Workshop Proceedings},
pages = {32--42},
publisher = {CEUR-WS.org},
series = {CEUR Workshop Proceedings},
address = {Aachen},
venue = {Liverpool, United Kingdom},
issn = {1613-0073},
url = {https://ceur-ws.org/Vol-3764/paper3.pdf},
eventdate = {2024-04-25},
}
```
**APA:**
Skrynnyk, M., Disassa, G., Krachkov, A., & DeVera, J. (2024). SDGi Corpus: A Comprehensive Multilingual Dataset for Text Classification by Sustainable Development Goals. In P. Sen, T. Saha, & D. Bollegala (Eds.), Proceedings of the 2nd Symposium on NLP for Social Good (Vol. 3764, pp. 32β42). CEUR-WS.org. https://ceur-ws.org/Vol-3764/paper3.pdf
## Glossary
<!-- If relevant, include terms and calculations in this section that can help readers understand the dataset or dataset card. -->
- **SDGs (Sustainable Development Goals)** : A collection of 17 global goals set by the United Nations General Assembly in 2015 for the year 2030. They cover social and economic development issues including poverty, hunger, health, education, climate change, gender equality, water, sanitation, energy, urbanization, environment and social justice.
- **VLR (Voluntary Local Review)**: A process undertaken by local and regional governments to evaluate their progress towards the 2030 Agenda. Note that unlike VNRs, VLRs were not originally envisioned in the 2030 Agenda but emerged as a popular means of communication about SDG localisation.
- **VNR (Voluntary National Review)**: A process undertaken by national governments to evaluate their progress towards the 2030 Agenda.
## More Information
The dataset is a product of the DFx. [Data Futures Platform (DFx)](https://data.undp.org) is an open-source, central hub for data innovation for development impact.
Guided by UNDPβs thematic focus areas, we use a systems approach and advanced analytics to identify actions to
accelerate sustainable development around the world.
## Dataset Card Contact
For inquiries regarding data sources, technical assistance, or general information, please feel free to reach out to us at [email protected].
|