Datasets:
Size:
10K<n<100K
License:
| pretty_name: "Multi-EuP v2: European Parliament Debates with MEP Metadata (24 languages)" | |
| dataset_name: multi-eup-v2 | |
| configs: | |
| - config_name: default | |
| data_files: "clean_all_with_did_qid.MEP.csv" | |
| license: cc-by-4.0 | |
| multilinguality: multilingual | |
| task_categories: | |
| - text-classification | |
| - text-retrieval | |
| - text-generation | |
| language: | |
| - bg | |
| - cs | |
| - da | |
| - de | |
| - el | |
| - en | |
| - es | |
| - et | |
| - fi | |
| - fr | |
| - ga | |
| - hr | |
| - hu | |
| - it | |
| - lt | |
| - lv | |
| - mt | |
| - nl | |
| - pl | |
| - pt | |
| - ro | |
| - sk | |
| - sl | |
| - sv | |
| size_categories: | |
| - 10K<n<100K | |
| homepage: "" | |
| repository: "" | |
| paper: "https://aclanthology.org/2024.mrl-1.23/" | |
| tags: | |
| - multilingual | |
| - european-parliament | |
| - political-discourse | |
| - metadata | |
| - mep | |
| # Multi-EuP-v2 | |
| This dataset card documents **Multi-EuP-v2**, a multilingual corpus of European Parliament debate speeches enriched with Member of European Parliament (MEP) metadata and multilingual debate titles/IDs. It supports research on political text analysis, speaker-attribute prediction, stance/vote prediction, multilingual NLP, and retrieval. | |
| ## Dataset Details | |
| ### Dataset Description | |
| **Multi-EuP-v2** aggregates **50,337** debate speeches (each a unique `did`) in **24 languages**. Each row contains the speech text (`TEXT`), speaker identity (`NAME`, `MEPID`), language (`LANGUAGE`), political group (`PARTY`), country and gender of the MEP, date, video timestamps, plus **multilingual debate titles `title_<LANG>`** and **per-language debate/vote linkage IDs `qid_<LANG>`**. | |
| - **Curated by:** Jinrui Yang, Fan Jiang, Timothy Baldwin | |
| - **Funded by:** Melbourne Research Scholarship; LIEF HPC-GPGPU Facility (LE170100200) | |
| - **Shared by:** University of Melbourne | |
| - **Language(s) (NLP):** `bg, cs, da, de, el, en, es, et, fi, fr, ga, hr, hu, it, lt, lv, mt, nl, pl, pt, ro, sk, sl, sv` (24 total) | |
| - **License:** cc-by-4.0 | |
| ### Dataset Sources | |
| - **Repository:** [https://github.com/jrnlp/MLIR_language_bias] | |
| - **Paper:** https://aclanthology.org/2024.mrl-1.23/ | |
| ## Uses | |
| ### Direct Use | |
| - **Text classification:** predict `gender`, `PARTY`, or `country` from `TEXT`. | |
| - **Stance/vote prediction:** link `qid_<LANG>` to external roll-call vote labels. | |
| - **Multilingual representation learning:** train/evaluate models across 24 EU languages. | |
| - **Information retrieval:** index `TEXT` and use `title_*`/`qid_*` as multilingual query anchors. | |
| ### Out-of-Scope Use | |
| - Inferring private attributes beyond public MEP metadata. | |
| - Automated profiling for sensitive decisions. | |
| - Misrepresenting model outputs as factual statements. | |
| ## Dataset Structure | |
| Each row corresponds to a single speech/document. | |
| **Core fields:** | |
| - `did` *(string)* β unique speech ID | |
| - `TEXT` *(string)* β speech text | |
| - `DATE` *(string/date)* β debate date | |
| - `LANGUAGE` *(string)* β language code | |
| - `NAME` *(string)* β MEP name | |
| - `MEPID` *(string)* β MEP ID | |
| - `PARTY` *(string)* β political group | |
| - `country` *(string)* β MEP's country | |
| - `gender` *(string)* β `Female`, `Male`, or `Unknown` | |
| - Additional provenance fields: `PRESIDENT`, `TEXTID`, `CODICT`, `VOD-START`, `VOD-END` | |
| **Multilingual metadata:** | |
| - `title_<LANG>` *(string)* β debate title in that language | |
| - `qid_<LANG>` *(string)* β debate/vote linkage ID in that language | |
| **Splits:** Single CSV, no predefined splits. | |
| **Basic stats:** | |
| - Rows: 50,337 | |
| - Languages: 24 | |
| - Top political groups: PPE 8,869; S-D 8,468; Renew 5,313; ECR 4,130; Verts/ALE 4,001; ID 3,286; The Left 2,951; NI 2,539; GUE/NGL 468 | |
| - Gender counts: Female 25,536; Male 23,461; Unknown 349 | |
| - Top countries: Germany 7,226; France 6,158; Poland 3,706; Spain 3,312; Italy 3,222; Netherlands 1,924; Greece 1,756; Romania 1,701; Czechia 1,661; Portugal 1,150; Belgium 1,134; Hungary 1,106 | |
| ## Dataset Creation | |
| ### Curation Rationale | |
| Support multilingual political text research, enabling standardized tasks in gender/group prediction, stance/vote prediction, and IR. | |
| ### Source Data | |
| #### Data Collection and Processing | |
| - **Source:** Official EP debates. | |
| - **Processing:** metadata linking, language verification, deduplication, multilingual title extraction. | |
| - **Quality checks:** consistency in language tags and IDs. | |
| #### Who are the source data producers? | |
| MEPs speaking in plenary debates; titles from official EP records. | |
| ### Annotations | |
| #### Annotation process | |
| Metadata compiled from public records; no manual stance labels. | |
| #### Personal and Sensitive Information | |
| Contains names and political opinions of public officials. | |
| ## Bias, Risks, and Limitations | |
| - Domain bias: formal political discourse. | |
| - Risk in demographic inference tasks. | |
| - Language/script differences affect comparability. | |
| ### Recommendations | |
| - Report per-language metrics. | |
| - Avoid over-claiming causal interpretations. | |
| ## Citation | |
| **BibTeX:** | |
| ```bibtex | |
| @inproceedings{yang-etal-2024-language-bias, | |
| title = {Language Bias in Multilingual Information Retrieval: The Nature of the Beast and Mitigation Methods}, | |
| author = {Yang, Jinrui and Jiang, Fan and Baldwin, Timothy}, | |
| booktitle = {Proceedings of the Fourth Workshop on Multilingual Representation Learning (MRL 2024)}, | |
| year = {2024}, | |
| pages = {280--292}, | |
| publisher = {Association for Computational Linguistics}, | |
| url = {https://aclanthology.org/2024.mrl-1.23/}, | |
| doi = {10.18653/v1/2024.mrl-1.23} | |
| } | |
| ``` | |
| **APA:** Yang, J., Jiang, F., & Baldwin, T. (2024). *Language Bias in Multilingual Information Retrieval: The Nature of the Beast and Mitigation Methods*. In MRL 2024. ACL. | |
| ## Contact | |
| [email protected] |