Update README.md
Browse files
README.md
CHANGED
|
@@ -16,7 +16,8 @@ This repo contains the data for the first [UNFCCC Global Stocktake](https://unfc
|
|
| 16 |
The files in this dataset are as follows:
|
| 17 |
|
| 18 |
- `metadata.csv`: a CSV containing document metadata for each document we have collected. **This metadata may not be the same as what's stored in the source databases** – we have cleaned and added metadata where it's corrupted or missing.
|
| 19 |
-
- `full_text.
|
|
|
|
| 20 |
|
| 21 |
A research tool you can use to view this data and the results of some classifiers run on it is at [gst1.org](https://gst1.org).
|
| 22 |
|
|
@@ -78,7 +79,7 @@ metadata = pd.read_csv("metadata.csv")
|
|
| 78 |
|
| 79 |
### Loading text block data
|
| 80 |
|
| 81 |
-
**
|
| 82 |
|
| 83 |
``` py
|
| 84 |
# Reading from parquet
|
|
|
|
| 16 |
The files in this dataset are as follows:
|
| 17 |
|
| 18 |
- `metadata.csv`: a CSV containing document metadata for each document we have collected. **This metadata may not be the same as what's stored in the source databases** – we have cleaned and added metadata where it's corrupted or missing.
|
| 19 |
+
- `full_text.parquet`: a parquet file containing the full text of each document we have parsed. Each row is a text block (paragraph) with all the associated text block and document metadata.
|
| 20 |
+
- `full_text.jsonl`: a JSON file containing the same data as the parquet file. **It's recommended to use the parquet file as it stores the data 10x more efficiently.**
|
| 21 |
|
| 22 |
A research tool you can use to view this data and the results of some classifiers run on it is at [gst1.org](https://gst1.org).
|
| 23 |
|
|
|
|
| 79 |
|
| 80 |
### Loading text block data
|
| 81 |
|
| 82 |
+
As mentioned at the top of this README **the parquet file is recommended over JSON** as it stores the data much more (10x) more efficiently, meaning lower load on your system's memory requirements.
|
| 83 |
|
| 84 |
``` py
|
| 85 |
# Reading from parquet
|