Update README.md
Browse files
README.md
CHANGED
|
@@ -54,7 +54,9 @@ This model is case-sensitive: it makes a difference between English and English.
|
|
| 54 |
The model has 6 layers, 768 dimension and 12 heads, totalizing 82M parameters (compared to 125M parameters for RoBERTa-base).
|
| 55 |
On average DistilRoBERTa is twice as fast as Roberta-base.
|
| 56 |
|
| 57 |
-
## Training
|
|
|
|
|
|
|
| 58 |
|
| 59 |
## Training procedure
|
| 60 |
|
|
|
|
| 54 |
The model has 6 layers, 768 dimension and 12 heads, totalizing 82M parameters (compared to 125M parameters for RoBERTa-base).
|
| 55 |
On average DistilRoBERTa is twice as fast as Roberta-base.
|
| 56 |
|
| 57 |
+
## Training Data
|
| 58 |
+
|
| 59 |
+
Polar sentiment dataset of sentences from financial news. The dataset consists of 4840 sentences from English language financial news categorised by sentiment. The dataset is divided by agreement rate of 5-8 annotators.
|
| 60 |
|
| 61 |
## Training procedure
|
| 62 |
|