Update README.md
Browse files
README.md
CHANGED
|
@@ -14,6 +14,23 @@ tags:
|
|
| 14 |
base_model: unsloth/tinyllama-bnb-4bit
|
| 15 |
---
|
| 16 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
# Uploaded model
|
| 18 |
|
| 19 |
- **Developed by:** Ramikan-BR
|
|
@@ -24,43 +41,4 @@ This llama model was trained 2x faster with [Unsloth](https://github.com/unsloth
|
|
| 24 |
|
| 25 |
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
| 26 |
|
| 27 |
-
---
|
| 28 |
-
==((====))== Unsloth - 2x faster free finetuning | Num GPUs = 1
|
| 29 |
-
\\ /| Num examples = 967 | Num Epochs = 1
|
| 30 |
-
O^O/ \_/ \ Batch size per device = 2 | Gradient Accumulation steps = 16
|
| 31 |
-
\ / Total batch size = 32 | Total steps = 30
|
| 32 |
-
"-____-" Number of trainable parameters = 100,925,440
|
| 33 |
-
[30/30 26:26, Epoch 0/1]
|
| 34 |
-
Step Training Loss
|
| 35 |
|
| 36 |
-
1 1.737000
|
| 37 |
-
2 1.738000
|
| 38 |
-
3 1.384700
|
| 39 |
-
4 1.086400
|
| 40 |
-
5 1.009600
|
| 41 |
-
6 0.921000
|
| 42 |
-
7 0.830400
|
| 43 |
-
8 0.808900
|
| 44 |
-
9 0.774500
|
| 45 |
-
10 0.759900
|
| 46 |
-
11 0.736100
|
| 47 |
-
12 0.721200
|
| 48 |
-
13 0.733200
|
| 49 |
-
14 0.701000
|
| 50 |
-
15 0.711700
|
| 51 |
-
16 0.701400
|
| 52 |
-
17 0.689500
|
| 53 |
-
18 0.678800
|
| 54 |
-
19 0.675200
|
| 55 |
-
20 0.680500
|
| 56 |
-
21 0.685800
|
| 57 |
-
22 0.681200
|
| 58 |
-
23 0.672000
|
| 59 |
-
24 0.679900
|
| 60 |
-
25 0.675500
|
| 61 |
-
26 0.666600
|
| 62 |
-
27 0.687900
|
| 63 |
-
28 0.653600
|
| 64 |
-
29 0.672500
|
| 65 |
-
30 0.660900
|
| 66 |
-
---
|
|
|
|
| 14 |
base_model: unsloth/tinyllama-bnb-4bit
|
| 15 |
---
|
| 16 |
|
| 17 |
+
--- >-
|
| 18 |
+
==((====))== Unsloth - 2x faster free finetuning | Num GPUs = 1 \\ /| Num
|
| 19 |
+
examples = 967 | Num Epochs = 1 O^O/ \_/ \ Batch size per device = 2 | Gradient
|
| 20 |
+
Accumulation steps = 16 \ / Total batch size = 32 | Total steps = 30 "-____-"
|
| 21 |
+
Number of trainable parameters = 100,925,440 [30/30 26:26, Epoch 0/1] Step
|
| 22 |
+
Training Loss
|
| 23 |
+
|
| 24 |
+
1 1.737000 2 1.738000 3 1.384700 4 1.086400 5 1.009600 6 0.921000 7 0.830400 8
|
| 25 |
+
0.808900 9 0.774500 10 0.759900 11 0.736100 12 0.721200 13 0.733200 14 0.701000
|
| 26 |
+
15 0.711700 16 0.701400 17 0.689500 18 0.678800 19 0.675200 20 0.680500 21
|
| 27 |
+
0.685800 22 0.681200 23 0.672000 24 0.679900 25 0.675500 26 0.666600 27 0.687900
|
| 28 |
+
28 0.653600 29 0.672500 30 0.660900
|
| 29 |
+
|
| 30 |
+
---
|
| 31 |
+
null
|
| 32 |
+
|
| 33 |
+
|
| 34 |
# Uploaded model
|
| 35 |
|
| 36 |
- **Developed by:** Ramikan-BR
|
|
|
|
| 41 |
|
| 42 |
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
| 43 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|