Trained on the Falcon 7b
Browse files- README.md +2 -7
- adapter_config.json +1 -1
- adapter_model.safetensors +2 -2
README.md
CHANGED
|
@@ -1,11 +1,6 @@
|
|
| 1 |
---
|
| 2 |
library_name: peft
|
| 3 |
-
|
| 4 |
-
base_model: bigscience/bloom-7b1
|
| 5 |
-
license: apache-2.0
|
| 6 |
-
tags:
|
| 7 |
-
- text-generation-inference
|
| 8 |
-
pipeline_tag: text-generation
|
| 9 |
---
|
| 10 |
|
| 11 |
# Model Card for Model ID
|
|
@@ -222,4 +217,4 @@ The following `bitsandbytes` quantization config was used during training:
|
|
| 222 |
### Framework versions
|
| 223 |
|
| 224 |
|
| 225 |
-
- PEFT 0.6.3.dev0
|
|
|
|
| 1 |
---
|
| 2 |
library_name: peft
|
| 3 |
+
base_model: tiiuae/falcon-7b
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
---
|
| 5 |
|
| 6 |
# Model Card for Model ID
|
|
|
|
| 217 |
### Framework versions
|
| 218 |
|
| 219 |
|
| 220 |
+
- PEFT 0.6.3.dev0
|
adapter_config.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
{
|
| 2 |
"alpha_pattern": {},
|
| 3 |
"auto_mapping": null,
|
| 4 |
-
"base_model_name_or_path": "
|
| 5 |
"bias": "none",
|
| 6 |
"fan_in_fan_out": false,
|
| 7 |
"inference_mode": true,
|
|
|
|
| 1 |
{
|
| 2 |
"alpha_pattern": {},
|
| 3 |
"auto_mapping": null,
|
| 4 |
+
"base_model_name_or_path": "tiiuae/falcon-7b",
|
| 5 |
"bias": "none",
|
| 6 |
"fan_in_fan_out": false,
|
| 7 |
"inference_mode": true,
|
adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f24ca81da2ed2a1804ef875a0a761fd691cc80bfc53ddcaeeb786d30aa636b14
|
| 3 |
+
size 18883912
|