Initial model upload
Browse files- config.json +1 -1
- generation_config.json +1 -1
- model-00001-of-00006.safetensors +1 -1
- model-00002-of-00006.safetensors +1 -1
- model-00003-of-00006.safetensors +1 -1
- model-00004-of-00006.safetensors +1 -1
- model-00005-of-00006.safetensors +1 -1
- model-00006-of-00006.safetensors +1 -1
- tokenizer_config.json +1 -2
config.json
CHANGED
|
@@ -21,7 +21,7 @@
|
|
| 21 |
"rope_theta": 500000,
|
| 22 |
"tie_word_embeddings": false,
|
| 23 |
"torch_dtype": "float32",
|
| 24 |
-
"transformers_version": "4.
|
| 25 |
"use_cache": true,
|
| 26 |
"vocab_size": 100352
|
| 27 |
}
|
|
|
|
| 21 |
"rope_theta": 500000,
|
| 22 |
"tie_word_embeddings": false,
|
| 23 |
"torch_dtype": "float32",
|
| 24 |
+
"transformers_version": "4.47.1",
|
| 25 |
"use_cache": true,
|
| 26 |
"vocab_size": 100352
|
| 27 |
}
|
generation_config.json
CHANGED
|
@@ -3,5 +3,5 @@
|
|
| 3 |
"bos_token_id": 100257,
|
| 4 |
"eos_token_id": 100257,
|
| 5 |
"pad_token_id": 100277,
|
| 6 |
-
"transformers_version": "4.
|
| 7 |
}
|
|
|
|
| 3 |
"bos_token_id": 100257,
|
| 4 |
"eos_token_id": 100257,
|
| 5 |
"pad_token_id": 100277,
|
| 6 |
+
"transformers_version": "4.47.1"
|
| 7 |
}
|
model-00001-of-00006.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4949546184
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:39f32672d6f3ca5c1c46dd23ba465bcfaba8f44c2d03ba90d49a480e40a55f4f
|
| 3 |
size 4949546184
|
model-00002-of-00006.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4991622816
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:aabd965c225fb67963876a4b004d127a0e0d1931d5f0afebc564cdd2076fc52f
|
| 3 |
size 4991622816
|
model-00003-of-00006.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4924546896
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a2bc26cadf9d85e858cec68d8ea845a674e83c0d1d0f7a5b6a230fcad8315860
|
| 3 |
size 4924546896
|
model-00004-of-00006.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4857404928
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8858570eb05bcd69d93dc313b045f6eaff89a6e0c1f150cd6fc2ed40e54eb22e
|
| 3 |
size 4857404928
|
model-00005-of-00006.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4857404928
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f85c34dc2bf7b6c4883c824b5a79f6dc554d0d271d24d2e40684a0109a91229a
|
| 3 |
size 4857404928
|
model-00006-of-00006.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4613984792
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:430df3110cfd3d5668bd97f99b80cecefab23e34860e1c6d000b481cb00a31a8
|
| 3 |
size 4613984792
|
tokenizer_config.json
CHANGED
|
@@ -179,11 +179,10 @@
|
|
| 179 |
}
|
| 180 |
},
|
| 181 |
"bos_token": "<|endoftext|>",
|
| 182 |
-
"chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
| 183 |
"clean_up_tokenization_spaces": false,
|
| 184 |
"eos_token": "<|endoftext|>",
|
| 185 |
"extra_special_tokens": {},
|
| 186 |
-
"model_max_length":
|
| 187 |
"pad_token": "<|pad|>",
|
| 188 |
"tokenizer_class": "GPT2Tokenizer",
|
| 189 |
"unk_token": "<|endoftext|>"
|
|
|
|
| 179 |
}
|
| 180 |
},
|
| 181 |
"bos_token": "<|endoftext|>",
|
|
|
|
| 182 |
"clean_up_tokenization_spaces": false,
|
| 183 |
"eos_token": "<|endoftext|>",
|
| 184 |
"extra_special_tokens": {},
|
| 185 |
+
"model_max_length": 1000000000000000019884624838656,
|
| 186 |
"pad_token": "<|pad|>",
|
| 187 |
"tokenizer_class": "GPT2Tokenizer",
|
| 188 |
"unk_token": "<|endoftext|>"
|