Upload 3 files
Browse files
config.json
CHANGED
|
@@ -1,13 +1,19 @@
|
|
| 1 |
{
|
| 2 |
"_sliding_window_pattern": 6,
|
| 3 |
"architectures": [
|
| 4 |
-
"
|
| 5 |
],
|
| 6 |
"attention_bias": false,
|
| 7 |
"attention_dropout": 0.0,
|
| 8 |
"attn_logit_softcapping": null,
|
|
|
|
|
|
|
|
|
|
|
|
|
| 9 |
"bos_token_id": 2,
|
| 10 |
"cache_implementation": "hybrid",
|
|
|
|
|
|
|
| 11 |
"eos_token_id": [
|
| 12 |
1,
|
| 13 |
106
|
|
@@ -52,7 +58,8 @@
|
|
| 52 |
"sliding_attention"
|
| 53 |
],
|
| 54 |
"max_position_embeddings": 32768,
|
| 55 |
-
"model_type": "
|
|
|
|
| 56 |
"num_attention_heads": 4,
|
| 57 |
"num_hidden_layers": 26,
|
| 58 |
"num_key_value_heads": 1,
|
|
|
|
| 1 |
{
|
| 2 |
"_sliding_window_pattern": 6,
|
| 3 |
"architectures": [
|
| 4 |
+
"FlashHeadGemma3ForCausalLM"
|
| 5 |
],
|
| 6 |
"attention_bias": false,
|
| 7 |
"attention_dropout": 0.0,
|
| 8 |
"attn_logit_softcapping": null,
|
| 9 |
+
"auto_map": {
|
| 10 |
+
"AutoConfig": "configuration_flash_head_gemma3_text.FlashHeadGemma3TextConfig",
|
| 11 |
+
"AutoModelForCausalLM": "modeling_flash_head_gemma3_text.FlashHeadGemma3ForCausalLM"
|
| 12 |
+
},
|
| 13 |
"bos_token_id": 2,
|
| 14 |
"cache_implementation": "hybrid",
|
| 15 |
+
"creation_time": 1753799722.0919366,
|
| 16 |
+
"enforce_equal_cluster_sizes": true,
|
| 17 |
"eos_token_id": [
|
| 18 |
1,
|
| 19 |
106
|
|
|
|
| 58 |
"sliding_attention"
|
| 59 |
],
|
| 60 |
"max_position_embeddings": 32768,
|
| 61 |
+
"model_type": "flash_head_gemma3_text",
|
| 62 |
+
"n_clusters": 16384,
|
| 63 |
"num_attention_heads": 4,
|
| 64 |
"num_hidden_layers": 26,
|
| 65 |
"num_key_value_heads": 1,
|
configuration_flash_head_gemma3_text.py
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
from embedl.models.gemma.modeling_flash_head import FlashHeadGemma3TextConfig
|
modeling_flash_head_gemma3_text.py
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
from embedl.models.gemma.modeling_flash_head import FlashHeadGemma3ForCausalLM
|