swaze commited on
Commit
9961a45
·
verified ·
1 Parent(s): 1aa132b

Upload 3 files

Browse files
config.json CHANGED
@@ -1,13 +1,19 @@
1
  {
2
  "_sliding_window_pattern": 6,
3
  "architectures": [
4
- "Gemma3ForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "attn_logit_softcapping": null,
 
 
 
 
9
  "bos_token_id": 2,
10
  "cache_implementation": "hybrid",
 
 
11
  "eos_token_id": [
12
  1,
13
  106
@@ -52,7 +58,8 @@
52
  "sliding_attention"
53
  ],
54
  "max_position_embeddings": 32768,
55
- "model_type": "gemma3_text",
 
56
  "num_attention_heads": 4,
57
  "num_hidden_layers": 26,
58
  "num_key_value_heads": 1,
 
1
  {
2
  "_sliding_window_pattern": 6,
3
  "architectures": [
4
+ "FlashHeadGemma3ForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "attn_logit_softcapping": null,
9
+ "auto_map": {
10
+ "AutoConfig": "configuration_flash_head_gemma3_text.FlashHeadGemma3TextConfig",
11
+ "AutoModelForCausalLM": "modeling_flash_head_gemma3_text.FlashHeadGemma3ForCausalLM"
12
+ },
13
  "bos_token_id": 2,
14
  "cache_implementation": "hybrid",
15
+ "creation_time": 1753799722.0919366,
16
+ "enforce_equal_cluster_sizes": true,
17
  "eos_token_id": [
18
  1,
19
  106
 
58
  "sliding_attention"
59
  ],
60
  "max_position_embeddings": 32768,
61
+ "model_type": "flash_head_gemma3_text",
62
+ "n_clusters": 16384,
63
  "num_attention_heads": 4,
64
  "num_hidden_layers": 26,
65
  "num_key_value_heads": 1,
configuration_flash_head_gemma3_text.py ADDED
@@ -0,0 +1 @@
 
 
1
+ from embedl.models.gemma.modeling_flash_head import FlashHeadGemma3TextConfig
modeling_flash_head_gemma3_text.py ADDED
@@ -0,0 +1 @@
 
 
1
+ from embedl.models.gemma.modeling_flash_head import FlashHeadGemma3ForCausalLM