aivfkesavan commited on
Commit
fb88ec2
·
1 Parent(s): 8900a38

Enable largefiles + add MLX 4-bit weights

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,16 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - moe
4
+ - minimax
5
+ - bfloat16
6
+ - sglang
7
+ - gguf
8
+ - mlx
9
+ license: mit
10
+ datasets:
11
+ - nick007x/github-code-2025
12
+ - tatsu-lab/alpaca
13
+ base_model: VibeStudio/MiniMax-M2-THRIFT-55
14
+ pipeline_tag: text-generation
15
+ library_name: mlx
16
+ ---
added_tokens.json ADDED
@@ -0,0 +1,56 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</minimax:tool_call>": 200053,
3
+ "</think>": 200051,
4
+ "<add_file>": 200036,
5
+ "<code_context>": 200043,
6
+ "<code_interpreter>": 200023,
7
+ "<commit_after>": 200018,
8
+ "<commit_before>": 200016,
9
+ "<commit_message>": 200040,
10
+ "<commit_msg>": 200017,
11
+ "<delete_file>": 200037,
12
+ "<edit_file>": 200039,
13
+ "<empty_output>": 200015,
14
+ "<empty_source_file>": 200041,
15
+ "<file_content>": 200044,
16
+ "<file_sep>": 200049,
17
+ "<filename>": 200006,
18
+ "<filepath>": 200048,
19
+ "<fim_middle>": 200002,
20
+ "<fim_pad>": 200004,
21
+ "<fim_prefix>": 200001,
22
+ "<fim_suffix>": 200003,
23
+ "<function_call>": 200022,
24
+ "<gh_stars>": 200007,
25
+ "<issue_closed>": 200010,
26
+ "<issue_comment>": 200009,
27
+ "<issue_start>": 200008,
28
+ "<jupyter_code>": 200013,
29
+ "<jupyter_error>": 200035,
30
+ "<jupyter_output>": 200014,
31
+ "<jupyter_start>": 200011,
32
+ "<jupyter_text>": 200012,
33
+ "<minimax:tool_call>": 200052,
34
+ "<pr_start>": 200046,
35
+ "<rename_file>": 200038,
36
+ "<repo_struct>": 200042,
37
+ "<reponame>": 200005,
38
+ "<review_comment>": 200047,
39
+ "<source_files>": 200045,
40
+ "<think>": 200050,
41
+ "[e~[": 200020,
42
+ "]!d~[": 200021,
43
+ "]!p~[": 200000,
44
+ "]<]end of image[>[": 200030,
45
+ "]<]end of speech[>[": 200028,
46
+ "]<]end of video[>[": 200032,
47
+ "]<]image[>[": 200025,
48
+ "]<]speech[>[": 200024,
49
+ "]<]start of image[>[": 200029,
50
+ "]<]start of speech[>[": 200027,
51
+ "]<]start of video[>[": 200031,
52
+ "]<]video[>[": 200026,
53
+ "]<]vision pad[>[": 200033,
54
+ "]~!b[": 200034,
55
+ "]~b]": 200019
56
+ }
chat_template.jinja ADDED
@@ -0,0 +1,159 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {# ----------‑‑‑ special token variables ‑‑‑---------- #}
2
+ {%- set toolcall_begin_token = '<minimax:tool_call>' -%}
3
+ {%- set toolcall_end_token = '</minimax:tool_call>' -%}
4
+ {#- Tool Rendering Functions ============================================== -#}
5
+ {%- macro render_tool_namespace(namespace_name, tool_list) -%}
6
+ {%- for tool in tool_list -%}
7
+ <tool>{{ tool.function | tojson(ensure_ascii=False) }}</tool>
8
+ {% endfor -%}
9
+ {%- endmacro -%}
10
+ {%- macro visible_text(content) -%}
11
+ {%- if content is string -%}
12
+ {{ content }}
13
+ {%- elif content is iterable and content is not mapping -%}
14
+ {%- for item in content -%}
15
+ {%- if item is mapping and item.type == 'text' -%}
16
+ {{- item.text }}
17
+ {%- elif item is string -%}
18
+ {{- item }}
19
+ {%- endif -%}
20
+ {%- endfor -%}
21
+ {%- else -%}
22
+ {{- content }}
23
+ {%- endif -%}
24
+ {%- endmacro -%}
25
+ {#- System Message Construction ============================================ -#}
26
+ {%- macro build_system_message(system_message) -%}
27
+ {%- if system_message and system_message.content -%}
28
+ {{- visible_text(system_message.content) }}
29
+ {%- else -%}
30
+ {%- if model_identity is not defined -%}
31
+ {%- set model_identity = "You are a helpful assistant." -%}
32
+ {%- endif -%}
33
+ {{- model_identity }}
34
+ {%- endif -%}
35
+
36
+ {#- Handle current_date -#}
37
+ {%- if system_message and system_message.current_date -%}
38
+ {{- '\n' ~ 'Current date: ' + system_message.current_date }}
39
+ {%- endif -%}
40
+ {#- Handle current_location -#}
41
+ {%- if system_message and system_message.current_location -%}
42
+ {{- '\n' ~ 'Current location: ' + system_message.current_location }}
43
+ {%- endif -%}
44
+ {%- endmacro -%}
45
+ {#- Main Template Logic ================================================= -#}
46
+ {#- Extract system message (only first message if it's system) -#}
47
+ {%- set system_message = none -%}
48
+ {%- set conversation_messages = messages -%}
49
+ {%- if messages and messages[0].role == "system" -%}
50
+ {%- set system_message = messages[0] -%}
51
+ {%- set conversation_messages = messages[1:] -%}
52
+ {%- endif -%}
53
+ {#- Get the last user message turn, for interleved thinking -#}
54
+ {%- set ns = namespace(last_user_index=-1) %}
55
+ {% for m in conversation_messages %}
56
+ {%- if m.role == 'user' %}
57
+ {% set ns.last_user_index = loop.index0 -%}
58
+ {%- endif %}
59
+ {%- endfor %}
60
+ {#- Render system message -#}
61
+ {{- ']~!b[' ~ ']~b]system' ~ '\n' }}
62
+ {{- build_system_message(system_message) }}
63
+ {#- Render tools if available -#}
64
+ {%- if tools -%}
65
+ {{- '\n\n' ~ '# Tools' ~ '\n' ~ 'You may call one or more tools to assist with the user query.\nHere are the tools available in JSONSchema format:' ~ '\n' }}
66
+ {{- '\n' ~ '<tools>' ~ '\n' }}
67
+ {{- render_tool_namespace("functions", tools) }}
68
+ {{- '</tools>' ~ '\n\n' }}
69
+ {{- 'When making tool calls, use XML format to invoke tools and pass parameters:' ~ '\n' }}
70
+ {{- '\n' ~ toolcall_begin_token }}
71
+ <invoke name="tool-name-1">
72
+ <parameter name="param-key-1">param-value-1</parameter>
73
+ <parameter name="param-key-2">param-value-2</parameter>
74
+ ...
75
+ </invoke>
76
+ {{- '\n' ~ toolcall_end_token }}
77
+ {%- endif -%}
78
+ {{- '[e~[\n' }}
79
+
80
+ {#- Render messages -#}
81
+ {%- set last_tool_call = namespace(name=none) -%}
82
+ {%- for message in conversation_messages -%}
83
+ {%- if message.role == 'assistant' -%}
84
+ {#- Only render reasoning_content if no user message follows -#}
85
+ {{- ']~b]ai' ~ '\n' }}
86
+
87
+ {%- set reasoning_content = '' %}
88
+ {%- set content = visible_text(message.content) %}
89
+ {%- if message.reasoning_content is string %}
90
+ {%- set reasoning_content = message.reasoning_content %}
91
+ {%- else %}
92
+ {%- if '</think>' in content %}
93
+ {%- set reasoning_content = content.split('</think>')[0].strip('\n').split('<think>')[-1].strip('\n') %}
94
+ {%- set content = content.split('</think>')[-1].strip('\n') %}
95
+ {%- endif %}
96
+ {%- endif %}
97
+ {%- if reasoning_content and loop.index0 > ns.last_user_index -%}
98
+ {{- '<think>' ~ '\n' ~ reasoning_content ~ '\n' ~ '</think>' ~ '\n\n' }}
99
+ {%- endif -%}
100
+ {%- if content -%}
101
+ {{- content }}
102
+ {%- endif -%}
103
+ {%- if message.tool_calls -%}
104
+ {{- '\n' ~ toolcall_begin_token ~ '\n' }}
105
+
106
+ {%- for tool_call in message.tool_calls -%}
107
+ {%- if tool_call.function %}
108
+ {%- set tool_call = tool_call.function %}
109
+ {%- endif %}
110
+ {{- '<invoke name="' + tool_call.name + '">' }}
111
+ {% set _args = tool_call.arguments %}
112
+ {%- for k, v in _args.items() %}
113
+ {{- '<parameter name="' + k + '">' }}
114
+ {{- v | tojson(ensure_ascii=False) if v is not string else v }}
115
+ {{- '</parameter>' }}
116
+ {% endfor %}
117
+ {{- '</invoke>' ~ '\n' }}
118
+ {%- endfor -%}
119
+
120
+ {{- toolcall_end_token}}
121
+ {%- set last_tool_call.name = message.tool_calls[-1].name -%}
122
+ {%- else -%}
123
+ {%- set last_tool_call.name = none -%}
124
+ {%- endif -%}
125
+ {{- '[e~[' ~ '\n' }}
126
+
127
+ {%- elif message.role == 'tool' -%}
128
+ {%- if last_tool_call.name is none -%}
129
+ {{- raise_exception("Message has tool role, but there was no previous assistant message with a tool call!") }}
130
+ {%- endif -%}
131
+ {%- if loop.first or (conversation_messages[loop.index0 - 1].role != 'tool') -%}
132
+ {{- ']~b]tool' }}
133
+ {%- endif -%}
134
+ {%- if message.content is string -%}
135
+ {{- '\n<response>' }}
136
+ {{- message.content }}
137
+ {{- '</response>' }}
138
+ {%- else -%}
139
+ {%- for tr in message.content -%}
140
+ {{- '\n<response>' }}
141
+ {{- tr.output if tr.output is defined else (tr.text if tr.type == 'text' and tr.text is defined else tr) }}
142
+ {{- '\n</response>' }}
143
+ {%- endfor -%}
144
+ {%- endif -%}
145
+ {%- if loop.last or (conversation_messages[loop.index0 + 1].role != 'tool') -%}
146
+ {{- '[e~[\n' -}}
147
+ {%- endif -%}
148
+
149
+ {%- elif message.role == 'user' -%}
150
+ {{- ']~b]user' ~ '\n' }}
151
+ {{- visible_text(message.content) }}
152
+ {{- '[e~[' ~ '\n' }}
153
+ {%- endif -%}
154
+ {%- endfor -%}
155
+
156
+ {#- Generation prompt -#}
157
+ {%- if add_generation_prompt -%}
158
+ {{- ']~b]ai' ~ '\n' ~ '<think>' ~ '\n' }}
159
+ {%- endif -%}
config.json ADDED
@@ -0,0 +1,626 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MiniMaxM2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "attn_type_list": [
7
+ 1,
8
+ 1,
9
+ 1,
10
+ 1,
11
+ 1,
12
+ 1,
13
+ 1,
14
+ 1,
15
+ 1,
16
+ 1,
17
+ 1,
18
+ 1,
19
+ 1,
20
+ 1,
21
+ 1,
22
+ 1,
23
+ 1,
24
+ 1,
25
+ 1,
26
+ 1,
27
+ 1,
28
+ 1,
29
+ 1,
30
+ 1,
31
+ 1,
32
+ 1,
33
+ 1,
34
+ 1,
35
+ 1,
36
+ 1,
37
+ 1,
38
+ 1,
39
+ 1,
40
+ 1,
41
+ 1,
42
+ 1,
43
+ 1,
44
+ 1,
45
+ 1,
46
+ 1,
47
+ 1,
48
+ 1,
49
+ 1,
50
+ 1,
51
+ 1,
52
+ 1,
53
+ 1,
54
+ 1,
55
+ 1,
56
+ 1,
57
+ 1,
58
+ 1,
59
+ 1,
60
+ 1,
61
+ 1,
62
+ 1,
63
+ 1,
64
+ 1,
65
+ 1,
66
+ 1,
67
+ 1,
68
+ 1
69
+ ],
70
+ "attn_window_size": null,
71
+ "auto_map": {
72
+ "AutoConfig": "configuration_minimax_m2.MiniMaxM2Config",
73
+ "AutoModelForCausalLM": "modeling_minimax_m2.MiniMaxM2ForCausalLM"
74
+ },
75
+ "head_dim": 128,
76
+ "hidden_act": "silu",
77
+ "hidden_size": 3072,
78
+ "initializer_range": 0.02,
79
+ "intermediate_size": 1536,
80
+ "layernorm_full_attention_beta": 1.0,
81
+ "layernorm_linear_attention_beta": 1.0,
82
+ "layernorm_mlp_beta": 1.0,
83
+ "max_model_len": null,
84
+ "max_position_embeddings": 196608,
85
+ "mlp_intermediate_size": 8192,
86
+ "model_type": "minimax",
87
+ "mtp_transformer_layers": 1,
88
+ "num_attention_heads": 48,
89
+ "num_expert_group": null,
90
+ "num_experts_per_tok": 8,
91
+ "num_hidden_layers": 62,
92
+ "num_key_value_heads": 8,
93
+ "num_local_experts": 116,
94
+ "num_mtp_modules": 3,
95
+ "output_router_logits": false,
96
+ "partial_rotary_factor": 0.5,
97
+ "qk_norm_type": "per_layer",
98
+ "quantization": {
99
+ "group_size": 32,
100
+ "bits": 4,
101
+ "mode": "mxfp4",
102
+ "model.layers.0.block_sparse_moe.gate": {
103
+ "group_size": 64,
104
+ "bits": 8
105
+ },
106
+ "model.layers.1.block_sparse_moe.gate": {
107
+ "group_size": 64,
108
+ "bits": 8
109
+ },
110
+ "model.layers.2.block_sparse_moe.gate": {
111
+ "group_size": 64,
112
+ "bits": 8
113
+ },
114
+ "model.layers.3.block_sparse_moe.gate": {
115
+ "group_size": 64,
116
+ "bits": 8
117
+ },
118
+ "model.layers.4.block_sparse_moe.gate": {
119
+ "group_size": 64,
120
+ "bits": 8
121
+ },
122
+ "model.layers.5.block_sparse_moe.gate": {
123
+ "group_size": 64,
124
+ "bits": 8
125
+ },
126
+ "model.layers.6.block_sparse_moe.gate": {
127
+ "group_size": 64,
128
+ "bits": 8
129
+ },
130
+ "model.layers.7.block_sparse_moe.gate": {
131
+ "group_size": 64,
132
+ "bits": 8
133
+ },
134
+ "model.layers.8.block_sparse_moe.gate": {
135
+ "group_size": 64,
136
+ "bits": 8
137
+ },
138
+ "model.layers.9.block_sparse_moe.gate": {
139
+ "group_size": 64,
140
+ "bits": 8
141
+ },
142
+ "model.layers.10.block_sparse_moe.gate": {
143
+ "group_size": 64,
144
+ "bits": 8
145
+ },
146
+ "model.layers.11.block_sparse_moe.gate": {
147
+ "group_size": 64,
148
+ "bits": 8
149
+ },
150
+ "model.layers.12.block_sparse_moe.gate": {
151
+ "group_size": 64,
152
+ "bits": 8
153
+ },
154
+ "model.layers.13.block_sparse_moe.gate": {
155
+ "group_size": 64,
156
+ "bits": 8
157
+ },
158
+ "model.layers.14.block_sparse_moe.gate": {
159
+ "group_size": 64,
160
+ "bits": 8
161
+ },
162
+ "model.layers.15.block_sparse_moe.gate": {
163
+ "group_size": 64,
164
+ "bits": 8
165
+ },
166
+ "model.layers.16.block_sparse_moe.gate": {
167
+ "group_size": 64,
168
+ "bits": 8
169
+ },
170
+ "model.layers.17.block_sparse_moe.gate": {
171
+ "group_size": 64,
172
+ "bits": 8
173
+ },
174
+ "model.layers.18.block_sparse_moe.gate": {
175
+ "group_size": 64,
176
+ "bits": 8
177
+ },
178
+ "model.layers.19.block_sparse_moe.gate": {
179
+ "group_size": 64,
180
+ "bits": 8
181
+ },
182
+ "model.layers.20.block_sparse_moe.gate": {
183
+ "group_size": 64,
184
+ "bits": 8
185
+ },
186
+ "model.layers.21.block_sparse_moe.gate": {
187
+ "group_size": 64,
188
+ "bits": 8
189
+ },
190
+ "model.layers.22.block_sparse_moe.gate": {
191
+ "group_size": 64,
192
+ "bits": 8
193
+ },
194
+ "model.layers.23.block_sparse_moe.gate": {
195
+ "group_size": 64,
196
+ "bits": 8
197
+ },
198
+ "model.layers.24.block_sparse_moe.gate": {
199
+ "group_size": 64,
200
+ "bits": 8
201
+ },
202
+ "model.layers.25.block_sparse_moe.gate": {
203
+ "group_size": 64,
204
+ "bits": 8
205
+ },
206
+ "model.layers.26.block_sparse_moe.gate": {
207
+ "group_size": 64,
208
+ "bits": 8
209
+ },
210
+ "model.layers.27.block_sparse_moe.gate": {
211
+ "group_size": 64,
212
+ "bits": 8
213
+ },
214
+ "model.layers.28.block_sparse_moe.gate": {
215
+ "group_size": 64,
216
+ "bits": 8
217
+ },
218
+ "model.layers.29.block_sparse_moe.gate": {
219
+ "group_size": 64,
220
+ "bits": 8
221
+ },
222
+ "model.layers.30.block_sparse_moe.gate": {
223
+ "group_size": 64,
224
+ "bits": 8
225
+ },
226
+ "model.layers.31.block_sparse_moe.gate": {
227
+ "group_size": 64,
228
+ "bits": 8
229
+ },
230
+ "model.layers.32.block_sparse_moe.gate": {
231
+ "group_size": 64,
232
+ "bits": 8
233
+ },
234
+ "model.layers.33.block_sparse_moe.gate": {
235
+ "group_size": 64,
236
+ "bits": 8
237
+ },
238
+ "model.layers.34.block_sparse_moe.gate": {
239
+ "group_size": 64,
240
+ "bits": 8
241
+ },
242
+ "model.layers.35.block_sparse_moe.gate": {
243
+ "group_size": 64,
244
+ "bits": 8
245
+ },
246
+ "model.layers.36.block_sparse_moe.gate": {
247
+ "group_size": 64,
248
+ "bits": 8
249
+ },
250
+ "model.layers.37.block_sparse_moe.gate": {
251
+ "group_size": 64,
252
+ "bits": 8
253
+ },
254
+ "model.layers.38.block_sparse_moe.gate": {
255
+ "group_size": 64,
256
+ "bits": 8
257
+ },
258
+ "model.layers.39.block_sparse_moe.gate": {
259
+ "group_size": 64,
260
+ "bits": 8
261
+ },
262
+ "model.layers.40.block_sparse_moe.gate": {
263
+ "group_size": 64,
264
+ "bits": 8
265
+ },
266
+ "model.layers.41.block_sparse_moe.gate": {
267
+ "group_size": 64,
268
+ "bits": 8
269
+ },
270
+ "model.layers.42.block_sparse_moe.gate": {
271
+ "group_size": 64,
272
+ "bits": 8
273
+ },
274
+ "model.layers.43.block_sparse_moe.gate": {
275
+ "group_size": 64,
276
+ "bits": 8
277
+ },
278
+ "model.layers.44.block_sparse_moe.gate": {
279
+ "group_size": 64,
280
+ "bits": 8
281
+ },
282
+ "model.layers.45.block_sparse_moe.gate": {
283
+ "group_size": 64,
284
+ "bits": 8
285
+ },
286
+ "model.layers.46.block_sparse_moe.gate": {
287
+ "group_size": 64,
288
+ "bits": 8
289
+ },
290
+ "model.layers.47.block_sparse_moe.gate": {
291
+ "group_size": 64,
292
+ "bits": 8
293
+ },
294
+ "model.layers.48.block_sparse_moe.gate": {
295
+ "group_size": 64,
296
+ "bits": 8
297
+ },
298
+ "model.layers.49.block_sparse_moe.gate": {
299
+ "group_size": 64,
300
+ "bits": 8
301
+ },
302
+ "model.layers.50.block_sparse_moe.gate": {
303
+ "group_size": 64,
304
+ "bits": 8
305
+ },
306
+ "model.layers.51.block_sparse_moe.gate": {
307
+ "group_size": 64,
308
+ "bits": 8
309
+ },
310
+ "model.layers.52.block_sparse_moe.gate": {
311
+ "group_size": 64,
312
+ "bits": 8
313
+ },
314
+ "model.layers.53.block_sparse_moe.gate": {
315
+ "group_size": 64,
316
+ "bits": 8
317
+ },
318
+ "model.layers.54.block_sparse_moe.gate": {
319
+ "group_size": 64,
320
+ "bits": 8
321
+ },
322
+ "model.layers.55.block_sparse_moe.gate": {
323
+ "group_size": 64,
324
+ "bits": 8
325
+ },
326
+ "model.layers.56.block_sparse_moe.gate": {
327
+ "group_size": 64,
328
+ "bits": 8
329
+ },
330
+ "model.layers.57.block_sparse_moe.gate": {
331
+ "group_size": 64,
332
+ "bits": 8
333
+ },
334
+ "model.layers.58.block_sparse_moe.gate": {
335
+ "group_size": 64,
336
+ "bits": 8
337
+ },
338
+ "model.layers.59.block_sparse_moe.gate": {
339
+ "group_size": 64,
340
+ "bits": 8
341
+ },
342
+ "model.layers.60.block_sparse_moe.gate": {
343
+ "group_size": 64,
344
+ "bits": 8
345
+ },
346
+ "model.layers.61.block_sparse_moe.gate": {
347
+ "group_size": 64,
348
+ "bits": 8
349
+ }
350
+ },
351
+ "quantization_config": {
352
+ "group_size": 32,
353
+ "bits": 4,
354
+ "mode": "mxfp4",
355
+ "model.layers.0.block_sparse_moe.gate": {
356
+ "group_size": 64,
357
+ "bits": 8
358
+ },
359
+ "model.layers.1.block_sparse_moe.gate": {
360
+ "group_size": 64,
361
+ "bits": 8
362
+ },
363
+ "model.layers.2.block_sparse_moe.gate": {
364
+ "group_size": 64,
365
+ "bits": 8
366
+ },
367
+ "model.layers.3.block_sparse_moe.gate": {
368
+ "group_size": 64,
369
+ "bits": 8
370
+ },
371
+ "model.layers.4.block_sparse_moe.gate": {
372
+ "group_size": 64,
373
+ "bits": 8
374
+ },
375
+ "model.layers.5.block_sparse_moe.gate": {
376
+ "group_size": 64,
377
+ "bits": 8
378
+ },
379
+ "model.layers.6.block_sparse_moe.gate": {
380
+ "group_size": 64,
381
+ "bits": 8
382
+ },
383
+ "model.layers.7.block_sparse_moe.gate": {
384
+ "group_size": 64,
385
+ "bits": 8
386
+ },
387
+ "model.layers.8.block_sparse_moe.gate": {
388
+ "group_size": 64,
389
+ "bits": 8
390
+ },
391
+ "model.layers.9.block_sparse_moe.gate": {
392
+ "group_size": 64,
393
+ "bits": 8
394
+ },
395
+ "model.layers.10.block_sparse_moe.gate": {
396
+ "group_size": 64,
397
+ "bits": 8
398
+ },
399
+ "model.layers.11.block_sparse_moe.gate": {
400
+ "group_size": 64,
401
+ "bits": 8
402
+ },
403
+ "model.layers.12.block_sparse_moe.gate": {
404
+ "group_size": 64,
405
+ "bits": 8
406
+ },
407
+ "model.layers.13.block_sparse_moe.gate": {
408
+ "group_size": 64,
409
+ "bits": 8
410
+ },
411
+ "model.layers.14.block_sparse_moe.gate": {
412
+ "group_size": 64,
413
+ "bits": 8
414
+ },
415
+ "model.layers.15.block_sparse_moe.gate": {
416
+ "group_size": 64,
417
+ "bits": 8
418
+ },
419
+ "model.layers.16.block_sparse_moe.gate": {
420
+ "group_size": 64,
421
+ "bits": 8
422
+ },
423
+ "model.layers.17.block_sparse_moe.gate": {
424
+ "group_size": 64,
425
+ "bits": 8
426
+ },
427
+ "model.layers.18.block_sparse_moe.gate": {
428
+ "group_size": 64,
429
+ "bits": 8
430
+ },
431
+ "model.layers.19.block_sparse_moe.gate": {
432
+ "group_size": 64,
433
+ "bits": 8
434
+ },
435
+ "model.layers.20.block_sparse_moe.gate": {
436
+ "group_size": 64,
437
+ "bits": 8
438
+ },
439
+ "model.layers.21.block_sparse_moe.gate": {
440
+ "group_size": 64,
441
+ "bits": 8
442
+ },
443
+ "model.layers.22.block_sparse_moe.gate": {
444
+ "group_size": 64,
445
+ "bits": 8
446
+ },
447
+ "model.layers.23.block_sparse_moe.gate": {
448
+ "group_size": 64,
449
+ "bits": 8
450
+ },
451
+ "model.layers.24.block_sparse_moe.gate": {
452
+ "group_size": 64,
453
+ "bits": 8
454
+ },
455
+ "model.layers.25.block_sparse_moe.gate": {
456
+ "group_size": 64,
457
+ "bits": 8
458
+ },
459
+ "model.layers.26.block_sparse_moe.gate": {
460
+ "group_size": 64,
461
+ "bits": 8
462
+ },
463
+ "model.layers.27.block_sparse_moe.gate": {
464
+ "group_size": 64,
465
+ "bits": 8
466
+ },
467
+ "model.layers.28.block_sparse_moe.gate": {
468
+ "group_size": 64,
469
+ "bits": 8
470
+ },
471
+ "model.layers.29.block_sparse_moe.gate": {
472
+ "group_size": 64,
473
+ "bits": 8
474
+ },
475
+ "model.layers.30.block_sparse_moe.gate": {
476
+ "group_size": 64,
477
+ "bits": 8
478
+ },
479
+ "model.layers.31.block_sparse_moe.gate": {
480
+ "group_size": 64,
481
+ "bits": 8
482
+ },
483
+ "model.layers.32.block_sparse_moe.gate": {
484
+ "group_size": 64,
485
+ "bits": 8
486
+ },
487
+ "model.layers.33.block_sparse_moe.gate": {
488
+ "group_size": 64,
489
+ "bits": 8
490
+ },
491
+ "model.layers.34.block_sparse_moe.gate": {
492
+ "group_size": 64,
493
+ "bits": 8
494
+ },
495
+ "model.layers.35.block_sparse_moe.gate": {
496
+ "group_size": 64,
497
+ "bits": 8
498
+ },
499
+ "model.layers.36.block_sparse_moe.gate": {
500
+ "group_size": 64,
501
+ "bits": 8
502
+ },
503
+ "model.layers.37.block_sparse_moe.gate": {
504
+ "group_size": 64,
505
+ "bits": 8
506
+ },
507
+ "model.layers.38.block_sparse_moe.gate": {
508
+ "group_size": 64,
509
+ "bits": 8
510
+ },
511
+ "model.layers.39.block_sparse_moe.gate": {
512
+ "group_size": 64,
513
+ "bits": 8
514
+ },
515
+ "model.layers.40.block_sparse_moe.gate": {
516
+ "group_size": 64,
517
+ "bits": 8
518
+ },
519
+ "model.layers.41.block_sparse_moe.gate": {
520
+ "group_size": 64,
521
+ "bits": 8
522
+ },
523
+ "model.layers.42.block_sparse_moe.gate": {
524
+ "group_size": 64,
525
+ "bits": 8
526
+ },
527
+ "model.layers.43.block_sparse_moe.gate": {
528
+ "group_size": 64,
529
+ "bits": 8
530
+ },
531
+ "model.layers.44.block_sparse_moe.gate": {
532
+ "group_size": 64,
533
+ "bits": 8
534
+ },
535
+ "model.layers.45.block_sparse_moe.gate": {
536
+ "group_size": 64,
537
+ "bits": 8
538
+ },
539
+ "model.layers.46.block_sparse_moe.gate": {
540
+ "group_size": 64,
541
+ "bits": 8
542
+ },
543
+ "model.layers.47.block_sparse_moe.gate": {
544
+ "group_size": 64,
545
+ "bits": 8
546
+ },
547
+ "model.layers.48.block_sparse_moe.gate": {
548
+ "group_size": 64,
549
+ "bits": 8
550
+ },
551
+ "model.layers.49.block_sparse_moe.gate": {
552
+ "group_size": 64,
553
+ "bits": 8
554
+ },
555
+ "model.layers.50.block_sparse_moe.gate": {
556
+ "group_size": 64,
557
+ "bits": 8
558
+ },
559
+ "model.layers.51.block_sparse_moe.gate": {
560
+ "group_size": 64,
561
+ "bits": 8
562
+ },
563
+ "model.layers.52.block_sparse_moe.gate": {
564
+ "group_size": 64,
565
+ "bits": 8
566
+ },
567
+ "model.layers.53.block_sparse_moe.gate": {
568
+ "group_size": 64,
569
+ "bits": 8
570
+ },
571
+ "model.layers.54.block_sparse_moe.gate": {
572
+ "group_size": 64,
573
+ "bits": 8
574
+ },
575
+ "model.layers.55.block_sparse_moe.gate": {
576
+ "group_size": 64,
577
+ "bits": 8
578
+ },
579
+ "model.layers.56.block_sparse_moe.gate": {
580
+ "group_size": 64,
581
+ "bits": 8
582
+ },
583
+ "model.layers.57.block_sparse_moe.gate": {
584
+ "group_size": 64,
585
+ "bits": 8
586
+ },
587
+ "model.layers.58.block_sparse_moe.gate": {
588
+ "group_size": 64,
589
+ "bits": 8
590
+ },
591
+ "model.layers.59.block_sparse_moe.gate": {
592
+ "group_size": 64,
593
+ "bits": 8
594
+ },
595
+ "model.layers.60.block_sparse_moe.gate": {
596
+ "group_size": 64,
597
+ "bits": 8
598
+ },
599
+ "model.layers.61.block_sparse_moe.gate": {
600
+ "group_size": 64,
601
+ "bits": 8
602
+ }
603
+ },
604
+ "rms_norm_eps": 1e-06,
605
+ "rope_scaling": null,
606
+ "rope_theta": 5000000,
607
+ "rotary_dim": 64,
608
+ "routed_scaling_factor": 1.0,
609
+ "router_aux_loss_coef": 0.001,
610
+ "router_jitter_noise": 0.0,
611
+ "scoring_func": "sigmoid",
612
+ "shared_intermediate_size": 0,
613
+ "shared_moe_mode": "sigmoid",
614
+ "sliding_window": null,
615
+ "swa_rope_theta": -1.0,
616
+ "tie_word_embeddings": false,
617
+ "topk_group": null,
618
+ "torch_dtype": "bfloat16",
619
+ "transformers_version": "4.55.0",
620
+ "use_cache": false,
621
+ "use_grouped_topk": true,
622
+ "use_mtp": true,
623
+ "use_qk_norm": true,
624
+ "use_routing_bias": true,
625
+ "vocab_size": 200064
626
+ }
configuration_minimax_m2.py ADDED
@@ -0,0 +1,131 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # SPDX-FileCopyrightText: 2024-2025 ModelCloud.ai
2
+ # SPDX-FileCopyrightText: 2024-2025 qubitium@modelcloud.ai
3
+ # SPDX-License-Identifier: Apache-2.0
4
+ # Contact: qubitium@modelcloud.ai, x.com/qubitium
5
+
6
+ """Configuration for the MiniMax M2 architecture."""
7
+
8
+ from __future__ import annotations
9
+
10
+ from typing import List, Optional, Union
11
+
12
+ from transformers.configuration_utils import PretrainedConfig
13
+
14
+
15
+ class MiniMaxM2Config(PretrainedConfig):
16
+ model_type = "minimax"
17
+
18
+ def __init__(
19
+ self,
20
+ vocab_size: int = 200_064,
21
+ hidden_size: int = 3_072,
22
+ intermediate_size: int = 1_536,
23
+ mlp_intermediate_size: int = 8_192,
24
+ num_hidden_layers: int = 62,
25
+ num_attention_heads: int = 48,
26
+ num_key_value_heads: int = 8,
27
+ head_dim: Optional[int] = 128,
28
+ num_local_experts: int = 256,
29
+ num_experts_per_tok: int = 8,
30
+ attn_type_list: Optional[List[int]] = None,
31
+ attention_dropout: float = 0.0,
32
+ hidden_act: str = "silu",
33
+ rms_norm_eps: float = 1e-6,
34
+ max_position_embeddings: int = 196_608,
35
+ rope_theta: float = 5_000_000.0,
36
+ rotary_dim: int = 64,
37
+ rope_scaling: Optional[dict] = None,
38
+ use_qk_norm: bool = True,
39
+ qk_norm_type: str = "per_layer",
40
+ use_routing_bias: bool = True,
41
+ scoring_func: str = "sigmoid",
42
+ router_aux_loss_coef: float = 0.001,
43
+ router_jitter_noise: float = 0.0,
44
+ output_router_logits: bool = False,
45
+ use_grouped_topk: bool = True,
46
+ num_expert_group: Optional[int] = None,
47
+ topk_group: Optional[int] = None,
48
+ routed_scaling_factor: float = 1.0,
49
+ layernorm_full_attention_beta: float = 1.0,
50
+ layernorm_linear_attention_beta: float = 1.0,
51
+ layernorm_mlp_beta: float = 1.0,
52
+ shared_intermediate_size: int = 0,
53
+ shared_moe_mode: str = "sigmoid",
54
+ use_mtp: bool = True,
55
+ num_mtp_modules: int = 3,
56
+ mtp_transformer_layers: int = 1,
57
+ attn_window_size: Optional[Union[int, List[int]]] = None,
58
+ swa_rope_theta: float = -1.0,
59
+ sliding_window: Optional[int] = None,
60
+ initializer_range: float = 0.02,
61
+ tie_word_embeddings: bool = False,
62
+ max_model_len: Optional[int] = None,
63
+ bos_token_id: Optional[int] = None,
64
+ eos_token_id: Optional[int] = None,
65
+ pad_token_id: Optional[int] = None,
66
+ use_cache: bool = True,
67
+ **kwargs,
68
+ ) -> None:
69
+ quantization_config = kwargs.pop("quantization_config", None)
70
+ transformers_version = kwargs.pop("transformers_version", None)
71
+
72
+ super().__init__(
73
+ bos_token_id=bos_token_id,
74
+ eos_token_id=eos_token_id,
75
+ tie_word_embeddings=tie_word_embeddings,
76
+ pad_token_id=pad_token_id,
77
+ **kwargs,
78
+ )
79
+
80
+ self.vocab_size = vocab_size
81
+ self.hidden_size = hidden_size
82
+ self.intermediate_size = intermediate_size
83
+ self.mlp_intermediate_size = mlp_intermediate_size
84
+ self.num_hidden_layers = num_hidden_layers
85
+ self.num_attention_heads = num_attention_heads
86
+ self.num_key_value_heads = num_key_value_heads
87
+ self.head_dim = head_dim or hidden_size // num_attention_heads
88
+ self.num_local_experts = num_local_experts
89
+ self.num_experts_per_tok = num_experts_per_tok
90
+ self.attn_type_list = attn_type_list or [1] * num_hidden_layers
91
+ self.attention_dropout = attention_dropout
92
+ self.hidden_act = hidden_act
93
+ self.rms_norm_eps = rms_norm_eps
94
+ self.max_position_embeddings = max_position_embeddings
95
+ self.rope_theta = rope_theta
96
+ self.rotary_dim = rotary_dim
97
+ self.rope_scaling = rope_scaling
98
+ self.use_qk_norm = use_qk_norm
99
+ self.qk_norm_type = qk_norm_type
100
+ self.use_routing_bias = use_routing_bias
101
+ self.scoring_func = scoring_func
102
+ self.router_aux_loss_coef = router_aux_loss_coef
103
+ self.router_jitter_noise = router_jitter_noise
104
+ self.output_router_logits = output_router_logits
105
+ self.use_grouped_topk = use_grouped_topk
106
+ self.num_expert_group = num_expert_group
107
+ self.topk_group = topk_group
108
+ self.routed_scaling_factor = routed_scaling_factor
109
+ self.layernorm_full_attention_beta = layernorm_full_attention_beta
110
+ self.layernorm_linear_attention_beta = layernorm_linear_attention_beta
111
+ self.layernorm_mlp_beta = layernorm_mlp_beta
112
+ self.shared_intermediate_size = shared_intermediate_size
113
+ self.shared_moe_mode = shared_moe_mode
114
+ self.use_mtp = use_mtp
115
+ self.num_mtp_modules = num_mtp_modules
116
+ self.mtp_transformer_layers = mtp_transformer_layers
117
+ self.attn_window_size = attn_window_size
118
+ self.swa_rope_theta = swa_rope_theta
119
+ self.sliding_window = sliding_window
120
+ self.initializer_range = initializer_range
121
+ self.max_model_len = max_model_len
122
+ self.use_cache = use_cache
123
+
124
+ # Convenient accessor used by rotary embedding helper
125
+ self.partial_rotary_factor = float(self.rotary_dim) / float(self.head_dim)
126
+ if quantization_config is not None:
127
+ self.quantization_config = quantization_config
128
+ self.transformers_version = transformers_version
129
+
130
+
131
+ __all__ = ["MiniMaxM2Config"]
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model-00001-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e87544928809f2f8e58da28b582734d53325cb6faa320fda33601f88e0d770cc
3
+ size 5121851158
model-00002-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:787d2c6410febdab92b040e105a003531060b4666cefcf5a52332038681ed95a
3
+ size 5359821650
model-00003-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a0e2a47867df3b19733043431fed5da647d35babcd05d887764898664446aa3
3
+ size 5103248235
model-00004-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7245755b526989f395843972108130ce88fbf279e18105ce371952eec332cf14
3
+ size 5366505963
model-00005-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c176cd93cc5466e8157f1288549b6149243025e143bc04a343977d8d2e7467d1
3
+ size 5363556730
model-00006-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a78919a335b644f0d3506a2f1b295a20df4b87c3c97713d3ce51f6721ae0f32
3
+ size 5359808638
model-00007-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32d24817f9eefb541e301a37dd308e2a3c70a98dae3f20110b8495bd9eb8f502
3
+ size 5103248299
model-00008-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48b95bf7a21153866a25ec5098ec093d9ce4c4d2cda633ee49d85708da2fda4f
3
+ size 5359821761
model-00009-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7587b5559fbf0db4d92b25d3f461cc2f8dcd6a7dfadabdab5b38236e95efa93f
3
+ size 5103248313
model-00010-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc066d1d4de449a22f133c0ff4f03372d226873a6ddfbdd8ef1173d4a77635b7
3
+ size 5359821749
model-00011-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:79bc1aa7d4192eee98d9c39a609c92d9a48d7a52180d32345fb7513f8c78d9cf
3
+ size 3613659423
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<code_interpreter>",
4
+ "<commit_after>",
5
+ "<commit_before>",
6
+ "<commit_msg>",
7
+ "<empty_output>",
8
+ "<filename>",
9
+ "<fim_middle>",
10
+ "<fim_pad>",
11
+ "<fim_prefix>",
12
+ "<fim_suffix>",
13
+ "<function_call>",
14
+ "<gh_stars>",
15
+ "]<]speech[>[",
16
+ "]<]image[>[",
17
+ "]<]video[>[",
18
+ "]<]start of speech[>[",
19
+ "]<]end of speech[>[",
20
+ "]<]start of image[>[",
21
+ "]<]end of image[>[",
22
+ "]<]start of video[>[",
23
+ "]<]end of video[>[",
24
+ "]<]vision pad[>[",
25
+ "]~!b[",
26
+ "<issue_closed>",
27
+ "<issue_comment>",
28
+ "<issue_start>",
29
+ "<jupyter_code>",
30
+ "<jupyter_output>",
31
+ "<jupyter_start>",
32
+ "<jupyter_text>",
33
+ "<reponame>",
34
+ "[e~[",
35
+ "]!d~[",
36
+ "]!p~[",
37
+ "]~b]",
38
+ "<jupyter_error>",
39
+ "<add_file>",
40
+ "<delete_file>",
41
+ "<rename_file>",
42
+ "<edit_file>",
43
+ "<commit_message>",
44
+ "<empty_source_file>",
45
+ "<repo_struct>",
46
+ "<code_context>",
47
+ "<file_content>",
48
+ "<source_files>",
49
+ "<pr_start>",
50
+ "<review_comment>",
51
+ "<filepath>",
52
+ "<file_sep>"
53
+ ],
54
+ "bos_token": {
55
+ "content": "]~!b[",
56
+ "lstrip": false,
57
+ "normalized": false,
58
+ "rstrip": false,
59
+ "single_word": false
60
+ },
61
+ "eos_token": {
62
+ "content": "[e~[",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false
67
+ },
68
+ "unk_token": {
69
+ "content": "]!d~[",
70
+ "lstrip": false,
71
+ "normalized": false,
72
+ "rstrip": false,
73
+ "single_word": false
74
+ }
75
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7b90ed7f55d905175bc26771d6d7d33b40b46742f073675bc816fedaf482ea1
3
+ size 15522763
tokenizer_config.json ADDED
@@ -0,0 +1,496 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "200000": {
5
+ "content": "]!p~[",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "200001": {
13
+ "content": "<fim_prefix>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "200002": {
21
+ "content": "<fim_middle>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "200003": {
29
+ "content": "<fim_suffix>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
+ },
36
+ "200004": {
37
+ "content": "<fim_pad>",
38
+ "lstrip": false,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": true
43
+ },
44
+ "200005": {
45
+ "content": "<reponame>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": true
51
+ },
52
+ "200006": {
53
+ "content": "<filename>",
54
+ "lstrip": false,
55
+ "normalized": false,
56
+ "rstrip": false,
57
+ "single_word": false,
58
+ "special": true
59
+ },
60
+ "200007": {
61
+ "content": "<gh_stars>",
62
+ "lstrip": false,
63
+ "normalized": false,
64
+ "rstrip": false,
65
+ "single_word": false,
66
+ "special": true
67
+ },
68
+ "200008": {
69
+ "content": "<issue_start>",
70
+ "lstrip": false,
71
+ "normalized": false,
72
+ "rstrip": false,
73
+ "single_word": false,
74
+ "special": true
75
+ },
76
+ "200009": {
77
+ "content": "<issue_comment>",
78
+ "lstrip": false,
79
+ "normalized": false,
80
+ "rstrip": false,
81
+ "single_word": false,
82
+ "special": true
83
+ },
84
+ "200010": {
85
+ "content": "<issue_closed>",
86
+ "lstrip": false,
87
+ "normalized": false,
88
+ "rstrip": false,
89
+ "single_word": false,
90
+ "special": true
91
+ },
92
+ "200011": {
93
+ "content": "<jupyter_start>",
94
+ "lstrip": false,
95
+ "normalized": false,
96
+ "rstrip": false,
97
+ "single_word": false,
98
+ "special": true
99
+ },
100
+ "200012": {
101
+ "content": "<jupyter_text>",
102
+ "lstrip": false,
103
+ "normalized": false,
104
+ "rstrip": false,
105
+ "single_word": false,
106
+ "special": true
107
+ },
108
+ "200013": {
109
+ "content": "<jupyter_code>",
110
+ "lstrip": false,
111
+ "normalized": false,
112
+ "rstrip": false,
113
+ "single_word": false,
114
+ "special": true
115
+ },
116
+ "200014": {
117
+ "content": "<jupyter_output>",
118
+ "lstrip": false,
119
+ "normalized": false,
120
+ "rstrip": false,
121
+ "single_word": false,
122
+ "special": true
123
+ },
124
+ "200015": {
125
+ "content": "<empty_output>",
126
+ "lstrip": false,
127
+ "normalized": false,
128
+ "rstrip": false,
129
+ "single_word": false,
130
+ "special": true
131
+ },
132
+ "200016": {
133
+ "content": "<commit_before>",
134
+ "lstrip": false,
135
+ "normalized": false,
136
+ "rstrip": false,
137
+ "single_word": false,
138
+ "special": true
139
+ },
140
+ "200017": {
141
+ "content": "<commit_msg>",
142
+ "lstrip": false,
143
+ "normalized": false,
144
+ "rstrip": false,
145
+ "single_word": false,
146
+ "special": true
147
+ },
148
+ "200018": {
149
+ "content": "<commit_after>",
150
+ "lstrip": false,
151
+ "normalized": false,
152
+ "rstrip": false,
153
+ "single_word": false,
154
+ "special": true
155
+ },
156
+ "200019": {
157
+ "content": "]~b]",
158
+ "lstrip": false,
159
+ "normalized": false,
160
+ "rstrip": false,
161
+ "single_word": false,
162
+ "special": true
163
+ },
164
+ "200020": {
165
+ "content": "[e~[",
166
+ "lstrip": false,
167
+ "normalized": false,
168
+ "rstrip": false,
169
+ "single_word": false,
170
+ "special": true
171
+ },
172
+ "200021": {
173
+ "content": "]!d~[",
174
+ "lstrip": false,
175
+ "normalized": false,
176
+ "rstrip": false,
177
+ "single_word": false,
178
+ "special": true
179
+ },
180
+ "200022": {
181
+ "content": "<function_call>",
182
+ "lstrip": false,
183
+ "normalized": false,
184
+ "rstrip": false,
185
+ "single_word": false,
186
+ "special": true
187
+ },
188
+ "200023": {
189
+ "content": "<code_interpreter>",
190
+ "lstrip": false,
191
+ "normalized": false,
192
+ "rstrip": false,
193
+ "single_word": false,
194
+ "special": true
195
+ },
196
+ "200024": {
197
+ "content": "]<]speech[>[",
198
+ "lstrip": false,
199
+ "normalized": false,
200
+ "rstrip": false,
201
+ "single_word": false,
202
+ "special": true
203
+ },
204
+ "200025": {
205
+ "content": "]<]image[>[",
206
+ "lstrip": false,
207
+ "normalized": false,
208
+ "rstrip": false,
209
+ "single_word": false,
210
+ "special": true
211
+ },
212
+ "200026": {
213
+ "content": "]<]video[>[",
214
+ "lstrip": false,
215
+ "normalized": false,
216
+ "rstrip": false,
217
+ "single_word": false,
218
+ "special": true
219
+ },
220
+ "200027": {
221
+ "content": "]<]start of speech[>[",
222
+ "lstrip": false,
223
+ "normalized": false,
224
+ "rstrip": false,
225
+ "single_word": false,
226
+ "special": true
227
+ },
228
+ "200028": {
229
+ "content": "]<]end of speech[>[",
230
+ "lstrip": false,
231
+ "normalized": false,
232
+ "rstrip": false,
233
+ "single_word": false,
234
+ "special": true
235
+ },
236
+ "200029": {
237
+ "content": "]<]start of image[>[",
238
+ "lstrip": false,
239
+ "normalized": false,
240
+ "rstrip": false,
241
+ "single_word": false,
242
+ "special": true
243
+ },
244
+ "200030": {
245
+ "content": "]<]end of image[>[",
246
+ "lstrip": false,
247
+ "normalized": false,
248
+ "rstrip": false,
249
+ "single_word": false,
250
+ "special": true
251
+ },
252
+ "200031": {
253
+ "content": "]<]start of video[>[",
254
+ "lstrip": false,
255
+ "normalized": false,
256
+ "rstrip": false,
257
+ "single_word": false,
258
+ "special": true
259
+ },
260
+ "200032": {
261
+ "content": "]<]end of video[>[",
262
+ "lstrip": false,
263
+ "normalized": false,
264
+ "rstrip": false,
265
+ "single_word": false,
266
+ "special": true
267
+ },
268
+ "200033": {
269
+ "content": "]<]vision pad[>[",
270
+ "lstrip": false,
271
+ "normalized": false,
272
+ "rstrip": false,
273
+ "single_word": false,
274
+ "special": true
275
+ },
276
+ "200034": {
277
+ "content": "]~!b[",
278
+ "lstrip": false,
279
+ "normalized": false,
280
+ "rstrip": false,
281
+ "single_word": false,
282
+ "special": true
283
+ },
284
+ "200035": {
285
+ "content": "<jupyter_error>",
286
+ "lstrip": false,
287
+ "normalized": false,
288
+ "rstrip": false,
289
+ "single_word": false,
290
+ "special": true
291
+ },
292
+ "200036": {
293
+ "content": "<add_file>",
294
+ "lstrip": false,
295
+ "normalized": false,
296
+ "rstrip": false,
297
+ "single_word": false,
298
+ "special": true
299
+ },
300
+ "200037": {
301
+ "content": "<delete_file>",
302
+ "lstrip": false,
303
+ "normalized": false,
304
+ "rstrip": false,
305
+ "single_word": false,
306
+ "special": true
307
+ },
308
+ "200038": {
309
+ "content": "<rename_file>",
310
+ "lstrip": false,
311
+ "normalized": false,
312
+ "rstrip": false,
313
+ "single_word": false,
314
+ "special": true
315
+ },
316
+ "200039": {
317
+ "content": "<edit_file>",
318
+ "lstrip": false,
319
+ "normalized": false,
320
+ "rstrip": false,
321
+ "single_word": false,
322
+ "special": true
323
+ },
324
+ "200040": {
325
+ "content": "<commit_message>",
326
+ "lstrip": false,
327
+ "normalized": false,
328
+ "rstrip": false,
329
+ "single_word": false,
330
+ "special": true
331
+ },
332
+ "200041": {
333
+ "content": "<empty_source_file>",
334
+ "lstrip": false,
335
+ "normalized": false,
336
+ "rstrip": false,
337
+ "single_word": false,
338
+ "special": true
339
+ },
340
+ "200042": {
341
+ "content": "<repo_struct>",
342
+ "lstrip": false,
343
+ "normalized": false,
344
+ "rstrip": false,
345
+ "single_word": false,
346
+ "special": true
347
+ },
348
+ "200043": {
349
+ "content": "<code_context>",
350
+ "lstrip": false,
351
+ "normalized": false,
352
+ "rstrip": false,
353
+ "single_word": false,
354
+ "special": true
355
+ },
356
+ "200044": {
357
+ "content": "<file_content>",
358
+ "lstrip": false,
359
+ "normalized": false,
360
+ "rstrip": false,
361
+ "single_word": false,
362
+ "special": true
363
+ },
364
+ "200045": {
365
+ "content": "<source_files>",
366
+ "lstrip": false,
367
+ "normalized": false,
368
+ "rstrip": false,
369
+ "single_word": false,
370
+ "special": true
371
+ },
372
+ "200046": {
373
+ "content": "<pr_start>",
374
+ "lstrip": false,
375
+ "normalized": false,
376
+ "rstrip": false,
377
+ "single_word": false,
378
+ "special": true
379
+ },
380
+ "200047": {
381
+ "content": "<review_comment>",
382
+ "lstrip": false,
383
+ "normalized": false,
384
+ "rstrip": false,
385
+ "single_word": false,
386
+ "special": true
387
+ },
388
+ "200048": {
389
+ "content": "<filepath>",
390
+ "lstrip": false,
391
+ "normalized": false,
392
+ "rstrip": false,
393
+ "single_word": false,
394
+ "special": true
395
+ },
396
+ "200049": {
397
+ "content": "<file_sep>",
398
+ "lstrip": false,
399
+ "normalized": false,
400
+ "rstrip": false,
401
+ "single_word": false,
402
+ "special": true
403
+ },
404
+ "200050": {
405
+ "content": "<think>",
406
+ "lstrip": false,
407
+ "normalized": false,
408
+ "rstrip": false,
409
+ "single_word": false,
410
+ "special": false
411
+ },
412
+ "200051": {
413
+ "content": "</think>",
414
+ "lstrip": false,
415
+ "normalized": false,
416
+ "rstrip": false,
417
+ "single_word": false,
418
+ "special": false
419
+ },
420
+ "200052": {
421
+ "content": "<minimax:tool_call>",
422
+ "lstrip": false,
423
+ "normalized": false,
424
+ "rstrip": false,
425
+ "single_word": false,
426
+ "special": false
427
+ },
428
+ "200053": {
429
+ "content": "</minimax:tool_call>",
430
+ "lstrip": false,
431
+ "normalized": false,
432
+ "rstrip": false,
433
+ "single_word": false,
434
+ "special": false
435
+ }
436
+ },
437
+ "additional_special_tokens": [
438
+ "<code_interpreter>",
439
+ "<commit_after>",
440
+ "<commit_before>",
441
+ "<commit_msg>",
442
+ "<empty_output>",
443
+ "<filename>",
444
+ "<fim_middle>",
445
+ "<fim_pad>",
446
+ "<fim_prefix>",
447
+ "<fim_suffix>",
448
+ "<function_call>",
449
+ "<gh_stars>",
450
+ "]<]speech[>[",
451
+ "]<]image[>[",
452
+ "]<]video[>[",
453
+ "]<]start of speech[>[",
454
+ "]<]end of speech[>[",
455
+ "]<]start of image[>[",
456
+ "]<]end of image[>[",
457
+ "]<]start of video[>[",
458
+ "]<]end of video[>[",
459
+ "]<]vision pad[>[",
460
+ "]~!b[",
461
+ "<issue_closed>",
462
+ "<issue_comment>",
463
+ "<issue_start>",
464
+ "<jupyter_code>",
465
+ "<jupyter_output>",
466
+ "<jupyter_start>",
467
+ "<jupyter_text>",
468
+ "<reponame>",
469
+ "[e~[",
470
+ "]!d~[",
471
+ "]!p~[",
472
+ "]~b]",
473
+ "<jupyter_error>",
474
+ "<add_file>",
475
+ "<delete_file>",
476
+ "<rename_file>",
477
+ "<edit_file>",
478
+ "<commit_message>",
479
+ "<empty_source_file>",
480
+ "<repo_struct>",
481
+ "<code_context>",
482
+ "<file_content>",
483
+ "<source_files>",
484
+ "<pr_start>",
485
+ "<review_comment>",
486
+ "<filepath>",
487
+ "<file_sep>"
488
+ ],
489
+ "bos_token": "]~!b[",
490
+ "clean_up_tokenization_spaces": false,
491
+ "eos_token": "[e~[",
492
+ "extra_special_tokens": {},
493
+ "model_max_length": 40960000,
494
+ "tokenizer_class": "GPT2Tokenizer",
495
+ "unk_token": "]!d~["
496
+ }
vocab.json ADDED
The diff for this file is too large to render. See raw diff