Infinite3214 commited on
Commit
4fb8489
·
verified ·
1 Parent(s): 31f9003

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +1 -0
  2. README.md +253 -0
  3. chat_template.jinja +86 -0
  4. config.json +43 -0
  5. generation_config.json +10 -0
  6. model-00001-of-00102.safetensors +3 -0
  7. model-00002-of-00102.safetensors +3 -0
  8. model-00003-of-00102.safetensors +3 -0
  9. model-00004-of-00102.safetensors +3 -0
  10. model-00005-of-00102.safetensors +3 -0
  11. model-00006-of-00102.safetensors +3 -0
  12. model-00007-of-00102.safetensors +3 -0
  13. model-00008-of-00102.safetensors +3 -0
  14. model-00009-of-00102.safetensors +3 -0
  15. model-00010-of-00102.safetensors +3 -0
  16. model-00011-of-00102.safetensors +3 -0
  17. model-00012-of-00102.safetensors +3 -0
  18. model-00013-of-00102.safetensors +3 -0
  19. model-00014-of-00102.safetensors +3 -0
  20. model-00015-of-00102.safetensors +3 -0
  21. model-00016-of-00102.safetensors +3 -0
  22. model-00017-of-00102.safetensors +3 -0
  23. model-00018-of-00102.safetensors +3 -0
  24. model-00019-of-00102.safetensors +3 -0
  25. model-00020-of-00102.safetensors +3 -0
  26. model-00021-of-00102.safetensors +3 -0
  27. model-00022-of-00102.safetensors +3 -0
  28. model-00023-of-00102.safetensors +3 -0
  29. model-00024-of-00102.safetensors +3 -0
  30. model-00025-of-00102.safetensors +3 -0
  31. model-00026-of-00102.safetensors +3 -0
  32. model-00027-of-00102.safetensors +3 -0
  33. model-00028-of-00102.safetensors +3 -0
  34. model-00029-of-00102.safetensors +3 -0
  35. model-00030-of-00102.safetensors +3 -0
  36. model-00031-of-00102.safetensors +3 -0
  37. model-00032-of-00102.safetensors +3 -0
  38. model-00033-of-00102.safetensors +3 -0
  39. model-00034-of-00102.safetensors +3 -0
  40. model-00035-of-00102.safetensors +3 -0
  41. model-00036-of-00102.safetensors +3 -0
  42. model-00037-of-00102.safetensors +3 -0
  43. model-00038-of-00102.safetensors +3 -0
  44. model-00039-of-00102.safetensors +3 -0
  45. model-00040-of-00102.safetensors +3 -0
  46. model-00041-of-00102.safetensors +3 -0
  47. model-00042-of-00102.safetensors +3 -0
  48. model-00043-of-00102.safetensors +3 -0
  49. model-00044-of-00102.safetensors +3 -0
  50. model-00045-of-00102.safetensors +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,253 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ library_name: transformers
5
+ tags:
6
+ - glm
7
+ - glm4
8
+ - MOE
9
+ - pruning
10
+ - compression
11
+ - reap
12
+ - cerebras
13
+ - code
14
+ - function-calling
15
+ - agentic
16
+ license: apache-2.0
17
+ pipeline_tag: text-generation
18
+ base_model:
19
+ - zai/glm-4.7
20
+ ---
21
+
22
+ <p align="center">
23
+ <em>𓌳 <strong>REAP</strong>𓌳 the Experts: Why Pruning Prevails for One-Shot MoE Compression</em><br>
24
+ <a href="https://arxiv.org/abs/2510.13999">📄 Paper</a> • <a href="https://github.com/CerebrasResearch/reap">💻 Code</a> • <a href="https://www.cerebras.ai/blog/reap">📝 Blog</a>
25
+ </p>
26
+
27
+ # GLM-4.7-REAP-30
28
+
29
+ ## ✨ Highlights
30
+
31
+ **30% Expert-Pruned** GLM-4.7 optimized for **code generation**, **function calling**, and **agentic workflows**.
32
+
33
+ Created using **[REAP (Router-weighted Expert Activation Pruning)](https://arxiv.org/abs/2510.13999)** by Cerebras:
34
+
35
+ - **358B → 251B**: 30% of MoE experts pruned (112/160 remaining)
36
+ - **Calibrated for Code & Tools**: Preserves coding and function-calling capabilities
37
+ - **One-Shot Compression**: No fine-tuning required
38
+ - **Drop-in Compatible**: Works with vLLM, Transformers, SGLang
39
+
40
+ ### 🙏 Acknowledgments
41
+
42
+ - **[Prime Intellect](https://www.primeintellect.ai/)** — Compute sponsorship (8x H200 cluster)
43
+ - **[Cerebras](https://www.cerebras.net/)** — [REAP methodology](https://arxiv.org/abs/2510.13999)
44
+
45
+ ---
46
+
47
+ ## 📋 Model Specifications
48
+
49
+ | Property | Value |
50
+ |----------|-------|
51
+ | **Base Model** | [zai/glm-4.7](https://huggingface.co/zai/glm-4.7) |
52
+ | **Architecture** | Sparse Mixture-of-Experts (SMoE) |
53
+ | **Original Parameters** | 358B |
54
+ | **Pruned Parameters** | 251B |
55
+ | **Compression** | 30% experts removed |
56
+ | **Experts per Layer** | 112 (was 160) |
57
+ | **MoE Layers** | 92 |
58
+ | **Activated Experts** | 8 per token |
59
+ | **Precision** | BF16 |
60
+ | **Disk Size** | ~470GB |
61
+ | **VRAM Required** | ~470GB |
62
+
63
+ ---
64
+
65
+
66
+ ## 🔬 Calibration Dataset: Deep Dive
67
+
68
+ REAP's effectiveness depends critically on **calibration data that represents the target use case**. We specifically optimized for **code generation**, **function/tool calling**, and **agentic workflows**.
69
+
70
+ ### Why These 3 Datasets?
71
+
72
+ | Dataset | Samples | Purpose | Why It Matters |
73
+ |---------|---------|---------|----------------|
74
+ | [evol-codealpaca-v1](https://huggingface.co/datasets/theblackcat102/evol-codealpaca-v1) | 700 | Code generation | **51% of mix** — Code tasks activate specific expert pathways; pruning without code calibration destroys coding ability |
75
+ | [xlam-function-calling-60k](https://huggingface.co/datasets/Salesforce/xlam-function-calling-60k) | 330 | Function/tool calling | **24% of mix** — Tool use requires structured JSON output; experts handling schema generation must be preserved |
76
+ | [SWE-smith-trajectories](https://huggingface.co/datasets/SWE-bench/SWE-smith-trajectories) | 330 | Agentic multi-turn | **24% of mix** — Real SWE-bench trajectories with tool calls, file edits, and multi-step reasoning |
77
+
78
+ ### The Science Behind Dataset Selection
79
+
80
+ ```
81
+ REAP Algorithm:
82
+ 1. Forward pass calibration samples through model
83
+ 2. Record which experts activate and their magnitudes
84
+ 3. Compute saliency = router_weight × activation_norm
85
+ 4. Prune lowest-saliency experts
86
+
87
+ Key Insight: Experts are TASK-SPECIFIC
88
+ ├── Some experts specialize in natural language
89
+ ├── Some experts specialize in code syntax
90
+ ├── Some experts specialize in JSON/structured output
91
+ └── Some experts specialize in multi-turn context
92
+
93
+ If calibration lacks code → code-specialized experts appear "unused" → get pruned → model loses coding ability
94
+ ```
95
+
96
+ ### Cerebras' Original Mix (from paper)
97
+
98
+ Cerebras used the same 3 datasets in their GLM-4.6 REAP experiments:
99
+ - evol-codealpaca-v1 for code generation
100
+ - xlam-function-calling-60k for tool calling
101
+ - SWE-smith-trajectories for agentic tasks
102
+
103
+ We followed this exact recipe for reproducibility.
104
+
105
+ ### Combined Dataset
106
+
107
+ Our calibration mix: [0xSero/glm47-reap-calibration-v2](https://huggingface.co/datasets/0xSero/glm47-reap-calibration-v2)
108
+
109
+
110
+ ---
111
+
112
+ ## 📦 Related Models
113
+
114
+ | Model | Params | Experts | Size | Format |
115
+ |-------|--------|---------|------|--------|
116
+ | [GLM-4.7-REAP-30](https://huggingface.co/0xSero/GLM-4.7-REAP-30) | 251B | 112 | ~470GB | BF16 |
117
+ | [GLM-4.7-REAP-35](https://huggingface.co/0xSero/GLM-4.7-REAP-35) | 233B | 104 | ~439GB | BF16 |
118
+ | [GLM-4.7-REAP-40](https://huggingface.co/0xSero/GLM-4.7-REAP-40) | 218B | 96 | ~407GB | BF16 |
119
+ | [GLM-4.7-REAP-45](https://huggingface.co/0xSero/GLM-4.7-REAP-45) | 197B | 88 | ~370GB | BF16 |
120
+ | [GLM-4.7-REAP-50](https://huggingface.co/0xSero/GLM-4.7-REAP-50) | 179B | 80 | ~345GB | BF16 |
121
+ | [GLM-4.7-REAP-40-W4A16](https://huggingface.co/0xSero/GLM-4.7-REAP-40-W4A16) | 218B | 96 | ~108GB | GPTQ |
122
+ | [GLM-4.7-REAP-50-W4A16](https://huggingface.co/0xSero/GLM-4.7-REAP-50-W4A16) | 179B | 80 | ~92GB | GPTQ |
123
+
124
+ ---
125
+
126
+ ## 🚀 Deployment
127
+
128
+ ### vLLM (Recommended)
129
+
130
+ ```bash
131
+ vllm serve 0xSero/GLM-4.7-REAP-30 \
132
+ --tensor-parallel-size 8 \
133
+ --trust-remote-code \
134
+ --dtype bfloat16
135
+ ```
136
+
137
+ ### Transformers
138
+
139
+ ```python
140
+ import torch
141
+ from transformers import AutoModelForCausalLM, AutoTokenizer
142
+
143
+ model = AutoModelForCausalLM.from_pretrained(
144
+ "0xSero/GLM-4.7-REAP-30",
145
+ torch_dtype=torch.bfloat16,
146
+ device_map="auto",
147
+ trust_remote_code=True
148
+ )
149
+ tokenizer = AutoTokenizer.from_pretrained("0xSero/GLM-4.7-REAP-30", trust_remote_code=True)
150
+
151
+ messages = [{"role": "user", "content": "Write a Python function to merge two sorted lists."}]
152
+ inputs = tokenizer.apply_chat_template(messages, return_tensors="pt", add_generation_prompt=True)
153
+ outputs = model.generate(inputs.to(model.device), max_new_tokens=512, temperature=0.7)
154
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
155
+ ```
156
+
157
+ ---
158
+
159
+ ## 🧩 Reproduction
160
+
161
+ ### REAP Pruning Script
162
+
163
+
164
+ ```python
165
+ #!/usr/bin/env python3
166
+ """
167
+ REAP Pruning Script for MoE Models
168
+ Adapted from: https://github.com/CerebrasResearch/reap
169
+ """
170
+
171
+ import subprocess
172
+ import sys
173
+
174
+ def run_reap(
175
+ model_path: str,
176
+ compression_ratio: float,
177
+ dataset: str = "0xSero/glm47-reap-calibration-v2",
178
+ samples: int = 1360,
179
+ seed: int = 42,
180
+ distance: str = "angular",
181
+ reuse_observations: str = None,
182
+ ):
183
+ """
184
+ Run REAP expert pruning.
185
+
186
+ Args:
187
+ model_path: Path to base model
188
+ compression_ratio: 0.30 = prune 30%, keep 70%
189
+ dataset: Calibration dataset (code + tools + agentic)
190
+ samples: Number of calibration samples
191
+ seed: Random seed for reproducibility
192
+ distance: Distance metric for expert clustering
193
+ reuse_observations: Path to pre-computed observations for instant pruning
194
+ """
195
+ cmd = [
196
+ sys.executable, "src/reap/prune.py",
197
+ "--model-name", model_path,
198
+ "--dataset-name", dataset,
199
+ "--compression-ratio", str(compression_ratio),
200
+ "--prune-method", "reap",
201
+ "--seed", str(seed),
202
+ "--samples_per_category", str(samples),
203
+ "--model_max_length", "2048",
204
+ "--distance_measure", distance,
205
+ "--record_pruning_metrics_only", "true",
206
+ ]
207
+
208
+ if reuse_observations:
209
+ # Instant pruning: skip calibration, reuse precomputed expert scores
210
+ cmd.extend(["--load_observations", reuse_observations])
211
+
212
+ subprocess.run(cmd, check=True)
213
+
214
+ # Example: Create 40% pruned model
215
+ run_reap(
216
+ model_path="/path/to/GLM-4.7",
217
+ compression_ratio=0.40, # Prune 40% of experts
218
+ )
219
+ ```
220
+
221
+
222
+ ### Observation Reuse (Instant Multi-Ratio Pruning)
223
+
224
+ REAP computes expert saliency scores during calibration. These scores are **compression-ratio independent**, enabling instant pruning at any ratio:
225
+
226
+ ```bash
227
+ # First run: compute observations (~5 hours)
228
+ python prune.py --compression-ratio 0.40 --output_file_name observations.pt
229
+
230
+ # Subsequent runs: instant pruning (<5 minutes)
231
+ python prune.py --compression-ratio 0.30 --load_observations observations.pt
232
+ python prune.py --compression-ratio 0.50 --load_observations observations.pt
233
+ ```
234
+
235
+ ---
236
+
237
+ ## ⚖️ License
238
+
239
+ Apache 2.0 (inherited from GLM-4)
240
+
241
+ ---
242
+
243
+ ## 🧾 Citation
244
+
245
+ ```bibtex
246
+ @article{lasby2025reap,
247
+ title={REAP the Experts: Why Pruning Prevails for One-Shot MoE Compression},
248
+ author={Lasby, Mike and Lazarevich, Ivan and Sinnadurai, Nish and Lie, Sean and Ioannou, Yani and Thangarasa, Vithursan},
249
+ journal={arXiv preprint arXiv:2510.13999},
250
+ year={2025},
251
+ url={https://arxiv.org/abs/2510.13999}
252
+ }
253
+ ```
chat_template.jinja ADDED
@@ -0,0 +1,86 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [gMASK]<sop>
2
+ {%- if tools -%}
3
+ <|system|>
4
+ # Tools
5
+
6
+ You may call one or more functions to assist with the user query.
7
+
8
+ You are provided with function signatures within <tools></tools> XML tags:
9
+ <tools>
10
+ {% for tool in tools %}
11
+ {{ tool | tojson(ensure_ascii=False) }}
12
+ {% endfor %}
13
+ </tools>
14
+
15
+ For each function call, output the function name and arguments within the following XML format:
16
+ <tool_call>{function-name}<arg_key>{arg-key-1}</arg_key><arg_value>{arg-value-1}</arg_value><arg_key>{arg-key-2}</arg_key><arg_value>{arg-value-2}</arg_value>...</tool_call>{%- endif -%}
17
+ {%- macro visible_text(content) -%}
18
+ {%- if content is string -%}
19
+ {{- content }}
20
+ {%- elif content is iterable and content is not mapping -%}
21
+ {%- for item in content -%}
22
+ {%- if item is mapping and item.type == 'text' -%}
23
+ {{- item.text }}
24
+ {%- elif item is string -%}
25
+ {{- item }}
26
+ {%- endif -%}
27
+ {%- endfor -%}
28
+ {%- else -%}
29
+ {{- content }}
30
+ {%- endif -%}
31
+ {%- endmacro -%}
32
+ {%- set ns = namespace(last_user_index=-1) %}
33
+ {%- for m in messages %}
34
+ {%- if m.role == 'user' %}
35
+ {% set ns.last_user_index = loop.index0 -%}
36
+ {%- endif %}
37
+ {%- endfor %}
38
+ {% for m in messages %}
39
+ {%- if m.role == 'user' -%}<|user|>{{ visible_text(m.content) }}
40
+ {%- elif m.role == 'assistant' -%}
41
+ <|assistant|>
42
+ {%- set reasoning_content = '' %}
43
+ {%- set content = visible_text(m.content) %}
44
+ {%- if m.reasoning_content is string %}
45
+ {%- set reasoning_content = m.reasoning_content %}
46
+ {%- else %}
47
+ {%- if '</think>' in content %}
48
+ {%- set reasoning_content = content.split('</think>')[0].rstrip('\n').split('<think>')[-1].lstrip('\n') %}
49
+ {%- set content = content.split('</think>')[-1].lstrip('\n') %}
50
+ {%- endif %}
51
+ {%- endif %}
52
+ {%- if ((clear_thinking is defined and not clear_thinking) or loop.index0 > ns.last_user_index) and reasoning_content -%}
53
+ {{ '<think>' + reasoning_content.strip() + '</think>'}}
54
+ {%- else -%}
55
+ {{ '</think>' }}
56
+ {%- endif -%}
57
+ {%- if content.strip() -%}
58
+ {{ content.strip() }}
59
+ {%- endif -%}
60
+ {% if m.tool_calls %}
61
+ {% for tc in m.tool_calls %}
62
+ {%- if tc.function %}
63
+ {%- set tc = tc.function %}
64
+ {%- endif %}
65
+ {{- '<tool_call>' + tc.name -}}
66
+ {% set _args = tc.arguments %}{% for k, v in _args.items() %}<arg_key>{{ k }}</arg_key><arg_value>{{ v | tojson(ensure_ascii=False) if v is not string else v }}</arg_value>{% endfor %}</tool_call>{% endfor %}
67
+ {% endif %}
68
+ {%- elif m.role == 'tool' -%}
69
+ {%- if m.content is string -%}
70
+ {%- if loop.first or (messages[loop.index0 - 1].role != "tool") %}
71
+ {{- '<|observation|>' }}
72
+ {%- endif %}
73
+ {{- '<tool_response>' }}
74
+ {{- m.content }}
75
+ {{- '</tool_response>' }}
76
+ {%- else -%}
77
+ <|observation|>{% for tr in m.content %}
78
+ <tool_response>{{ tr.output if tr.output is defined else tr }}</tool_response>{% endfor -%}
79
+ {% endif -%}
80
+ {%- elif m.role == 'system' -%}
81
+ <|system|>{{ visible_text(m.content) }}
82
+ {%- endif -%}
83
+ {%- endfor -%}
84
+ {%- if add_generation_prompt -%}
85
+ <|assistant|>{{- '</think>' if (enable_thinking is defined and not enable_thinking) else '<think>' -}}
86
+ {%- endif -%}
config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Glm4MoeForCausalLM"
4
+ ],
5
+ "attention_bias": true,
6
+ "attention_dropout": 0.0,
7
+ "dtype": "bfloat16",
8
+ "eos_token_id": [
9
+ 151329,
10
+ 151336,
11
+ 151338
12
+ ],
13
+ "first_k_dense_replace": 3,
14
+ "head_dim": 128,
15
+ "hidden_act": "silu",
16
+ "hidden_size": 5120,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 12288,
19
+ "max_position_embeddings": 202752,
20
+ "model_type": "glm4_moe",
21
+ "moe_intermediate_size": 1536,
22
+ "n_group": 1,
23
+ "n_routed_experts": 112,
24
+ "n_shared_experts": 1,
25
+ "norm_topk_prob": true,
26
+ "num_attention_heads": 96,
27
+ "num_experts_per_tok": 8,
28
+ "num_hidden_layers": 92,
29
+ "num_key_value_heads": 8,
30
+ "num_nextn_predict_layers": 1,
31
+ "pad_token_id": 151329,
32
+ "partial_rotary_factor": 0.5,
33
+ "rms_norm_eps": 1e-05,
34
+ "rope_scaling": null,
35
+ "rope_theta": 1000000,
36
+ "routed_scaling_factor": 2.5,
37
+ "tie_word_embeddings": false,
38
+ "topk_group": 1,
39
+ "transformers_version": "4.57.3",
40
+ "use_cache": true,
41
+ "use_qk_norm": true,
42
+ "vocab_size": 151552
43
+ }
generation_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "eos_token_id": [
4
+ 151329,
5
+ 151336,
6
+ 151338
7
+ ],
8
+ "pad_token_id": 151329,
9
+ "transformers_version": "4.57.3"
10
+ }
model-00001-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a43d3af802b963551e4cc900aa5de17bee95beb16e8f6834e8306cf2305020e0
3
+ size 4986172648
model-00002-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b0c53afa704399e0b37e66f21b008898c5dbaa956c663105e59f0e233ce80c3
3
+ size 4992457518
model-00003-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d0be7d1ed38ffc53a242ceb40c6cbdf85f725923d34718934dcc3a2f2a84670
3
+ size 4882356341
model-00004-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88a3f995df536e59a650dc0a9c3b31776fd82457a6da853484252d3d8f084c36
3
+ size 4986018121
model-00005-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:898b72aee53773a1a154ad417968e15ba87138caa815199283fdf2fe1afef591
3
+ size 4992457505
model-00006-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc3a3b2fd43b54f1e1124e779448616e8873741cf4b37b0efd2290c23e7d2596
3
+ size 4992457518
model-00007-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c70c84d714654ca11ffdabc90c14ee4fb6b9a6da75bcd2c60d81de66b38fc12b
3
+ size 4992457518
model-00008-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c993c74e5e24fe092e20e63d2f8868c6f7da21069d9b64dad09ca0163e15f0be
3
+ size 4992457518
model-00009-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:550a4f87dd09dacc1fa5684dfba29a3fe53419467a776cabce60ea4048bd024c
3
+ size 4992457651
model-00010-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94dd67c4819699e14818591ba7282ea3767b1159a7b01f360ee196f423db1b4d
3
+ size 4992457831
model-00011-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30216766736313fbf7e0dc4c63fce3e86708d746e82f7b62897a70656465cbfd
3
+ size 4992457831
model-00012-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef408d8810f183952adfaa816b2560a096f0abfaa71a5e8d27dc256d7ef727e0
3
+ size 4913814182
model-00013-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96d826bd17aceaef9fb04f6399bad992cac3113a69ae6c0fcdbecc4ca58bc127
3
+ size 4986018433
model-00014-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9b37126d375e81558ad82da3b13b8884871defb3afadf72f670c5c1adebed2c
3
+ size 4992457818
model-00015-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d15dfc5ba0b524aa232010ad8ad93531f0e333a749ffd8c4d353e674aec67324
3
+ size 4992457831
model-00016-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ae0da60e597d319b4bf7eb5964f7691488ac858fd6d9e1173e259f290a12c29
3
+ size 4992457831
model-00017-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b196d3bfe354e7943cd7e4b7df0ebd97c1c2a9655cb1255894fd697f51b19dde
3
+ size 4992457831
model-00018-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e98d254f090756ac038b577c8c74b7e6b84594bb9e4cd67dcfa5670144ccc31
3
+ size 4992457831
model-00019-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ea97e6561ce7846ecd5cc4665bbdc8d7f4e6ca868aba2314d639a0fc2bdde4f
3
+ size 4992457831
model-00020-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:95b0f562da7cfc6c76021276414f031f14d60440c7625e99638d7e77c2defc43
3
+ size 4992457831
model-00021-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea4cc3ed947ed7b6a0fbf44224b5321504a52b00bb546bb2ba733c8bcacb5f06
3
+ size 4913814182
model-00022-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c637c3a293cd337d61e072220bc7b3add0a8f6fbe146cdbd6f06dba482723a8e
3
+ size 4986018433
model-00023-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af57de8b509ac281ca2e0f44dab40c80c599da2d5f0b983627e0eead8419db83
3
+ size 4992457818
model-00024-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6bb20249efd752d731689843d44fdffd67143b6a38c126c913502bfcabd19fa
3
+ size 4992457831
model-00025-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b479481783608d15500e98f661df5ea2a9e76de64b6ef6e7edac06f4fad4351a
3
+ size 4992457831
model-00026-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b09e38c735c105d53b20326b60aeeffa902bd5787eabab08840b78c833b56e72
3
+ size 4992457831
model-00027-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffcd268cfb2627f28591fade32a75ee645a9685d1abad6a56e8f0e59ec4601e7
3
+ size 4992457831
model-00028-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:078a008efc4d2ee6ec7531b44a88134452d492cdfb3e7265896b0fbc71e188c0
3
+ size 4992457831
model-00029-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb889f88bc2926bf2363c23ec9283c0b3df757dcf8f53bf7d2686b3f4cb7793f
3
+ size 4992457831
model-00030-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2918cc7fb0ed106d479430bc3f74d3831b473a86a60078fa31fb5941d515b40
3
+ size 4913814182
model-00031-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ced59fd3cf22cfe026f291af6f434bc70f5b5ce1599f3cb432bb656166b8f1ac
3
+ size 4986018433
model-00032-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:33b23b5a60913dd1d49750b67767573cd636e730e38c2ba9d4c1fb7eb78bedc9
3
+ size 4992457818
model-00033-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4a9d5585840e43775d9643fd34dd5bafbfe04b01bc061e1c1ffdf4301679161e
3
+ size 4992457831
model-00034-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d48b7a22ed9ded39421ebea6eba9dbe0b6cc1c32e9a9083f1cc20e9de7b6e876
3
+ size 4992457831
model-00035-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8dcfd26fb610d11f65cddca971637729413ea3325da47463397d47130865f81a
3
+ size 4992457831
model-00036-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:beacf32349a6c13b4cb1fa2cc2a1f627c8ecb5c4d6b3743dc8d5bc4a74957ac4
3
+ size 4992457831
model-00037-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17e37706aa7186c4ef75826f447090f9bfebaabc41cbdb40753d333b9032475a
3
+ size 4992457831
model-00038-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bb7cc0d0918a2730f66a96fb5e513f054bbcf67d88f41b89b29d5e2de176343
3
+ size 4992457831
model-00039-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ba3f657e339c5da5a15dcc4c6e892eb9f7d4aaff22bbf6c588a8730628019e0
3
+ size 4913814182
model-00040-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efd82016e19376d9743b36dd3693b3956edda3586a0567d726b72bb58380db7c
3
+ size 4986018433
model-00041-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:990ceb748e37b51cd7fb7dcc4081373a883b4cf5a007df2fde176b136006958f
3
+ size 4992457818
model-00042-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a9425cc80346df15ff801c39274de1a7d125c4ca40959d0cef06c6475397ae3f
3
+ size 4992457831
model-00043-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36a65571eabb79a8f2bbadff3764fb0db6a410dafa40d9aae0e9ed6d7c39489f
3
+ size 4992457831
model-00044-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd38ad84194e23c569918051e575762f433a7f15ddd8fb49ac4a2060c9ac98a0
3
+ size 4992457831
model-00045-of-00102.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03d77c81d173adccbe8ca9c86db3cf41527e2e4005ba3398b3857cbe00d50637
3
+ size 4992457831