初始化项目,由ModelHub XC社区提供模型
Model: K-intelligence/Midm-2.0-Base-Instruct Source: Original Platform
This commit is contained in:
49
.gitattributes
vendored
Normal file
49
.gitattributes
vendored
Normal file
@@ -0,0 +1,49 @@
|
||||
*.7z filter=lfs diff=lfs merge=lfs -text
|
||||
*.arrow filter=lfs diff=lfs merge=lfs -text
|
||||
*.bin filter=lfs diff=lfs merge=lfs -text
|
||||
*.bin.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
||||
*.ftz filter=lfs diff=lfs merge=lfs -text
|
||||
*.gz filter=lfs diff=lfs merge=lfs -text
|
||||
*.h5 filter=lfs diff=lfs merge=lfs -text
|
||||
*.joblib filter=lfs diff=lfs merge=lfs -text
|
||||
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.model filter=lfs diff=lfs merge=lfs -text
|
||||
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
||||
*.onnx filter=lfs diff=lfs merge=lfs -text
|
||||
*.ot filter=lfs diff=lfs merge=lfs -text
|
||||
*.parquet filter=lfs diff=lfs merge=lfs -text
|
||||
*.pb filter=lfs diff=lfs merge=lfs -text
|
||||
*.pt filter=lfs diff=lfs merge=lfs -text
|
||||
*.pth filter=lfs diff=lfs merge=lfs -text
|
||||
*.rar filter=lfs diff=lfs merge=lfs -text
|
||||
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
||||
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.tflite filter=lfs diff=lfs merge=lfs -text
|
||||
*.tgz filter=lfs diff=lfs merge=lfs -text
|
||||
*.xz filter=lfs diff=lfs merge=lfs -text
|
||||
*.zip filter=lfs diff=lfs merge=lfs -text
|
||||
*.zstandard filter=lfs diff=lfs merge=lfs -text
|
||||
*.tfevents* filter=lfs diff=lfs merge=lfs -text
|
||||
*.db* filter=lfs diff=lfs merge=lfs -text
|
||||
*.ark* filter=lfs diff=lfs merge=lfs -text
|
||||
**/*ckpt*data* filter=lfs diff=lfs merge=lfs -text
|
||||
**/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text
|
||||
**/*ckpt*.index filter=lfs diff=lfs merge=lfs -text
|
||||
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
||||
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
||||
*.gguf* filter=lfs diff=lfs merge=lfs -text
|
||||
*.ggml filter=lfs diff=lfs merge=lfs -text
|
||||
*.llamafile* filter=lfs diff=lfs merge=lfs -text
|
||||
*.pt2 filter=lfs diff=lfs merge=lfs -text
|
||||
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
||||
*.npy filter=lfs diff=lfs merge=lfs -text
|
||||
*.npz filter=lfs diff=lfs merge=lfs -text
|
||||
*.pickle filter=lfs diff=lfs merge=lfs -text
|
||||
*.pkl filter=lfs diff=lfs merge=lfs -text
|
||||
*.tar filter=lfs diff=lfs merge=lfs -text
|
||||
*.wasm filter=lfs diff=lfs merge=lfs -text
|
||||
*.zst filter=lfs diff=lfs merge=lfs -text
|
||||
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
||||
|
||||
tokenizer.json filter=lfs diff=lfs merge=lfs -text
|
||||
21
LICENSE.txt
Normal file
21
LICENSE.txt
Normal file
@@ -0,0 +1,21 @@
|
||||
MIT License
|
||||
|
||||
Copyright (c) 2025 KT Corporation
|
||||
|
||||
Permission is hereby granted, free of charge, to any person obtaining a copy
|
||||
of this software and associated documentation files (the "Software"), to deal
|
||||
in the Software without restriction, including without limitation the rights
|
||||
to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
|
||||
copies of the Software, and to permit persons to whom the Software is
|
||||
furnished to do so, subject to the following conditions:
|
||||
|
||||
The above copyright notice and this permission notice shall be included in all
|
||||
copies or substantial portions of the Software.
|
||||
|
||||
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
||||
IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
||||
FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
|
||||
AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
||||
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
||||
SOFTWARE.
|
||||
597
README.md
Normal file
597
README.md
Normal file
@@ -0,0 +1,597 @@
|
||||
---
|
||||
license: mit
|
||||
language:
|
||||
- en
|
||||
- ko
|
||||
tags:
|
||||
- KT
|
||||
- K-intelligence
|
||||
- Mi:dm
|
||||
inference: true
|
||||
pipeline_tag: text-generation
|
||||
library_name: transformers
|
||||
---
|
||||
|
||||
|
||||
<p align="center">
|
||||
<br>
|
||||
<span style="font-size: 60px; font-weight: bold;">Mi:dm 2.0 Base</span>
|
||||
</br>
|
||||
</p>
|
||||
|
||||
<p align="center">
|
||||
🤗 <a href="https://huggingface.co/collections/K-intelligence/mi-dm-20-6866406c301e5f45a6926af8">Mi:dm 2.0 Models</a> |
|
||||
📜 <a href="https://github.com/K-intelligence-Midm/Midm-2.0/blob/main/Mi_dm2_0__technical_report.pdf">Mi:dm 2.0 Technical Report</a> |
|
||||
📕 <a href="https://kode.kt.com/blog/article/3935">Mi:dm 2.0 Technical Blog</a>
|
||||
</p>
|
||||
|
||||
|
||||
<br>
|
||||
|
||||
# News 📢
|
||||
|
||||
- 🔧`2025/10/29`: Added support for function calling on vLLM with Mi:dm 2.0 parser.
|
||||
- 📕`2025/08/08`: Published a technical blog article about Mi:dm 2.0 Model.
|
||||
- ⚡️`2025/07/04`: Released Mi:dm 2.0 Model collection on Hugging Face🤗.
|
||||
<br>
|
||||
<br>
|
||||
# Table of Contents
|
||||
|
||||
- ___Overview___
|
||||
- [Mi:dm 2.0](#midm-20)
|
||||
- [Quickstart](#quickstart)
|
||||
- [Evaluation](#evaluation)
|
||||
- ___Usage___
|
||||
- [Run on Friendli.AI](#run-on-friendliai)
|
||||
- [Run on Your Local Machine](#run-on-your-local-machine)
|
||||
- [Deployment](#deployment)
|
||||
- [Tutorials](#tutorials)
|
||||
- ___More Information___
|
||||
- [Limitation](#limitation)
|
||||
- [License](#license)
|
||||
- [Contact](#contact)
|
||||
|
||||
<br>
|
||||
<br>
|
||||
|
||||
# Overview
|
||||
|
||||
## Mi:dm 2.0
|
||||
|
||||
**Mi:dm 2.0** is a __"Korea-centric AI"__ model developed using KT's proprietary technology. The term __"Korea-centric AI"__ refers to a model that deeply internalizes the unique values, cognitive frameworks, and commonsense reasoning inherent to Korean society. It goes beyond simply processing or generating Korean text—it reflects a deeper understanding of the socio-cultural norms and values that define Korean society.
|
||||
|
||||
Mi:dm 2.0 is released in two versions:
|
||||
|
||||
- **Mi:dm 2.0 Base**
|
||||
An 11.5B parameter dense model designed to balance model size and performance.
|
||||
It extends an 8B-scale model by applying the Depth-up Scaling (DuS) method, making it suitable for real-world applications that require both performance and versatility.
|
||||
|
||||
- **Mi:dm 2.0 Mini**
|
||||
A lightweight 2.3B parameter dense model optimized for on-device environments and systems with limited GPU resources.
|
||||
It was derived from the Base model through pruning and distillation to enable compact deployment.
|
||||
|
||||
> [!Note]
|
||||
> Neither the pre-training nor the post-training data includes KT users' data.
|
||||
|
||||
<br>
|
||||
|
||||
## Quickstart
|
||||
|
||||
Here is the code snippet to run conversational inference with the model:
|
||||
|
||||
```python
|
||||
import torch
|
||||
from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig
|
||||
|
||||
model_name = "K-intelligence/Midm-2.0-Base-Instruct"
|
||||
|
||||
model = AutoModelForCausalLM.from_pretrained(
|
||||
model_name,
|
||||
torch_dtype=torch.bfloat16,
|
||||
trust_remote_code=True,
|
||||
device_map="auto"
|
||||
)
|
||||
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
||||
generation_config = GenerationConfig.from_pretrained(model_name)
|
||||
|
||||
prompt = "KT에 대해 소개해줘"
|
||||
|
||||
# message for inference
|
||||
messages = [
|
||||
{"role": "system",
|
||||
"content": "Mi:dm(믿:음)은 KT에서 개발한 AI 기반 어시스턴트이다."},
|
||||
{"role": "user", "content": prompt}
|
||||
]
|
||||
|
||||
input_ids = tokenizer.apply_chat_template(
|
||||
messages,
|
||||
tokenize=True,
|
||||
add_generation_prompt=True,
|
||||
return_tensors="pt"
|
||||
)
|
||||
|
||||
output = model.generate(
|
||||
input_ids.to("cuda"),
|
||||
generation_config=generation_config,
|
||||
eos_token_id=tokenizer.eos_token_id,
|
||||
max_new_tokens=128,
|
||||
do_sample=False,
|
||||
)
|
||||
print(tokenizer.decode(output[0]))
|
||||
```
|
||||
|
||||
> [!NOTE]
|
||||
> The `transformers` library should be version `4.45.0` or higher.
|
||||
|
||||
<br>
|
||||
|
||||
## Evaluation
|
||||
|
||||
|
||||
### Korean
|
||||
|
||||
<!-- first half table-->
|
||||
<table>
|
||||
<tr>
|
||||
<th rowspan="2">Model</th>
|
||||
<th colspan="5" align="center">Society & Culture</th>
|
||||
<th colspan="3" align="center">General Knowledge</th>
|
||||
<th colspan="3" align="center">Instruction Following</th>
|
||||
</tr>
|
||||
<tr>
|
||||
<th align="center">K-Refer<sup>*</sup></th>
|
||||
<th align="center">K-Refer-Hard<sup>*</sup></th>
|
||||
<th align="center">Ko-Sovereign<sup>*</sup></th>
|
||||
<th align="center">HAERAE</th>
|
||||
<th align="center">Avg.</th>
|
||||
<th align="center">KMMLU</th>
|
||||
<th align="center">Ko-Sovereign<sup>*</sup></th>
|
||||
<th align="center">Avg.</th>
|
||||
<th align="center">Ko-IFEval</th>
|
||||
<th align="center">Ko-MTBench</th>
|
||||
<th align="center">Avg.</th>
|
||||
</tr>
|
||||
|
||||
<!-- Small Models -->
|
||||
<tr>
|
||||
<td><strong>Qwen3-4B</strong></td>
|
||||
<td align="center">53.6</td>
|
||||
<td align="center">42.9</td>
|
||||
<td align="center">35.8</td>
|
||||
<td align="center">50.6</td>
|
||||
<td align="center">45.7</td>
|
||||
<td align="center"><strong>50.6</strong></td>
|
||||
<td align="center"><strong>42.5</strong></td>
|
||||
<td align="center"><strong>46.5</strong></td>
|
||||
<td align="center"><strong>75.9</strong></td>
|
||||
<td align="center">63.0</td>
|
||||
<td align="center">69.4</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Exaone-3.5-2.4B-inst</strong></td>
|
||||
<td align="center">64.0</td>
|
||||
<td align="center"><strong>67.1</strong></td>
|
||||
<td align="center"><strong>44.4</strong></td>
|
||||
<td align="center">61.3</td>
|
||||
<td align="center"><strong>59.2</strong></td>
|
||||
<td align="center">43.5</td>
|
||||
<td align="center">42.4</td>
|
||||
<td align="center">43.0</td>
|
||||
<td align="center">65.4</td>
|
||||
<td align="center"><strong>74.0</strong></td>
|
||||
<td align="center">68.9</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Mi:dm 2.0-Mini-inst</strong></td>
|
||||
<td align="center"><strong>66.4</strong></td>
|
||||
<td align="center">61.4</td>
|
||||
<td align="center">36.7</td>
|
||||
<td align="center"><strong>70.8</strong></td>
|
||||
<td align="center">58.8</td>
|
||||
<td align="center">45.1</td>
|
||||
<td align="center">42.4</td>
|
||||
<td align="center">43.8</td>
|
||||
<td align="center">73.3</td>
|
||||
<td align="center"><strong>74.0</strong></td>
|
||||
<td align="center"><strong>73.6</strong></td>
|
||||
</tr>
|
||||
|
||||
<!-- Spacer row -->
|
||||
<tr><td colspan="13"> </td></tr>
|
||||
|
||||
<!-- Large Models -->
|
||||
<tr>
|
||||
<td><strong>Qwen3-14B</strong></td>
|
||||
<td align="center">72.4</td>
|
||||
<td align="center">65.7</td>
|
||||
<td align="center">49.8</td>
|
||||
<td align="center">68.4</td>
|
||||
<td align="center">64.1</td>
|
||||
<td align="center">55.4</td>
|
||||
<td align="center">54.7</td>
|
||||
<td align="center">55.1</td>
|
||||
<td align="center"><strong>83.6</strong></td>
|
||||
<td align="center">71</td>
|
||||
<td align="center">77.3</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Llama-3.1-8B-inst</strong></td>
|
||||
<td align="center">43.2</td>
|
||||
<td align="center">36.4</td>
|
||||
<td align="center">33.8</td>
|
||||
<td align="center">49.5</td>
|
||||
<td align="center">40.7</td>
|
||||
<td align="center">33.0</td>
|
||||
<td align="center">36.7</td>
|
||||
<td align="center">34.8</td>
|
||||
<td align="center">60.1</td>
|
||||
<td align="center">57</td>
|
||||
<td align="center">58.5</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Exaone-3.5-7.8B-inst</strong></td>
|
||||
<td align="center">71.6</td>
|
||||
<td align="center">69.3</td>
|
||||
<td align="center">46.9</td>
|
||||
<td align="center">72.9</td>
|
||||
<td align="center">65.2</td>
|
||||
<td align="center">52.6</td>
|
||||
<td align="center">45.6</td>
|
||||
<td align="center">49.1</td>
|
||||
<td align="center">69.1</td>
|
||||
<td align="center">79.6</td>
|
||||
<td align="center">74.4</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Mi:dm 2.0-Base-inst</strong></td>
|
||||
<td align="center"><strong>89.6</strong></td>
|
||||
<td align="center"><strong>86.4</strong></td>
|
||||
<td align="center"><strong>56.3</strong></td>
|
||||
<td align="center"><strong>81.5</strong></td>
|
||||
<td align="center"><strong>78.4</strong></td>
|
||||
<td align="center"><strong>57.3</strong></td>
|
||||
<td align="center"><strong>58.0</strong></td>
|
||||
<td align="center"><strong>57.7</strong></td>
|
||||
<td align="center">82</td>
|
||||
<td align="center"><strong>89.7</strong></td>
|
||||
<td align="center"><strong>85.9</strong></td>
|
||||
</tr>
|
||||
</table>
|
||||
|
||||
<!-- second half table-->
|
||||
<table>
|
||||
<tr>
|
||||
<th rowspan="2" align="center">Model</th>
|
||||
<th colspan="5" align="center">Comprehension</th>
|
||||
<th colspan="5" align="center">Reasoning</th>
|
||||
</tr>
|
||||
<tr>
|
||||
<th align="center">K-Prag<sup>*</sup></th>
|
||||
<th align="center">K-Refer-Hard<sup>*</sup></th>
|
||||
<th align="center">Ko-Best</th>
|
||||
<th align="center">Ko-Sovereign<sup>*</sup></th>
|
||||
<th align="center">Avg.</th>
|
||||
<th align="center">Ko-Winogrande</th>
|
||||
<th align="center">Ko-Best</th>
|
||||
<th align="center">LogicKor</th>
|
||||
<th align="center">HRM8K</th>
|
||||
<th align="center">Avg.</th>
|
||||
</tr>
|
||||
|
||||
<!-- Small Models -->
|
||||
<tr>
|
||||
<td><strong>Qwen3-4B</strong></td>
|
||||
<td align="center"><strong>73.9<strong></td>
|
||||
<td align="center">56.7</td>
|
||||
<td align="center"><strong>91.5</strong></td>
|
||||
<td align="center"><strong>43.5</strong></td>
|
||||
<td align="center"><strong>66.6</strong></td>
|
||||
<td align="center"><strong>67.5</strong></td>
|
||||
<td align="center"><strong>69.2</strong></td>
|
||||
<td align="center">5.6</td>
|
||||
<td align="center"><strong>56.7</strong></td>
|
||||
<td align="center"><strong>43.8</strong></td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Exaone-3.5-2.4B-inst</strong></td>
|
||||
<td align="center">68.7</td>
|
||||
<td align="center"><strong>58.5</strong></td>
|
||||
<td align="center">87.2</td>
|
||||
<td align="center">38.0</td>
|
||||
<td align="center">62.5</td>
|
||||
<td align="center">60.3</td>
|
||||
<td align="center">64.1</td>
|
||||
<td align="center">7.4</td>
|
||||
<td align="center">38.5</td>
|
||||
<td align="center">36.7</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Mi:dm 2.0-Mini-inst</strong></td>
|
||||
<td align="center">69.5</td>
|
||||
<td align="center">55.4</td>
|
||||
<td align="center">80.5</td>
|
||||
<td align="center">42.5</td>
|
||||
<td align="center">61.9</td>
|
||||
<td align="center">61.7</td>
|
||||
<td align="center">64.5</td>
|
||||
<td align="center"><strong>7.7</strong></td>
|
||||
<td align="center">39.9</td>
|
||||
<td align="center">37.4</td>
|
||||
</tr>
|
||||
|
||||
<!-- Visual Spacer -->
|
||||
<tr><td colspan="11"> </td></tr>
|
||||
|
||||
<!-- Large Models -->
|
||||
<tr>
|
||||
<td><strong>Qwen3-14B</strong></td>
|
||||
<td align="center"><strong>86.7</strong></td>
|
||||
<td align="center"><strong>74.0</strong></td>
|
||||
<td align="center">93.9</td>
|
||||
<td align="center">52.0</td>
|
||||
<td align="center"><strong>76.8</strong></td>
|
||||
<td align="center"><strong>77.2</strong></td>
|
||||
<td align="center"><strong>75.4</strong></td>
|
||||
<td align="center">6.4</td>
|
||||
<td align="center"><strong>64.5</strong></td>
|
||||
<td align="center"><strong>48.8</strong></td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Llama-3.1-8B-inst</strong></td>
|
||||
<td align="center">59.9</td>
|
||||
<td align="center">48.6</td>
|
||||
<td align="center">77.4</td>
|
||||
<td align="center">31.5</td>
|
||||
<td align="center">51.5</td>
|
||||
<td align="center">40.1</td>
|
||||
<td align="center">26.0</td>
|
||||
<td align="center">2.4</td>
|
||||
<td align="center">30.9</td>
|
||||
<td align="center">19.8</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Exaone-3.5-7.8B-inst</strong></td>
|
||||
<td align="center">73.5</td>
|
||||
<td align="center">61.9</td>
|
||||
<td align="center">92.0</td>
|
||||
<td align="center">44.0</td>
|
||||
<td align="center">67.2</td>
|
||||
<td align="center">64.6</td>
|
||||
<td align="center">60.3</td>
|
||||
<td align="center"><strong>8.6</strong></td>
|
||||
<td align="center">49.7</td>
|
||||
<td align="center">39.5</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Mi:dm 2.0-Base-inst</strong></td>
|
||||
<td align="center">86.5</td>
|
||||
<td align="center">70.8</td>
|
||||
<td align="center"><strong>95.2</strong></td>
|
||||
<td align="center"><strong>53.0</strong></td>
|
||||
<td align="center">76.1</td>
|
||||
<td align="center">75.1</td>
|
||||
<td align="center">73.0</td>
|
||||
<td align="center"><strong>8.6</strong></td>
|
||||
<td align="center">52.9</td>
|
||||
<td align="center">44.8</td>
|
||||
</tr>
|
||||
</table>
|
||||
|
||||
`*` indicates KT proprietary evaluation resources.
|
||||
|
||||
<br>
|
||||
|
||||
|
||||
### English
|
||||
|
||||
|
||||
<table>
|
||||
<tr>
|
||||
<th rowspan="2" align="center">Model</th>
|
||||
<th align="center">Instruction</th>
|
||||
<th colspan="4" align="center">Reasoning</th>
|
||||
<th align="center">Math</th>
|
||||
<th align="center">Coding</th>
|
||||
<th colspan="3" align="center">General Knowledge</th>
|
||||
</tr>
|
||||
<tr>
|
||||
<th align="center">IFEval</th>
|
||||
<th align="center">BBH</th>
|
||||
<th align="center">GPQA</th>
|
||||
<th align="center">MuSR</th>
|
||||
<th align="center">Avg.</th>
|
||||
<th align="center">GSM8K</th>
|
||||
<th align="center">MBPP+</th>
|
||||
<th align="center">MMLU-pro</th>
|
||||
<th align="center">MMLU</th>
|
||||
<th align="center">Avg.</th>
|
||||
</tr>
|
||||
|
||||
<!-- Small Models -->
|
||||
<tr>
|
||||
<td><strong>Qwen3-4B</strong></td>
|
||||
<td align="center">79.7</td>
|
||||
<td align="center"><strong>79.0</strong></td>
|
||||
<td align="center"><strong>39.8</strong></td>
|
||||
<td align="center"><strong>58.5</strong></td>
|
||||
<td align="center"><strong>59.1</strong></td>
|
||||
<td align="center"><strong>90.4</strong></td>
|
||||
<td align="center">62.4</td>
|
||||
<td align="center">-</td>
|
||||
<td align="center"><strong>73.3</strong></td>
|
||||
<td align="center"><strong>73.3</strong></td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Exaone-3.5-2.4B-inst</strong></td>
|
||||
<td align="center"><strong>81.1</strong></td>
|
||||
<td align="center">46.4</td>
|
||||
<td align="center">28.1</td>
|
||||
<td align="center">49.7</td>
|
||||
<td align="center">41.4</td>
|
||||
<td align="center">82.5</td>
|
||||
<td align="center">59.8</td>
|
||||
<td align="center">-</td>
|
||||
<td align="center">59.5</td>
|
||||
<td align="center">59.5</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Mi:dm 2.0-Mini-inst</strong></td>
|
||||
<td align="center">73.6</td>
|
||||
<td align="center">44.5</td>
|
||||
<td align="center">26.6</td>
|
||||
<td align="center">51.7</td>
|
||||
<td align="center">40.9</td>
|
||||
<td align="center">83.1</td>
|
||||
<td align="center"><strong>60.9</strong></td>
|
||||
<td align="center">-</td>
|
||||
<td align="center">56.5</td>
|
||||
<td align="center">56.5</td>
|
||||
</tr>
|
||||
|
||||
<tr><td colspan="11"> </td></tr>
|
||||
|
||||
<!-- Large Models -->
|
||||
<tr>
|
||||
<td><strong>Qwen3-14B</strong></td>
|
||||
<td align="center">83.9</td>
|
||||
<td align="center"><strong>83.4</strong></td>
|
||||
<td align="center"><strong>49.8</strong></td>
|
||||
<td align="center"><strong>57.7</strong></td>
|
||||
<td align="center"><strong>63.6</strong></td>
|
||||
<td align="center">88.0</td>
|
||||
<td align="center">73.4</td>
|
||||
<td align="center"><strong>70.5</strong></td>
|
||||
<td align="center"><strong>82.7</strong></td>
|
||||
<td align="center"><strong>76.6</strong></td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Llama-3.1-8B-inst</strong></td>
|
||||
<td align="center">79.9</td>
|
||||
<td align="center">60.3</td>
|
||||
<td align="center">21.6</td>
|
||||
<td align="center">50.3</td>
|
||||
<td align="center">44.1</td>
|
||||
<td align="center">81.2</td>
|
||||
<td align="center"><strong>81.8</strong></td>
|
||||
<td align="center">47.6</td>
|
||||
<td align="center">70.7</td>
|
||||
<td align="center">59.2</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Exaone-3.5-7.8B-inst</strong></td>
|
||||
<td align="center">83.6</td>
|
||||
<td align="center">50.1</td>
|
||||
<td align="center">33.1</td>
|
||||
<td align="center">51.2</td>
|
||||
<td align="center">44.8</td>
|
||||
<td align="center">81.1</td>
|
||||
<td align="center">79.4</td>
|
||||
<td align="center">40.7</td>
|
||||
<td align="center">69.0</td>
|
||||
<td align="center">54.8</td>
|
||||
</tr>
|
||||
<tr>
|
||||
<td><strong>Mi:dm 2.0-Base-inst</strong></td>
|
||||
<td align="center"><strong>84.0</strong></td>
|
||||
<td align="center">77.7</td>
|
||||
<td align="center">33.5</td>
|
||||
<td align="center">51.9</td>
|
||||
<td align="center">54.4</td>
|
||||
<td align="center"><strong>91.6</strong></td>
|
||||
<td align="center">77.5</td>
|
||||
<td align="center">53.3</td>
|
||||
<td align="center">73.7</td>
|
||||
<td align="center">63.5</td>
|
||||
</tr>
|
||||
</table>
|
||||
|
||||
|
||||
<br>
|
||||
|
||||
# Usage
|
||||
|
||||
## Run on Friendli.AI
|
||||
You can try our model immediately via `Friendli.AI`. Simply click `Deploy` and then `Friendli Endpoints`.
|
||||
|
||||
> [!Note]
|
||||
> Please note that a login to `Friendli.AI` is required after your fifth chat interaction.
|
||||
|
||||
<p>
|
||||
<img src="./assets/image_1.png" alt="Left Image" width="36%" style="display:inline-block; margin-right:2%">
|
||||
<img src="./assets/image_2.png" alt="Right Image" width="36%" style="display:inline-block">
|
||||
</p>
|
||||
|
||||
|
||||
## Run on Your Local Machine
|
||||
We provide a detailed description about running Mi:dm 2.0 on your local machine using llama.cpp, LM Studio, and Ollama. Please check our [github](https://github.com/K-intelligence-Midm/Midm-2.0) for more information
|
||||
|
||||
|
||||
## Deployment
|
||||
|
||||
#### Basic Serving
|
||||
|
||||
To serve Mi:dm 2.0 using [vLLM](https://github.com/vllm-project/vllm)(`>=0.8.0`) with an OpenAI-compatible API:
|
||||
```bash
|
||||
vllm serve K-intelligence/Midm-2.0-Base-Instruct
|
||||
```
|
||||
|
||||
#### With Function Calling
|
||||
|
||||
For advanced function calling tasks, you can serve Mi:dm 2.0 with our own tool parser:
|
||||
1. Download and place [Mi:dm 2.0 parser file](https://github.com/K-intelligence-Midm/Midm-2.0/blob/main/tutorial/03_open-webui/modelfile/midm_parser.py) in your working directory.
|
||||
2. Run the following Docker command to launch the vLLM server with our custom parser file:
|
||||
```bash
|
||||
docker run --rm -it --gpus all -p 8000:8000 \
|
||||
-e HUGGING_FACE_HUB_TOKEN="<YOUR_HUGGINGFACE_TOKEN>" \
|
||||
-v "$(pwd)/midm_parser.py:/custom/midm_parser.py" \
|
||||
vllm/vllm-openai:v0.11.0 \
|
||||
--model K-intelligence/Midm-2.0-Base-Instruct \
|
||||
--enable-auto-tool-choice \
|
||||
--tool-parser-plugin /custom/midm_parser.py \
|
||||
--tool-call-parser midm-parser \
|
||||
--host 0.0.0.0
|
||||
```
|
||||
|
||||
>[!Note]
|
||||
> This setup is compatible with `vllm/vllm-openai:v0.8.0` and later, but we strongly recommend using `v0.11.0` for optimal stability and compatibility with our parser.
|
||||
|
||||
|
||||
## Tutorials
|
||||
To help our end-users easily use Mi:dm 2.0, we have provided comprehensive tutorials on [github](https://github.com/K-intelligence-Midm/Midm-2.0).
|
||||
<br>
|
||||
|
||||
<br>
|
||||
<br>
|
||||
|
||||
# More Information
|
||||
|
||||
## Limitation
|
||||
* The training data for both Mi:dm 2.0 models consists primarily of English and Korean. Understanding and generation in other languages are not guaranteed.
|
||||
|
||||
* The model is not guaranteed to provide reliable advice in fields that require professional expertise, such as law, medicine, or finance.
|
||||
|
||||
* Researchers have made efforts to exclude unethical content from the training data — such as profanity, slurs, bias, and discriminatory language. However, despite these efforts, the model may still produce inappropriate expressions or factual inaccuracies.
|
||||
|
||||
|
||||
## License
|
||||
|
||||
Mi:dm 2.0 is licensed under the [MIT License](./LICENSE).
|
||||
|
||||
<!-- ### Citation
|
||||
|
||||
```
|
||||
@misc{,
|
||||
title={},
|
||||
author={},
|
||||
year={2025},
|
||||
eprint={},
|
||||
archivePrefix={arXiv},
|
||||
primaryClass={cs.CL},
|
||||
url={},
|
||||
}
|
||||
``` -->
|
||||
## Contact
|
||||
Mi:dm 2.0 Technical Inquiries: midm-llm@kt.com
|
||||
|
||||
<br>
|
||||
|
||||
BIN
assets/image_1.png
Normal file
BIN
assets/image_1.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 66 KiB |
BIN
assets/image_2.png
Normal file
BIN
assets/image_2.png
Normal file
Binary file not shown.
|
After Width: | Height: | Size: 126 KiB |
31
config.json
Normal file
31
config.json
Normal file
@@ -0,0 +1,31 @@
|
||||
{
|
||||
"architectures": [
|
||||
"LlamaForCausalLM"
|
||||
],
|
||||
"attention_bias": false,
|
||||
"attention_dropout": 0.0,
|
||||
"attention_probs_dropout_prob": 0.1,
|
||||
"bos_token_id": 0,
|
||||
"eos_token_id": 2,
|
||||
"head_dim": 128,
|
||||
"hidden_act": "silu",
|
||||
"hidden_dropout_prob": 0.0,
|
||||
"hidden_size": 4096,
|
||||
"initializer_range": 0.02,
|
||||
"intermediate_size": 14336,
|
||||
"max_position_embeddings": 32768,
|
||||
"mlp_bias": false,
|
||||
"model_type": "llama",
|
||||
"num_attention_heads": 32,
|
||||
"num_hidden_layers": 48,
|
||||
"num_key_value_heads": 8,
|
||||
"pretraining_tp": 1,
|
||||
"rms_norm_eps": 1e-05,
|
||||
"rope_scaling": null,
|
||||
"rope_theta": 8000000,
|
||||
"tie_word_embeddings": false,
|
||||
"torch_dtype": "bfloat16",
|
||||
"transformers_version": "4.51.3",
|
||||
"use_cache": true,
|
||||
"vocab_size": 131384
|
||||
}
|
||||
1
configuration.json
Normal file
1
configuration.json
Normal file
@@ -0,0 +1 @@
|
||||
{"framework": "pytorch", "task": "text-generation", "allow_remote": true}
|
||||
10
generation_config.json
Normal file
10
generation_config.json
Normal file
@@ -0,0 +1,10 @@
|
||||
{
|
||||
"bos_token_id": 0,
|
||||
"do_sample": true,
|
||||
"eos_token_id": 2,
|
||||
"repetition_penalty": 1.05,
|
||||
"temperature": 0.8,
|
||||
"top_k": 20,
|
||||
"top_p": 0.7,
|
||||
"transformers_version": "4.51.3"
|
||||
}
|
||||
3
model-00001-of-00005.safetensors
Normal file
3
model-00001-of-00005.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:16a89c6f9a06f178441f63c914ed33b4deda73f1407563ee9e63b2c6ecb01973
|
||||
size 4884866008
|
||||
3
model-00002-of-00005.safetensors
Normal file
3
model-00002-of-00005.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:e8f01a498141a6d202a9323346f21d607489dcad22cc7a96838def6a955555ea
|
||||
size 4999819328
|
||||
3
model-00003-of-00005.safetensors
Normal file
3
model-00003-of-00005.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:2cd98cbe3301443a4512c4057b866e55ae04cc06649b1f1895eaf0c9ea93a38c
|
||||
size 4915916184
|
||||
3
model-00004-of-00005.safetensors
Normal file
3
model-00004-of-00005.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:e7d8d7e6fabbbbf4fed8c3d1b76a4bdad8b46cc05f7472da2cf0e299225229e6
|
||||
size 4915916176
|
||||
3
model-00005-of-00005.safetensors
Normal file
3
model-00005-of-00005.safetensors
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:2b06333f3782bbd8cbd02156b5f945ecde78fe63778e7aa519c4cbbb4b103520
|
||||
size 3374888640
|
||||
442
model.safetensors.index.json
Normal file
442
model.safetensors.index.json
Normal file
@@ -0,0 +1,442 @@
|
||||
{
|
||||
"metadata": {
|
||||
"total_size": 23091355648
|
||||
},
|
||||
"weight_map": {
|
||||
"lm_head.weight": "model-00005-of-00005.safetensors",
|
||||
"model.embed_tokens.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.0.input_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.0.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.0.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.0.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.0.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.0.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.0.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.0.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.0.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.1.input_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.1.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.1.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.1.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.1.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.1.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.1.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.1.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.1.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.10.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.10.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.10.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.10.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.10.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.10.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.10.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.10.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.10.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.11.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.11.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.11.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.11.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.11.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.11.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.11.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.11.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.11.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.12.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.12.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.12.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.12.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.12.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.12.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.12.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.12.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.12.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.13.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.13.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.13.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.13.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.13.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.13.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.13.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.13.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.13.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.14.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.14.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.14.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.14.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.14.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.14.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.14.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.14.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.14.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.15.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.15.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.15.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.15.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.15.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.15.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.15.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.15.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.15.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.16.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.16.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.16.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.16.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.16.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.16.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.16.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.16.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.16.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.17.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.17.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.17.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.17.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.17.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.17.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.17.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.17.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.17.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.18.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.18.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.18.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.18.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.18.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.18.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.18.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.18.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.18.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.19.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.19.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.19.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.19.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.19.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.19.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.19.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.19.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.19.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.2.input_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.2.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.2.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.2.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.2.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.2.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.2.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.2.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.20.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.20.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.20.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.20.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.20.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.20.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.20.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.20.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.20.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.21.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.21.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.21.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.21.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.21.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.21.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.21.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.21.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.21.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.22.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.22.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.22.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.22.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.22.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.22.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.22.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.22.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.22.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.23.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.23.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.23.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.23.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.23.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.23.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.23.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.23.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.23.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.24.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.24.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.24.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.24.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.24.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.24.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.24.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.24.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.24.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.25.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.25.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.25.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.25.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.25.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.25.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.25.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.25.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.25.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.26.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.26.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.26.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.26.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.26.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.26.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.26.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.26.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.26.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.27.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.27.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.27.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.27.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.27.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.27.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.27.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.27.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.27.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.28.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.28.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.28.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.28.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.28.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.28.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.28.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.28.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.28.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.29.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.29.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.29.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.29.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.29.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.29.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.29.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.29.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.29.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.3.input_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.3.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.3.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.3.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.3.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.3.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.3.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.3.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.3.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.30.input_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.30.mlp.down_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.30.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.30.mlp.up_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.30.post_attention_layernorm.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.30.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.30.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.30.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.30.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.31.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.31.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.31.mlp.gate_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.31.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.31.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.31.self_attn.k_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.31.self_attn.o_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.31.self_attn.q_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.31.self_attn.v_proj.weight": "model-00003-of-00005.safetensors",
|
||||
"model.layers.32.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.32.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.32.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.32.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.32.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.32.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.32.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.32.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.32.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.33.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.33.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.33.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.33.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.33.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.33.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.33.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.33.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.33.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.34.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.34.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.34.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.34.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.34.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.34.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.34.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.34.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.34.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.35.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.35.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.35.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.35.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.35.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.35.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.35.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.35.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.35.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.36.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.36.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.36.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.36.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.36.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.36.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.36.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.36.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.36.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.37.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.37.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.37.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.37.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.37.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.37.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.37.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.37.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.37.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.38.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.38.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.38.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.38.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.38.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.38.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.38.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.38.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.38.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.39.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.39.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.39.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.39.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.39.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.39.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.39.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.39.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.39.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.4.input_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.4.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.4.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.4.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.4.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.4.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.4.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.4.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.4.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.40.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.40.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.40.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.40.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.40.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.40.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.40.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.40.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.40.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.41.input_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.41.mlp.down_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.41.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.41.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.41.post_attention_layernorm.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.41.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.41.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.41.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.41.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.42.input_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.42.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.42.mlp.gate_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.42.mlp.up_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.42.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.42.self_attn.k_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.42.self_attn.o_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.42.self_attn.q_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.42.self_attn.v_proj.weight": "model-00004-of-00005.safetensors",
|
||||
"model.layers.43.input_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.43.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.43.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.43.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.43.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.43.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.43.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.43.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.43.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.44.input_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.44.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.44.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.44.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.44.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.44.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.44.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.44.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.44.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.45.input_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.45.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.45.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.45.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.45.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.45.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.45.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.45.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.45.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.46.input_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.46.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.46.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.46.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.46.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.46.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.46.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.46.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.46.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.47.input_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.47.mlp.down_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.47.mlp.gate_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.47.mlp.up_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.47.post_attention_layernorm.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.47.self_attn.k_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.47.self_attn.o_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.47.self_attn.q_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.47.self_attn.v_proj.weight": "model-00005-of-00005.safetensors",
|
||||
"model.layers.5.input_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.5.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.5.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.5.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.5.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.5.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.5.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.5.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.5.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.6.input_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.6.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.6.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.6.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.6.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.6.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.6.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.6.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.6.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.7.input_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.7.mlp.down_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.7.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.7.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.7.post_attention_layernorm.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.7.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.7.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.7.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.7.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.8.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.8.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.8.mlp.gate_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.8.mlp.up_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.8.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.8.self_attn.k_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.8.self_attn.o_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.8.self_attn.q_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.8.self_attn.v_proj.weight": "model-00001-of-00005.safetensors",
|
||||
"model.layers.9.input_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.9.mlp.down_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.9.mlp.gate_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.9.mlp.up_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.9.post_attention_layernorm.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.9.self_attn.k_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.9.self_attn.o_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.9.self_attn.q_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.layers.9.self_attn.v_proj.weight": "model-00002-of-00005.safetensors",
|
||||
"model.norm.weight": "model-00005-of-00005.safetensors"
|
||||
}
|
||||
}
|
||||
23
special_tokens_map.json
Normal file
23
special_tokens_map.json
Normal file
@@ -0,0 +1,23 @@
|
||||
{
|
||||
"bos_token": {
|
||||
"content": "<|begin_of_text|>",
|
||||
"lstrip": false,
|
||||
"normalized": false,
|
||||
"rstrip": false,
|
||||
"single_word": false
|
||||
},
|
||||
"eos_token": {
|
||||
"content": "<|end_of_text|>",
|
||||
"lstrip": false,
|
||||
"normalized": false,
|
||||
"rstrip": false,
|
||||
"single_word": false
|
||||
},
|
||||
"pad_token": {
|
||||
"content": "<|end_of_text|>",
|
||||
"lstrip": false,
|
||||
"normalized": false,
|
||||
"rstrip": false,
|
||||
"single_word": false
|
||||
}
|
||||
}
|
||||
3
tokenizer.json
Normal file
3
tokenizer.json
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:56af6093020d6923882ebc787e4c22b74ac26bce41cbdc0e7952dd70a91baad7
|
||||
size 10398090
|
||||
706
tokenizer_config.json
Normal file
706
tokenizer_config.json
Normal file
File diff suppressed because one or more lines are too long
Reference in New Issue
Block a user