commit 62d150aeb635dad57e9720924470894520817d4e Author: ModelHub XC Date: Sat Apr 11 08:53:03 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: louisbrulenaudet/Pearl-3x7B Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..968b8ba --- /dev/null +++ b/.gitattributes @@ -0,0 +1,69 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bin.* filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text + +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zstandard filter=lfs diff=lfs merge=lfs -text +*.tfevents* filter=lfs diff=lfs merge=lfs -text +*.db* filter=lfs diff=lfs merge=lfs -text +*.ark* filter=lfs diff=lfs merge=lfs -text +**/*ckpt*data* filter=lfs diff=lfs merge=lfs -text +**/*ckpt*.meta filter=lfs diff=lfs merge=lfs -text +**/*ckpt*.index filter=lfs diff=lfs merge=lfs -text + +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.gguf* filter=lfs diff=lfs merge=lfs -text +*.ggml filter=lfs diff=lfs merge=lfs -text +*.llamafile* filter=lfs diff=lfs merge=lfs -text +*.pt2 filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text + +model-00012-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00014-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00004-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +tokenizer.json filter=lfs diff=lfs merge=lfs -text +model-00016-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00015-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00006-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00009-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00003-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00005-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00010-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00008-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00018-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00013-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00001-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00017-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00002-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00007-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00019-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +model-00011-of-00019.safetensors filter=lfs diff=lfs merge=lfs -text +tokenizer.model filter=lfs diff=lfs merge=lfs -text \ No newline at end of file diff --git a/README.md b/README.md new file mode 100644 index 0000000..45f8dbe --- /dev/null +++ b/README.md @@ -0,0 +1,146 @@ +--- +license: apache-2.0 +tags: +- moe +- frankenmoe +- merge +- mergekit +- lazymergekit +- dvilasuero/DistilabelBeagle14-7B +- beowolx/CodeNinja-1.0-OpenChat-7B +- WizardLM/WizardMath-7B-V1.1 +- Maths +- Code +- Python +base_model: +- dvilasuero/DistilabelBeagle14-7B +- beowolx/CodeNinja-1.0-OpenChat-7B +- WizardLM/WizardMath-7B-V1.1 +language: +- en +library_name: transformers +pipeline_tag: text-generation +--- + +
+ +# Pearl-3x7B, an xtraordinary Mixture of Experts (MoE) for data science + +Pearl-3x7B is a Mixture of Experts (MoE) made with the following models : +* [dvilasuero/DistilabelBeagle14-7B](https://huggingface.co/dvilasuero/DistilabelBeagle14-7B) +* [beowolx/CodeNinja-1.0-OpenChat-7B](https://huggingface.co/beowolx/CodeNinja-1.0-OpenChat-7B) +* [WizardLM/WizardMath-7B-V1.1](https://huggingface.co/WizardLM/WizardMath-7B-V1.1) + +A Mixture of Experts (MoE) model represents a sophisticated architecture that amalgamates the capabilities of multiple specialized models to address a wide array of tasks within a unified framework. Within the realm of a MoE model tailored for a chat application, the integration of expertise spanning three distinct domains - chat, code, and mathematics - substantially enhances its capacity to furnish nuanced and precise responses to a diverse spectrum of user inquiries. + +The initial expert model, honed for chat applications, exhibits prowess in comprehending natural language nuances, conversational dynamics, and contextual cues. Drawing upon extensive conversational data, it adeptly generates engaging and contextually pertinent responses, thereby fostering meaningful interactions with users. + +The subsequent expert model, centered on code, brings to the fore proficiency in programming languages, algorithms, and software engineering principles. Possessing a deep-seated understanding of syntax, logical constructs, and problem-solving methodologies, it deftly tackles queries spanning coding challenges, debugging assistance, and software development inquiries. + +Lastly, the third expert model, specializing in mathematics, boasts expertise in mathematical reasoning, problem-solving strategies, and analytical techniques. Armed with a breadth of knowledge encompassing arithmetic, algebra, calculus, and beyond, it offers precise solutions, lucid explanations, and profound insights for mathematical queries, equations, and proofs. + +## Configuration + +```yaml +base_model: argilla/CapybaraHermes-2.5-Mistral-7B +experts: + - source_model: dvilasuero/DistilabelBeagle14-7B + positive_prompts: + - "chat" + - "assistant" + - "tell me" + - "explain" + - "help" + - "guide" + - "assist" + - "answer" + - "support" + - "clarify" + - "elaborate" + - "educate" + - "inform" + - "advise" + - "instruct" + - source_model: beowolx/CodeNinja-1.0-OpenChat-7B + positive_prompts: + - "code" + - "python" + - "javascript" + - "programming" + - "algorithm" + - "develop" + - "debug" + - "optimize" + - "software" + - "engineer" + - "web" + - "application" + - "framework" + - "library" + - "syntax" + - "logic" + - "compile" + - "execute" + - source_model: WizardLM/WizardMath-7B-V1.1 + positive_prompts: + - "reason" + - "math" + - "mathematics" + - "solve" + - "count" + - "calculate" + - "analyze" + - "derive" + - "compute" + - "numbers" + - "equation" + - "theorem" + - "proof" + - "geometry" + - "trigonometry" + - "statistics" + - "probability" + - "algebra" + - "integral" +``` + +## Usage + +```python +!pip install -qU transformers bitsandbytes accelerate + +from transformers import AutoTokenizer +import transformers +import torch + +model = "louisbrulenaudet/Pearl-3x7B" + +tokenizer = AutoTokenizer.from_pretrained(model) +pipeline = transformers.pipeline( + "text-generation", + model=model, + model_kwargs={"torch_dtype": torch.float16, "load_in_4bit": True}, +) + +messages = [{"role": "user", "content": "Explain what a Mixture of Experts is in less than 100 words."}] +prompt = pipeline.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) +outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) +print(outputs[0]["generated_text"]) +``` + +## Citing & Authors + +If you use this code in your research, please use the following BibTeX entry. + +```BibTeX +@misc{louisbrulenaudet2023, + author = {Louis Brulé Naudet}, + title = {Pearl-3x7B, an xtraordinary Mixture of Experts (MoE) for data science}, + year = {2023} + howpublished = {\url{https://huggingface.co/louisbrulenaudet/Pearl-3x7B}}, +} +``` + +## Feedback + +If you have any feedback, please reach out at [louisbrulenaudet@icloud.com](mailto:louisbrulenaudet@icloud.com). \ No newline at end of file diff --git a/added_tokens.json b/added_tokens.json new file mode 100644 index 0000000..e36863d --- /dev/null +++ b/added_tokens.json @@ -0,0 +1,4 @@ +{ + "<|im_end|>": 32000, + "<|im_start|>": 32001 +} diff --git a/config.json b/config.json new file mode 100644 index 0000000..a5f461c --- /dev/null +++ b/config.json @@ -0,0 +1,30 @@ +{ + "_name_or_path": "argilla/CapybaraHermes-2.5-Mistral-7B", + "architectures": [ + "MixtralForCausalLM" + ], + "attention_dropout": 0.0, + "bos_token_id": 1, + "eos_token_id": 32000, + "hidden_act": "silu", + "hidden_size": 4096, + "initializer_range": 0.02, + "intermediate_size": 14336, + "max_position_embeddings": 32768, + "model_type": "mixtral", + "num_attention_heads": 32, + "num_experts_per_tok": 2, + "num_hidden_layers": 32, + "num_key_value_heads": 8, + "num_local_experts": 3, + "output_router_logits": false, + "rms_norm_eps": 1e-05, + "rope_theta": 10000.0, + "router_aux_loss_coef": 0.001, + "sliding_window": null, + "tie_word_embeddings": false, + "torch_dtype": "float16", + "transformers_version": "4.37.2", + "use_cache": false, + "vocab_size": 32002 +} diff --git a/configuration.json b/configuration.json new file mode 100644 index 0000000..bbeeda1 --- /dev/null +++ b/configuration.json @@ -0,0 +1 @@ +{"framework": "pytorch", "task": "text-generation", "allow_remote": true} \ No newline at end of file diff --git a/mergekit_moe_config.yml b/mergekit_moe_config.yml new file mode 100644 index 0000000..2a9c319 --- /dev/null +++ b/mergekit_moe_config.yml @@ -0,0 +1,61 @@ + +base_model: argilla/CapybaraHermes-2.5-Mistral-7B +experts: + - source_model: dvilasuero/DistilabelBeagle14-7B + positive_prompts: + - "chat" + - "assistant" + - "tell me" + - "explain" + - "help" + - "guide" + - "assist" + - "answer" + - "support" + - "clarify" + - "elaborate" + - "educate" + - "inform" + - "advise" + - "instruct" + - source_model: beowolx/CodeNinja-1.0-OpenChat-7B + positive_prompts: + - "code" + - "python" + - "javascript" + - "programming" + - "algorithm" + - "develop" + - "debug" + - "optimize" + - "software" + - "engineer" + - "web" + - "application" + - "framework" + - "library" + - "syntax" + - "logic" + - "compile" + - "execute" + - source_model: WizardLM/WizardMath-7B-V1.1 + positive_prompts: + - "reason" + - "math" + - "mathematics" + - "solve" + - "count" + - "calculate" + - "analyze" + - "derive" + - "compute" + - "numbers" + - "equation" + - "theorem" + - "proof" + - "geometry" + - "trigonometry" + - "statistics" + - "probability" + - "algebra" + - "integral" diff --git a/model-00001-of-00019.safetensors b/model-00001-of-00019.safetensors new file mode 100644 index 0000000..16d8aae --- /dev/null +++ b/model-00001-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9fd98130c520239927ba4fb3eec2ac5db398482fa64085bc73e0b53e5710c620 +size 1933882656 diff --git a/model-00002-of-00019.safetensors b/model-00002-of-00019.safetensors new file mode 100644 index 0000000..18913cb --- /dev/null +++ b/model-00002-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f582751fd1e3f380a5261210a9bd303604065d6d38f10d468805ddc755986521 +size 1996490936 diff --git a/model-00003-of-00019.safetensors b/model-00003-of-00019.safetensors new file mode 100644 index 0000000..79e09d6 --- /dev/null +++ b/model-00003-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e072b7a97d69997a61aeaec63f7b8bc1470cd68a3bd1d1c786889b45413b75a +size 1996490952 diff --git a/model-00004-of-00019.safetensors b/model-00004-of-00019.safetensors new file mode 100644 index 0000000..d3cdb9f --- /dev/null +++ b/model-00004-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5225a7fe639cccbd1aff1d85be606583d5c7fc363e8d062a1c79a5ea25cac7c9 +size 1996490952 diff --git a/model-00005-of-00019.safetensors b/model-00005-of-00019.safetensors new file mode 100644 index 0000000..e18ab75 --- /dev/null +++ b/model-00005-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2f308145891966075878f1b5f37cd8c3e5a193f4a82b09ef384e2310fa7b46a7 +size 1996490952 diff --git a/model-00006-of-00019.safetensors b/model-00006-of-00019.safetensors new file mode 100644 index 0000000..55a9881 --- /dev/null +++ b/model-00006-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:609237bb2ba0bdc1a0c2528b681bbc1d463f3de3d04c687177f154f928b46a2a +size 1996490952 diff --git a/model-00007-of-00019.safetensors b/model-00007-of-00019.safetensors new file mode 100644 index 0000000..64f4874 --- /dev/null +++ b/model-00007-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:96d41d9c04c32231e3f431f9aeabb79a1bae6c64380c94ac4021cd77dbc17fbd +size 1996490936 diff --git a/model-00008-of-00019.safetensors b/model-00008-of-00019.safetensors new file mode 100644 index 0000000..5448659 --- /dev/null +++ b/model-00008-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1442748cdd7cc9bc39bd068f5119f9481800008ed34402a10b083a7ab6796e54 +size 1996490936 diff --git a/model-00009-of-00019.safetensors b/model-00009-of-00019.safetensors new file mode 100644 index 0000000..39da7d4 --- /dev/null +++ b/model-00009-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8a2e9de2c2509a207edd82115af511f122c7abfdf0758501867b1118e6391d41 +size 1996490952 diff --git a/model-00010-of-00019.safetensors b/model-00010-of-00019.safetensors new file mode 100644 index 0000000..03b253b --- /dev/null +++ b/model-00010-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7e2ab43dac0f23b30731f0b36ed7feceabe0a692baef03c5c57fb9dc431043de +size 1996490952 diff --git a/model-00011-of-00019.safetensors b/model-00011-of-00019.safetensors new file mode 100644 index 0000000..da0ca3f --- /dev/null +++ b/model-00011-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad78496cc40362f784b50e7962d43d251ab2585ac7d29a05512863671d72e9f0 +size 1996490952 diff --git a/model-00012-of-00019.safetensors b/model-00012-of-00019.safetensors new file mode 100644 index 0000000..3159f9d --- /dev/null +++ b/model-00012-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:baed25bf28aa66c718aaa6b541b62a49abb09099bc0569a634c05a14a188453d +size 1996490944 diff --git a/model-00013-of-00019.safetensors b/model-00013-of-00019.safetensors new file mode 100644 index 0000000..1e3e9a7 --- /dev/null +++ b/model-00013-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b86739adf7f1c0d8e2e3d686d4e76b1d7aca41087b4656b958643b8b8beba8f4 +size 1996490936 diff --git a/model-00014-of-00019.safetensors b/model-00014-of-00019.safetensors new file mode 100644 index 0000000..2bd1ead --- /dev/null +++ b/model-00014-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b346100083a795825011b0330ff289055e23cf4fcd75d3ad2011b9ac5d4922d1 +size 1996490944 diff --git a/model-00015-of-00019.safetensors b/model-00015-of-00019.safetensors new file mode 100644 index 0000000..37e08c0 --- /dev/null +++ b/model-00015-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9281a5a7cd4d0578a482bfb54e86a49881f582994fe57b2329b96e46305abb3f +size 1996490952 diff --git a/model-00016-of-00019.safetensors b/model-00016-of-00019.safetensors new file mode 100644 index 0000000..2f865f8 --- /dev/null +++ b/model-00016-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bb67656a258659da37da4c06c887208396967780d4cacd2bd41302a3a49ff91e +size 1996490952 diff --git a/model-00017-of-00019.safetensors b/model-00017-of-00019.safetensors new file mode 100644 index 0000000..d1ff075 --- /dev/null +++ b/model-00017-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2dafa2241d8db920255bb5be263a4bf6752ff6e913d31cbb17b0246e48e8c20a +size 1996490952 diff --git a/model-00018-of-00019.safetensors b/model-00018-of-00019.safetensors new file mode 100644 index 0000000..7ce1dbe --- /dev/null +++ b/model-00018-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eeb423411a57d2058bfe0270425cfbde22939f7c06dc1c2d8fdfc29dd85a29a1 +size 1996765088 diff --git a/model-00019-of-00019.safetensors b/model-00019-of-00019.safetensors new file mode 100644 index 0000000..b68294e --- /dev/null +++ b/model-00019-of-00019.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62ccf1aff3a6c89117a4fbc0096db353e3389a01bf38bc0bdb9693f41ccbc70e +size 1158422928 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000..39f1381 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1 @@ +{"metadata": {"mergekit_version": "0.0.4"}, "weight_map": {"model.embed_tokens.weight": "model-00001-of-00019.safetensors", "model.norm.weight": "model-00001-of-00019.safetensors", "lm_head.weight": "model-00001-of-00019.safetensors", "model.layers.0.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.1.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.2.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.3.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.4.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.5.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.6.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.7.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.8.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.9.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.10.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.11.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.12.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.13.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.14.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.15.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.16.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.17.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.18.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.19.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.20.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.21.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.22.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.23.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.24.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.25.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.26.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.27.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.28.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.29.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.30.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.31.input_layernorm.weight": "model-00001-of-00019.safetensors", "model.layers.0.block_sparse_moe.experts.0.w3.weight": "model-00001-of-00019.safetensors", "model.layers.0.block_sparse_moe.experts.1.w3.weight": "model-00001-of-00019.safetensors", "model.layers.0.block_sparse_moe.experts.2.w3.weight": "model-00001-of-00019.safetensors", "model.layers.1.block_sparse_moe.experts.0.w3.weight": "model-00001-of-00019.safetensors", "model.layers.1.block_sparse_moe.experts.1.w3.weight": "model-00001-of-00019.safetensors", "model.layers.1.block_sparse_moe.experts.2.w3.weight": "model-00001-of-00019.safetensors", "model.layers.2.block_sparse_moe.experts.0.w3.weight": "model-00001-of-00019.safetensors", "model.layers.2.block_sparse_moe.experts.1.w3.weight": "model-00001-of-00019.safetensors", "model.layers.2.block_sparse_moe.experts.2.w3.weight": "model-00001-of-00019.safetensors", "model.layers.3.block_sparse_moe.experts.0.w3.weight": "model-00001-of-00019.safetensors", "model.layers.3.block_sparse_moe.experts.1.w3.weight": "model-00001-of-00019.safetensors", "model.layers.3.block_sparse_moe.experts.2.w3.weight": "model-00001-of-00019.safetensors", "model.layers.4.block_sparse_moe.experts.0.w3.weight": "model-00002-of-00019.safetensors", "model.layers.4.block_sparse_moe.experts.1.w3.weight": "model-00002-of-00019.safetensors", "model.layers.4.block_sparse_moe.experts.2.w3.weight": "model-00002-of-00019.safetensors", "model.layers.5.block_sparse_moe.experts.0.w3.weight": "model-00002-of-00019.safetensors", "model.layers.5.block_sparse_moe.experts.1.w3.weight": "model-00002-of-00019.safetensors", "model.layers.5.block_sparse_moe.experts.2.w3.weight": "model-00002-of-00019.safetensors", "model.layers.6.block_sparse_moe.experts.0.w3.weight": "model-00002-of-00019.safetensors", "model.layers.6.block_sparse_moe.experts.1.w3.weight": "model-00002-of-00019.safetensors", "model.layers.6.block_sparse_moe.experts.2.w3.weight": "model-00002-of-00019.safetensors", "model.layers.7.block_sparse_moe.experts.0.w3.weight": "model-00002-of-00019.safetensors", "model.layers.7.block_sparse_moe.experts.1.w3.weight": "model-00002-of-00019.safetensors", "model.layers.7.block_sparse_moe.experts.2.w3.weight": "model-00002-of-00019.safetensors", "model.layers.8.block_sparse_moe.experts.0.w3.weight": "model-00002-of-00019.safetensors", "model.layers.8.block_sparse_moe.experts.1.w3.weight": "model-00002-of-00019.safetensors", "model.layers.8.block_sparse_moe.experts.2.w3.weight": "model-00002-of-00019.safetensors", "model.layers.9.block_sparse_moe.experts.0.w3.weight": "model-00002-of-00019.safetensors", "model.layers.9.block_sparse_moe.experts.1.w3.weight": "model-00002-of-00019.safetensors", "model.layers.9.block_sparse_moe.experts.2.w3.weight": "model-00003-of-00019.safetensors", "model.layers.10.block_sparse_moe.experts.0.w3.weight": "model-00003-of-00019.safetensors", "model.layers.10.block_sparse_moe.experts.1.w3.weight": "model-00003-of-00019.safetensors", "model.layers.10.block_sparse_moe.experts.2.w3.weight": "model-00003-of-00019.safetensors", "model.layers.11.block_sparse_moe.experts.0.w3.weight": "model-00003-of-00019.safetensors", "model.layers.11.block_sparse_moe.experts.1.w3.weight": "model-00003-of-00019.safetensors", "model.layers.11.block_sparse_moe.experts.2.w3.weight": "model-00003-of-00019.safetensors", "model.layers.12.block_sparse_moe.experts.0.w3.weight": "model-00003-of-00019.safetensors", "model.layers.12.block_sparse_moe.experts.1.w3.weight": "model-00003-of-00019.safetensors", "model.layers.12.block_sparse_moe.experts.2.w3.weight": "model-00003-of-00019.safetensors", "model.layers.13.block_sparse_moe.experts.0.w3.weight": "model-00003-of-00019.safetensors", "model.layers.13.block_sparse_moe.experts.1.w3.weight": "model-00003-of-00019.safetensors", "model.layers.13.block_sparse_moe.experts.2.w3.weight": "model-00003-of-00019.safetensors", "model.layers.14.block_sparse_moe.experts.0.w3.weight": "model-00003-of-00019.safetensors", "model.layers.14.block_sparse_moe.experts.1.w3.weight": "model-00003-of-00019.safetensors", "model.layers.14.block_sparse_moe.experts.2.w3.weight": "model-00003-of-00019.safetensors", "model.layers.15.block_sparse_moe.experts.0.w3.weight": "model-00003-of-00019.safetensors", "model.layers.15.block_sparse_moe.experts.1.w3.weight": "model-00004-of-00019.safetensors", "model.layers.15.block_sparse_moe.experts.2.w3.weight": "model-00004-of-00019.safetensors", "model.layers.16.block_sparse_moe.experts.0.w3.weight": "model-00004-of-00019.safetensors", "model.layers.16.block_sparse_moe.experts.1.w3.weight": "model-00004-of-00019.safetensors", "model.layers.16.block_sparse_moe.experts.2.w3.weight": "model-00004-of-00019.safetensors", "model.layers.17.block_sparse_moe.experts.0.w3.weight": "model-00004-of-00019.safetensors", "model.layers.17.block_sparse_moe.experts.1.w3.weight": "model-00004-of-00019.safetensors", "model.layers.17.block_sparse_moe.experts.2.w3.weight": "model-00004-of-00019.safetensors", "model.layers.18.block_sparse_moe.experts.0.w3.weight": "model-00004-of-00019.safetensors", "model.layers.18.block_sparse_moe.experts.1.w3.weight": "model-00004-of-00019.safetensors", "model.layers.18.block_sparse_moe.experts.2.w3.weight": "model-00004-of-00019.safetensors", "model.layers.19.block_sparse_moe.experts.0.w3.weight": "model-00004-of-00019.safetensors", "model.layers.19.block_sparse_moe.experts.1.w3.weight": "model-00004-of-00019.safetensors", "model.layers.19.block_sparse_moe.experts.2.w3.weight": "model-00004-of-00019.safetensors", "model.layers.20.block_sparse_moe.experts.0.w3.weight": "model-00004-of-00019.safetensors", "model.layers.20.block_sparse_moe.experts.1.w3.weight": "model-00004-of-00019.safetensors", "model.layers.20.block_sparse_moe.experts.2.w3.weight": "model-00004-of-00019.safetensors", "model.layers.21.block_sparse_moe.experts.0.w3.weight": "model-00005-of-00019.safetensors", "model.layers.21.block_sparse_moe.experts.1.w3.weight": "model-00005-of-00019.safetensors", "model.layers.21.block_sparse_moe.experts.2.w3.weight": "model-00005-of-00019.safetensors", "model.layers.22.block_sparse_moe.experts.0.w3.weight": "model-00005-of-00019.safetensors", "model.layers.22.block_sparse_moe.experts.1.w3.weight": "model-00005-of-00019.safetensors", "model.layers.22.block_sparse_moe.experts.2.w3.weight": "model-00005-of-00019.safetensors", "model.layers.23.block_sparse_moe.experts.0.w3.weight": "model-00005-of-00019.safetensors", "model.layers.23.block_sparse_moe.experts.1.w3.weight": "model-00005-of-00019.safetensors", "model.layers.23.block_sparse_moe.experts.2.w3.weight": "model-00005-of-00019.safetensors", "model.layers.24.block_sparse_moe.experts.0.w3.weight": "model-00005-of-00019.safetensors", "model.layers.24.block_sparse_moe.experts.1.w3.weight": "model-00005-of-00019.safetensors", "model.layers.24.block_sparse_moe.experts.2.w3.weight": "model-00005-of-00019.safetensors", "model.layers.25.block_sparse_moe.experts.0.w3.weight": "model-00005-of-00019.safetensors", "model.layers.25.block_sparse_moe.experts.1.w3.weight": "model-00005-of-00019.safetensors", "model.layers.25.block_sparse_moe.experts.2.w3.weight": "model-00005-of-00019.safetensors", "model.layers.26.block_sparse_moe.experts.0.w3.weight": "model-00005-of-00019.safetensors", "model.layers.26.block_sparse_moe.experts.1.w3.weight": "model-00005-of-00019.safetensors", "model.layers.26.block_sparse_moe.experts.2.w3.weight": "model-00006-of-00019.safetensors", "model.layers.27.block_sparse_moe.experts.0.w3.weight": "model-00006-of-00019.safetensors", "model.layers.27.block_sparse_moe.experts.1.w3.weight": "model-00006-of-00019.safetensors", "model.layers.27.block_sparse_moe.experts.2.w3.weight": "model-00006-of-00019.safetensors", "model.layers.28.block_sparse_moe.experts.0.w3.weight": "model-00006-of-00019.safetensors", "model.layers.28.block_sparse_moe.experts.1.w3.weight": "model-00006-of-00019.safetensors", "model.layers.28.block_sparse_moe.experts.2.w3.weight": "model-00006-of-00019.safetensors", "model.layers.29.block_sparse_moe.experts.0.w3.weight": "model-00006-of-00019.safetensors", "model.layers.29.block_sparse_moe.experts.1.w3.weight": "model-00006-of-00019.safetensors", "model.layers.29.block_sparse_moe.experts.2.w3.weight": "model-00006-of-00019.safetensors", "model.layers.30.block_sparse_moe.experts.0.w3.weight": "model-00006-of-00019.safetensors", "model.layers.30.block_sparse_moe.experts.1.w3.weight": "model-00006-of-00019.safetensors", "model.layers.30.block_sparse_moe.experts.2.w3.weight": "model-00006-of-00019.safetensors", "model.layers.31.block_sparse_moe.experts.0.w3.weight": "model-00006-of-00019.safetensors", "model.layers.31.block_sparse_moe.experts.1.w3.weight": "model-00006-of-00019.safetensors", "model.layers.31.block_sparse_moe.experts.2.w3.weight": "model-00006-of-00019.safetensors", "model.layers.0.block_sparse_moe.experts.0.w2.weight": "model-00006-of-00019.safetensors", "model.layers.0.block_sparse_moe.experts.1.w2.weight": "model-00007-of-00019.safetensors", "model.layers.0.block_sparse_moe.experts.2.w2.weight": "model-00007-of-00019.safetensors", "model.layers.1.block_sparse_moe.experts.0.w2.weight": "model-00007-of-00019.safetensors", "model.layers.1.block_sparse_moe.experts.1.w2.weight": "model-00007-of-00019.safetensors", "model.layers.1.block_sparse_moe.experts.2.w2.weight": "model-00007-of-00019.safetensors", "model.layers.2.block_sparse_moe.experts.0.w2.weight": "model-00007-of-00019.safetensors", "model.layers.2.block_sparse_moe.experts.1.w2.weight": "model-00007-of-00019.safetensors", "model.layers.2.block_sparse_moe.experts.2.w2.weight": "model-00007-of-00019.safetensors", "model.layers.3.block_sparse_moe.experts.0.w2.weight": "model-00007-of-00019.safetensors", "model.layers.3.block_sparse_moe.experts.1.w2.weight": "model-00007-of-00019.safetensors", "model.layers.3.block_sparse_moe.experts.2.w2.weight": "model-00007-of-00019.safetensors", "model.layers.4.block_sparse_moe.experts.0.w2.weight": "model-00007-of-00019.safetensors", "model.layers.4.block_sparse_moe.experts.1.w2.weight": "model-00007-of-00019.safetensors", "model.layers.4.block_sparse_moe.experts.2.w2.weight": "model-00007-of-00019.safetensors", "model.layers.5.block_sparse_moe.experts.0.w2.weight": "model-00007-of-00019.safetensors", "model.layers.5.block_sparse_moe.experts.1.w2.weight": "model-00007-of-00019.safetensors", "model.layers.5.block_sparse_moe.experts.2.w2.weight": "model-00007-of-00019.safetensors", "model.layers.6.block_sparse_moe.experts.0.w2.weight": "model-00008-of-00019.safetensors", "model.layers.6.block_sparse_moe.experts.1.w2.weight": "model-00008-of-00019.safetensors", "model.layers.6.block_sparse_moe.experts.2.w2.weight": "model-00008-of-00019.safetensors", "model.layers.7.block_sparse_moe.experts.0.w2.weight": "model-00008-of-00019.safetensors", "model.layers.7.block_sparse_moe.experts.1.w2.weight": "model-00008-of-00019.safetensors", "model.layers.7.block_sparse_moe.experts.2.w2.weight": "model-00008-of-00019.safetensors", "model.layers.8.block_sparse_moe.experts.0.w2.weight": "model-00008-of-00019.safetensors", "model.layers.8.block_sparse_moe.experts.1.w2.weight": "model-00008-of-00019.safetensors", "model.layers.8.block_sparse_moe.experts.2.w2.weight": "model-00008-of-00019.safetensors", "model.layers.9.block_sparse_moe.experts.0.w2.weight": "model-00008-of-00019.safetensors", "model.layers.9.block_sparse_moe.experts.1.w2.weight": "model-00008-of-00019.safetensors", "model.layers.9.block_sparse_moe.experts.2.w2.weight": "model-00008-of-00019.safetensors", "model.layers.10.block_sparse_moe.experts.0.w2.weight": "model-00008-of-00019.safetensors", "model.layers.10.block_sparse_moe.experts.1.w2.weight": "model-00008-of-00019.safetensors", "model.layers.10.block_sparse_moe.experts.2.w2.weight": "model-00008-of-00019.safetensors", "model.layers.11.block_sparse_moe.experts.0.w2.weight": "model-00008-of-00019.safetensors", "model.layers.11.block_sparse_moe.experts.1.w2.weight": "model-00008-of-00019.safetensors", "model.layers.11.block_sparse_moe.experts.2.w2.weight": "model-00009-of-00019.safetensors", "model.layers.12.block_sparse_moe.experts.0.w2.weight": "model-00009-of-00019.safetensors", "model.layers.12.block_sparse_moe.experts.1.w2.weight": "model-00009-of-00019.safetensors", "model.layers.12.block_sparse_moe.experts.2.w2.weight": "model-00009-of-00019.safetensors", "model.layers.13.block_sparse_moe.experts.0.w2.weight": "model-00009-of-00019.safetensors", "model.layers.13.block_sparse_moe.experts.1.w2.weight": "model-00009-of-00019.safetensors", "model.layers.13.block_sparse_moe.experts.2.w2.weight": "model-00009-of-00019.safetensors", "model.layers.14.block_sparse_moe.experts.0.w2.weight": "model-00009-of-00019.safetensors", "model.layers.14.block_sparse_moe.experts.1.w2.weight": "model-00009-of-00019.safetensors", "model.layers.14.block_sparse_moe.experts.2.w2.weight": "model-00009-of-00019.safetensors", "model.layers.15.block_sparse_moe.experts.0.w2.weight": "model-00009-of-00019.safetensors", "model.layers.15.block_sparse_moe.experts.1.w2.weight": "model-00009-of-00019.safetensors", "model.layers.15.block_sparse_moe.experts.2.w2.weight": "model-00009-of-00019.safetensors", "model.layers.16.block_sparse_moe.experts.0.w2.weight": "model-00009-of-00019.safetensors", "model.layers.16.block_sparse_moe.experts.1.w2.weight": "model-00009-of-00019.safetensors", "model.layers.16.block_sparse_moe.experts.2.w2.weight": "model-00009-of-00019.safetensors", "model.layers.17.block_sparse_moe.experts.0.w2.weight": "model-00009-of-00019.safetensors", "model.layers.17.block_sparse_moe.experts.1.w2.weight": "model-00010-of-00019.safetensors", "model.layers.17.block_sparse_moe.experts.2.w2.weight": "model-00010-of-00019.safetensors", "model.layers.18.block_sparse_moe.experts.0.w2.weight": "model-00010-of-00019.safetensors", "model.layers.18.block_sparse_moe.experts.1.w2.weight": "model-00010-of-00019.safetensors", "model.layers.18.block_sparse_moe.experts.2.w2.weight": "model-00010-of-00019.safetensors", "model.layers.19.block_sparse_moe.experts.0.w2.weight": "model-00010-of-00019.safetensors", "model.layers.19.block_sparse_moe.experts.1.w2.weight": "model-00010-of-00019.safetensors", "model.layers.19.block_sparse_moe.experts.2.w2.weight": "model-00010-of-00019.safetensors", "model.layers.20.block_sparse_moe.experts.0.w2.weight": "model-00010-of-00019.safetensors", "model.layers.20.block_sparse_moe.experts.1.w2.weight": "model-00010-of-00019.safetensors", "model.layers.20.block_sparse_moe.experts.2.w2.weight": "model-00010-of-00019.safetensors", "model.layers.21.block_sparse_moe.experts.0.w2.weight": "model-00010-of-00019.safetensors", "model.layers.21.block_sparse_moe.experts.1.w2.weight": "model-00010-of-00019.safetensors", "model.layers.21.block_sparse_moe.experts.2.w2.weight": "model-00010-of-00019.safetensors", "model.layers.22.block_sparse_moe.experts.0.w2.weight": "model-00010-of-00019.safetensors", "model.layers.22.block_sparse_moe.experts.1.w2.weight": "model-00010-of-00019.safetensors", "model.layers.22.block_sparse_moe.experts.2.w2.weight": "model-00010-of-00019.safetensors", "model.layers.23.block_sparse_moe.experts.0.w2.weight": "model-00011-of-00019.safetensors", "model.layers.23.block_sparse_moe.experts.1.w2.weight": "model-00011-of-00019.safetensors", "model.layers.23.block_sparse_moe.experts.2.w2.weight": "model-00011-of-00019.safetensors", "model.layers.24.block_sparse_moe.experts.0.w2.weight": "model-00011-of-00019.safetensors", "model.layers.24.block_sparse_moe.experts.1.w2.weight": "model-00011-of-00019.safetensors", "model.layers.24.block_sparse_moe.experts.2.w2.weight": "model-00011-of-00019.safetensors", "model.layers.25.block_sparse_moe.experts.0.w2.weight": "model-00011-of-00019.safetensors", "model.layers.25.block_sparse_moe.experts.1.w2.weight": "model-00011-of-00019.safetensors", "model.layers.25.block_sparse_moe.experts.2.w2.weight": "model-00011-of-00019.safetensors", "model.layers.26.block_sparse_moe.experts.0.w2.weight": "model-00011-of-00019.safetensors", "model.layers.26.block_sparse_moe.experts.1.w2.weight": "model-00011-of-00019.safetensors", "model.layers.26.block_sparse_moe.experts.2.w2.weight": "model-00011-of-00019.safetensors", "model.layers.27.block_sparse_moe.experts.0.w2.weight": "model-00011-of-00019.safetensors", "model.layers.27.block_sparse_moe.experts.1.w2.weight": "model-00011-of-00019.safetensors", "model.layers.27.block_sparse_moe.experts.2.w2.weight": "model-00011-of-00019.safetensors", "model.layers.28.block_sparse_moe.experts.0.w2.weight": "model-00011-of-00019.safetensors", "model.layers.28.block_sparse_moe.experts.1.w2.weight": "model-00011-of-00019.safetensors", "model.layers.28.block_sparse_moe.experts.2.w2.weight": "model-00012-of-00019.safetensors", "model.layers.29.block_sparse_moe.experts.0.w2.weight": "model-00012-of-00019.safetensors", "model.layers.29.block_sparse_moe.experts.1.w2.weight": "model-00012-of-00019.safetensors", "model.layers.29.block_sparse_moe.experts.2.w2.weight": "model-00012-of-00019.safetensors", "model.layers.30.block_sparse_moe.experts.0.w2.weight": "model-00012-of-00019.safetensors", "model.layers.30.block_sparse_moe.experts.1.w2.weight": "model-00012-of-00019.safetensors", "model.layers.30.block_sparse_moe.experts.2.w2.weight": "model-00012-of-00019.safetensors", "model.layers.31.block_sparse_moe.experts.0.w2.weight": "model-00012-of-00019.safetensors", "model.layers.31.block_sparse_moe.experts.1.w2.weight": "model-00012-of-00019.safetensors", "model.layers.31.block_sparse_moe.experts.2.w2.weight": "model-00012-of-00019.safetensors", "model.layers.0.block_sparse_moe.experts.0.w1.weight": "model-00012-of-00019.safetensors", "model.layers.0.block_sparse_moe.experts.1.w1.weight": "model-00012-of-00019.safetensors", "model.layers.0.block_sparse_moe.experts.2.w1.weight": "model-00012-of-00019.safetensors", "model.layers.1.block_sparse_moe.experts.0.w1.weight": "model-00012-of-00019.safetensors", "model.layers.1.block_sparse_moe.experts.1.w1.weight": "model-00012-of-00019.safetensors", "model.layers.1.block_sparse_moe.experts.2.w1.weight": "model-00012-of-00019.safetensors", "model.layers.2.block_sparse_moe.experts.0.w1.weight": "model-00012-of-00019.safetensors", "model.layers.2.block_sparse_moe.experts.1.w1.weight": "model-00013-of-00019.safetensors", "model.layers.2.block_sparse_moe.experts.2.w1.weight": "model-00013-of-00019.safetensors", "model.layers.3.block_sparse_moe.experts.0.w1.weight": "model-00013-of-00019.safetensors", "model.layers.3.block_sparse_moe.experts.1.w1.weight": "model-00013-of-00019.safetensors", "model.layers.3.block_sparse_moe.experts.2.w1.weight": "model-00013-of-00019.safetensors", "model.layers.4.block_sparse_moe.experts.0.w1.weight": "model-00013-of-00019.safetensors", "model.layers.4.block_sparse_moe.experts.1.w1.weight": "model-00013-of-00019.safetensors", "model.layers.4.block_sparse_moe.experts.2.w1.weight": "model-00013-of-00019.safetensors", "model.layers.5.block_sparse_moe.experts.0.w1.weight": "model-00013-of-00019.safetensors", "model.layers.5.block_sparse_moe.experts.1.w1.weight": "model-00013-of-00019.safetensors", "model.layers.5.block_sparse_moe.experts.2.w1.weight": "model-00013-of-00019.safetensors", "model.layers.6.block_sparse_moe.experts.0.w1.weight": "model-00013-of-00019.safetensors", "model.layers.6.block_sparse_moe.experts.1.w1.weight": "model-00013-of-00019.safetensors", "model.layers.6.block_sparse_moe.experts.2.w1.weight": "model-00013-of-00019.safetensors", "model.layers.7.block_sparse_moe.experts.0.w1.weight": "model-00013-of-00019.safetensors", "model.layers.7.block_sparse_moe.experts.1.w1.weight": "model-00013-of-00019.safetensors", "model.layers.7.block_sparse_moe.experts.2.w1.weight": "model-00013-of-00019.safetensors", "model.layers.8.block_sparse_moe.experts.0.w1.weight": "model-00014-of-00019.safetensors", "model.layers.8.block_sparse_moe.experts.1.w1.weight": "model-00014-of-00019.safetensors", "model.layers.8.block_sparse_moe.experts.2.w1.weight": "model-00014-of-00019.safetensors", "model.layers.9.block_sparse_moe.experts.0.w1.weight": "model-00014-of-00019.safetensors", "model.layers.9.block_sparse_moe.experts.1.w1.weight": "model-00014-of-00019.safetensors", "model.layers.9.block_sparse_moe.experts.2.w1.weight": "model-00014-of-00019.safetensors", "model.layers.10.block_sparse_moe.experts.0.w1.weight": "model-00014-of-00019.safetensors", "model.layers.10.block_sparse_moe.experts.1.w1.weight": "model-00014-of-00019.safetensors", "model.layers.10.block_sparse_moe.experts.2.w1.weight": "model-00014-of-00019.safetensors", "model.layers.11.block_sparse_moe.experts.0.w1.weight": "model-00014-of-00019.safetensors", "model.layers.11.block_sparse_moe.experts.1.w1.weight": "model-00014-of-00019.safetensors", "model.layers.11.block_sparse_moe.experts.2.w1.weight": "model-00014-of-00019.safetensors", "model.layers.12.block_sparse_moe.experts.0.w1.weight": "model-00014-of-00019.safetensors", "model.layers.12.block_sparse_moe.experts.1.w1.weight": "model-00014-of-00019.safetensors", "model.layers.12.block_sparse_moe.experts.2.w1.weight": "model-00014-of-00019.safetensors", "model.layers.13.block_sparse_moe.experts.0.w1.weight": "model-00014-of-00019.safetensors", "model.layers.13.block_sparse_moe.experts.1.w1.weight": "model-00014-of-00019.safetensors", "model.layers.13.block_sparse_moe.experts.2.w1.weight": "model-00015-of-00019.safetensors", "model.layers.14.block_sparse_moe.experts.0.w1.weight": "model-00015-of-00019.safetensors", "model.layers.14.block_sparse_moe.experts.1.w1.weight": "model-00015-of-00019.safetensors", "model.layers.14.block_sparse_moe.experts.2.w1.weight": "model-00015-of-00019.safetensors", "model.layers.15.block_sparse_moe.experts.0.w1.weight": "model-00015-of-00019.safetensors", "model.layers.15.block_sparse_moe.experts.1.w1.weight": "model-00015-of-00019.safetensors", "model.layers.15.block_sparse_moe.experts.2.w1.weight": "model-00015-of-00019.safetensors", "model.layers.16.block_sparse_moe.experts.0.w1.weight": "model-00015-of-00019.safetensors", "model.layers.16.block_sparse_moe.experts.1.w1.weight": "model-00015-of-00019.safetensors", "model.layers.16.block_sparse_moe.experts.2.w1.weight": "model-00015-of-00019.safetensors", "model.layers.17.block_sparse_moe.experts.0.w1.weight": "model-00015-of-00019.safetensors", "model.layers.17.block_sparse_moe.experts.1.w1.weight": "model-00015-of-00019.safetensors", "model.layers.17.block_sparse_moe.experts.2.w1.weight": "model-00015-of-00019.safetensors", "model.layers.18.block_sparse_moe.experts.0.w1.weight": "model-00015-of-00019.safetensors", "model.layers.18.block_sparse_moe.experts.1.w1.weight": "model-00015-of-00019.safetensors", "model.layers.18.block_sparse_moe.experts.2.w1.weight": "model-00015-of-00019.safetensors", "model.layers.19.block_sparse_moe.experts.0.w1.weight": "model-00015-of-00019.safetensors", "model.layers.19.block_sparse_moe.experts.1.w1.weight": "model-00016-of-00019.safetensors", "model.layers.19.block_sparse_moe.experts.2.w1.weight": "model-00016-of-00019.safetensors", "model.layers.20.block_sparse_moe.experts.0.w1.weight": "model-00016-of-00019.safetensors", "model.layers.20.block_sparse_moe.experts.1.w1.weight": "model-00016-of-00019.safetensors", "model.layers.20.block_sparse_moe.experts.2.w1.weight": "model-00016-of-00019.safetensors", "model.layers.21.block_sparse_moe.experts.0.w1.weight": "model-00016-of-00019.safetensors", "model.layers.21.block_sparse_moe.experts.1.w1.weight": "model-00016-of-00019.safetensors", "model.layers.21.block_sparse_moe.experts.2.w1.weight": "model-00016-of-00019.safetensors", "model.layers.22.block_sparse_moe.experts.0.w1.weight": "model-00016-of-00019.safetensors", "model.layers.22.block_sparse_moe.experts.1.w1.weight": "model-00016-of-00019.safetensors", "model.layers.22.block_sparse_moe.experts.2.w1.weight": "model-00016-of-00019.safetensors", "model.layers.23.block_sparse_moe.experts.0.w1.weight": "model-00016-of-00019.safetensors", "model.layers.23.block_sparse_moe.experts.1.w1.weight": "model-00016-of-00019.safetensors", "model.layers.23.block_sparse_moe.experts.2.w1.weight": "model-00016-of-00019.safetensors", "model.layers.24.block_sparse_moe.experts.0.w1.weight": "model-00016-of-00019.safetensors", "model.layers.24.block_sparse_moe.experts.1.w1.weight": "model-00016-of-00019.safetensors", "model.layers.24.block_sparse_moe.experts.2.w1.weight": "model-00016-of-00019.safetensors", "model.layers.25.block_sparse_moe.experts.0.w1.weight": "model-00017-of-00019.safetensors", "model.layers.25.block_sparse_moe.experts.1.w1.weight": "model-00017-of-00019.safetensors", "model.layers.25.block_sparse_moe.experts.2.w1.weight": "model-00017-of-00019.safetensors", "model.layers.26.block_sparse_moe.experts.0.w1.weight": "model-00017-of-00019.safetensors", "model.layers.26.block_sparse_moe.experts.1.w1.weight": "model-00017-of-00019.safetensors", "model.layers.26.block_sparse_moe.experts.2.w1.weight": "model-00017-of-00019.safetensors", "model.layers.27.block_sparse_moe.experts.0.w1.weight": "model-00017-of-00019.safetensors", "model.layers.27.block_sparse_moe.experts.1.w1.weight": "model-00017-of-00019.safetensors", "model.layers.27.block_sparse_moe.experts.2.w1.weight": "model-00017-of-00019.safetensors", "model.layers.28.block_sparse_moe.experts.0.w1.weight": "model-00017-of-00019.safetensors", "model.layers.28.block_sparse_moe.experts.1.w1.weight": "model-00017-of-00019.safetensors", "model.layers.28.block_sparse_moe.experts.2.w1.weight": "model-00017-of-00019.safetensors", "model.layers.29.block_sparse_moe.experts.0.w1.weight": "model-00017-of-00019.safetensors", "model.layers.29.block_sparse_moe.experts.1.w1.weight": "model-00017-of-00019.safetensors", "model.layers.29.block_sparse_moe.experts.2.w1.weight": "model-00017-of-00019.safetensors", "model.layers.30.block_sparse_moe.experts.0.w1.weight": "model-00017-of-00019.safetensors", "model.layers.30.block_sparse_moe.experts.1.w1.weight": "model-00017-of-00019.safetensors", "model.layers.30.block_sparse_moe.experts.2.w1.weight": "model-00018-of-00019.safetensors", "model.layers.31.block_sparse_moe.experts.0.w1.weight": "model-00018-of-00019.safetensors", "model.layers.31.block_sparse_moe.experts.1.w1.weight": "model-00018-of-00019.safetensors", "model.layers.31.block_sparse_moe.experts.2.w1.weight": "model-00018-of-00019.safetensors", "model.layers.0.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.1.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.2.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.3.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.4.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.5.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.6.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.7.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.8.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.9.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.10.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.11.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.12.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.13.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.14.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.15.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.16.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.17.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.18.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.19.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.20.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.21.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.22.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.23.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.24.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.25.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.26.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.27.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.28.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.29.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.30.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.31.post_attention_layernorm.weight": "model-00018-of-00019.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00018-of-00019.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.15.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00018-of-00019.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00018-of-00019.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00019-of-00019.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00019-of-00019.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00019.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00019-of-00019.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00019-of-00019.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00019-of-00019.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00019-of-00019.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00019-of-00019.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00019-of-00019.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00019-of-00019.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00019-of-00019.safetensors", "model.layers.0.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.1.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.2.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.3.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.4.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.5.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.6.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.7.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.8.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.9.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.10.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.11.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.12.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.13.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.14.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.15.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.16.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.17.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.18.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.19.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.20.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.21.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.22.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.23.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.24.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.25.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.26.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.27.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.28.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.29.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.30.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors", "model.layers.31.block_sparse_moe.gate.weight": "model-00019-of-00019.safetensors"}} \ No newline at end of file diff --git a/special_tokens_map.json b/special_tokens_map.json new file mode 100644 index 0000000..1576bc1 --- /dev/null +++ b/special_tokens_map.json @@ -0,0 +1,24 @@ +{ + "bos_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "<|im_end|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": "", + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/tokenizer.json b/tokenizer.json new file mode 100644 index 0000000..b5a36eb --- /dev/null +++ b/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c7cf57dadebd4785143c06f48ffeec2ec498062ef26c5c527152805cbaac14d8 +size 1795829 diff --git a/tokenizer.model b/tokenizer.model new file mode 100644 index 0000000..8b443ef --- /dev/null +++ b/tokenizer.model @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055 +size 493443 diff --git a/tokenizer_config.json b/tokenizer_config.json new file mode 100644 index 0000000..87ffc94 --- /dev/null +++ b/tokenizer_config.json @@ -0,0 +1,61 @@ +{ + "add_bos_token": true, + "add_eos_token": false, + "added_tokens_decoder": { + "0": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "1": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "2": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "32000": { + "content": "<|im_end|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "32001": { + "content": "<|im_start|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [], + "bos_token": "", + "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "legacy": true, + "model_max_length": 1000000000000000019884624838656, + "pad_token": "", + "sp_model_kwargs": {}, + "spaces_between_special_tokens": false, + "tokenizer_class": "LlamaTokenizer", + "trust_remote_code": false, + "unk_token": "", + "use_default_system_prompt": true, + "use_fast": true +}