From fb83a80c251058c896ac1242b26dea2300ff5ff0 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Tue, 5 May 2026 14:31:42 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: hongzhouyu/FineMedLM-o1 Source: Original Platform --- .gitattributes | 5 + README.md | 107 + config.json | 35 + generation_config.json | 9 + model-00001-of-00004.safetensors | 3 + model-00002-of-00004.safetensors | 3 + model-00003-of-00004.safetensors | 3 + model-00004-of-00004.safetensors | 3 + model.safetensors.index.json | 298 +++ ...941.dpo0--958cfdb21e51-doqv5oixhk.556397.0 | Bin 0 -> 23847 bytes special_tokens_map.json | 23 + tokenizer.json | 3 + tokenizer_config.json | 2064 +++++++++++++++++ trainer_state.json | 409 ++++ training_args.bin | Bin 0 -> 7800 bytes 15 files changed, 2965 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 config.json create mode 100644 generation_config.json create mode 100644 model-00001-of-00004.safetensors create mode 100644 model-00002-of-00004.safetensors create mode 100644 model-00003-of-00004.safetensors create mode 100644 model-00004-of-00004.safetensors create mode 100644 model.safetensors.index.json create mode 100644 runs/May29_08-03-21_dpo0--958cfdb21e51-doqv5oixhk/events.out.tfevents.1748505941.dpo0--958cfdb21e51-doqv5oixhk.556397.0 create mode 100644 special_tokens_map.json create mode 100644 tokenizer.json create mode 100644 tokenizer_config.json create mode 100644 trainer_state.json create mode 100644 training_args.bin diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..51a5b10 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,5 @@ +model-00001-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text +model-00002-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text +model-00003-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text +model-00004-of-00004.safetensors filter=lfs diff=lfs merge=lfs -text +tokenizer.json filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..09fbccb --- /dev/null +++ b/README.md @@ -0,0 +1,107 @@ +--- +license: mit +datasets: +- hongzhouyu/FineMed-SFT +- hongzhouyu/FineMed-DPO +language: +- en +- zh +base_model: +- meta-llama/Llama-3.1-8B +- hongzhouyu/FineMedLM +library_name: transformers +tags: +- medical +--- + +
+

+ FineMedLM-o1 +

+
+ +
+GitHub | Paper +
+ +# Introduction +**FineMedLM-o1** is a specialized medical LLM engineered for advanced medical reasoning. It employs a multi-step reasoning process, iteratively reflecting on and refining its thought process before delivering a final response. + +For more information, visit our GitHub repository. + +# Usage +You can use FineMedLM-o1 in the same way as `Llama-3.1-8B-Instruct`: + +(⚠️**Note**: Please use the system prompt we provide to achieve better inference results.) +```python +from transformers import AutoModelForCausalLM, AutoTokenizer + +main_model_name = "yuhongzhou/FineMedLM" +model = AutoModelForCausalLM.from_pretrained(main_model_name, device_map="auto") +tokenizer = AutoTokenizer.from_pretrained(main_model_name) + +prompt = ( + """The following are multiple choice questions (with answers) about health. Think step by step and then finish your answer with "the answer is (X)" where X is the correct letter choice. + + +Question: +Polio can be eradicated by which of the following? +Options: +A. Herbal remedies +B. Use of antibiotics +C. Regular intake of vitamins +D. Administration of tetanus vaccine +E. Attention to sewage control and hygiene +F. Natural immunity acquired through exposure +G. Use of antiviral drugs +Answer: Let's think step by step. +""" +) + +messages = [ + {"role": "system", "content": """You are a helpful professional doctor. You need to generate an answer based on the given problem and thoroughly explore the problem through a systematic and long-term thinking process to provide a final and accurate solution. This requires a comprehensive cycle of analysis, summary, exploration, re-evaluation, reflection, backtracking and iteration to form a thoughtful thinking process. Use the background information provided in the text to assist in formulating the answer. Follow these answer guidelines: +1. Please structure your response into two main sections: **Thought** and **Summarization**. +2. During the **Thought** phase, think step by step based on the given text content. If the text content is used, it must be expressed. +3. During the **Summarization** phase, based on the thinking process in the thinking phase, give the final answer to the question. +Here is the question: """}, + {"role": "user", "content": prompt} +] + +text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) +print(text) + +model_inputs = tokenizer(text, return_tensors="pt").to(model.device) + +print("-----start generate-----") +generated_ids = model.generate( + model_inputs.input_ids, + max_new_tokens=2048, + eos_token_id=tokenizer.eos_token_id +) + +answer = tokenizer.decode(generated_ids[0], skip_special_tokens=False) +print(answer) +``` + +FineMedLM-o1 adopts a *slow-thinking* approach, with outputs formatted as: + +``` +**Thought** +[Reasoning process] + +**Summarization** +[Output] +``` + +# Citation +``` +@misc{yu2025finemedlmo1enhancingmedicalreasoning, + title={FineMedLM-o1: Enhancing the Medical Reasoning Ability of LLM from Supervised Fine-Tuning to Test-Time Training}, + author={Hongzhou Yu and Tianhao Cheng and Ying Cheng and Rui Feng}, + year={2025}, + eprint={2501.09213}, + archivePrefix={arXiv}, + primaryClass={cs.CL}, + url={https://arxiv.org/abs/2501.09213}, +} +``` \ No newline at end of file diff --git a/config.json b/config.json new file mode 100644 index 0000000..a153282 --- /dev/null +++ b/config.json @@ -0,0 +1,35 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 128000, + "eos_token_id": 128001, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 4096, + "initializer_range": 0.02, + "intermediate_size": 14336, + "max_position_embeddings": 131072, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 32, + "num_hidden_layers": 32, + "num_key_value_heads": 8, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": { + "factor": 8.0, + "high_freq_factor": 4.0, + "low_freq_factor": 1.0, + "original_max_position_embeddings": 8192, + "rope_type": "llama3" + }, + "rope_theta": 500000.0, + "tie_word_embeddings": false, + "torch_dtype": "bfloat16", + "transformers_version": "4.50.0", + "use_cache": true, + "vocab_size": 128256 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000..6e76d5c --- /dev/null +++ b/generation_config.json @@ -0,0 +1,9 @@ +{ + "_from_model_config": true, + "bos_token_id": 128000, + "do_sample": true, + "eos_token_id": 128001, + "temperature": 0.6, + "top_p": 0.9, + "transformers_version": "4.50.0" +} diff --git a/model-00001-of-00004.safetensors b/model-00001-of-00004.safetensors new file mode 100644 index 0000000..bde9920 --- /dev/null +++ b/model-00001-of-00004.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c3602b0ac16f7bd9cdb705bf33c57c7fd7fc40475c3f17975b8908e83bef34a +size 4976698672 diff --git a/model-00002-of-00004.safetensors b/model-00002-of-00004.safetensors new file mode 100644 index 0000000..10aa1f6 --- /dev/null +++ b/model-00002-of-00004.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2a513d618ed31ba3493e4055aada0253388cea02e4585da3744764a0fca8f9e9 +size 4999802720 diff --git a/model-00003-of-00004.safetensors b/model-00003-of-00004.safetensors new file mode 100644 index 0000000..15f2c8f --- /dev/null +++ b/model-00003-of-00004.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3eff8581b3cae3e53a7962b2d615b809834568012c124bdd5076c11911432ec8 +size 4915916176 diff --git a/model-00004-of-00004.safetensors b/model-00004-of-00004.safetensors new file mode 100644 index 0000000..9fd0726 --- /dev/null +++ b/model-00004-of-00004.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:50696f6e2f8acd4578e5e15630d07d20070898f2d0d637c5ac6ec1554f6d263c +size 1168138808 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000..0fd8120 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,298 @@ +{ + "metadata": { + "total_size": 16060522496 + }, + "weight_map": { + "lm_head.weight": "model-00004-of-00004.safetensors", + "model.embed_tokens.weight": "model-00001-of-00004.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.10.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.11.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.12.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.13.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.14.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.15.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.16.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.17.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.18.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.19.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.2.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.20.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.21.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.22.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.23.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.24.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.25.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.26.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.27.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.28.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.29.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.3.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.30.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.31.input_layernorm.weight": "model-00004-of-00004.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00004-of-00004.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00004-of-00004.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.4.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.5.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.6.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.7.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.8.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.9.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.norm.weight": "model-00004-of-00004.safetensors" + } +} diff --git a/runs/May29_08-03-21_dpo0--958cfdb21e51-doqv5oixhk/events.out.tfevents.1748505941.dpo0--958cfdb21e51-doqv5oixhk.556397.0 b/runs/May29_08-03-21_dpo0--958cfdb21e51-doqv5oixhk/events.out.tfevents.1748505941.dpo0--958cfdb21e51-doqv5oixhk.556397.0 new file mode 100644 index 0000000000000000000000000000000000000000..6639cc20021b246ee457059363d728e42e65be2f GIT binary patch literal 23847 zcmchfd3+Pq`o;sO*suwuBFIvTECsu!g|@U!DYEYg72&vn{?yA#ZbKBoB>Ozv$v^PQj5Ua1KMf(`~hK zj^g-n4zr7Q#PhV3&SAS*a_#x68#@5_+0#`~cq2e?Xj&6!W6^&O+WHmY_p#H*yvyFE%g;Iu~_@bO~ML z3pq>hHM@g1nDs7nnYPn8$s=66shHMf6Qnl>8k+^1<LSI>Bb4_t{JAyhCT;3(b07 z>VYngbLsPSPV;!4v`Z3u2ik})LQT??CMLT(+%s%$iv_g_eTLG+>`TC@n%y#fri&hPmE8LoZ$?dRhP%RxFq_X_;j(caxq!Dm<}$JLd+I{ z=1dkK0op`CoHn%n4CvO+@`Jw~j3#dT$Bu?sx#i4VOnhb6W$Qf1OjSMFpiEC(qzIUsXWeMb=zs#W`SHyjF%otTG7k$tE6681e4A}h9>1Q z#?CpN=0f=u@*SOnw_D75jy;38F^{)r64FVi1&grV*L-?wG`{DBtASfId~@-V>}sL z7E2zdFCg6|!^o*KIs~hZu&FPw3uYUhw`_t=?>10S*gM=#@(jB>@GGA z1_dVie9mIjkvUgycG8K9+Q8dr_hLA=D=;O~$Low_QVi{m_~7ujTt*F@PL!G~0`ZxQ zSX%+PDn)KE*ufi>mSnlp(WyjF!l6;48&=-sFzczYd2SQA$%>S(jQ*EmtVQ4qbh&T{ zcAbTD;wZ%_SLRxI2hAfOqkR}BI0|?N?M0x=>^5OEae&yj38h(`_R(8Hmjmx7CKml!eRVEn;tx z|I`<EA2CxgtPGdt-rBuyHM`DPPYQFxs+CJM}W zDl$3^^g?n7ObMZzD6)+T-RbZZCvvnX#R0NQ z;K_=}5uivgxB@3xqg>KfehXPx+r2d!f$?*wVD>C*ka(xbZ?iYKBZcTGpAJ$T={eD5yQxqNLtp!JG#KGKzSf(raVKjC&uIu$gUu-m`r~*lPe!q5hGm+3<70< z1{#VJ3E3M64i|Yh(T$c^0^a~)Z-|lOIGZ6ahTIO!K7>vqnE~WMCK5`PFL5eSx+_kA zZZhS%$Rb2=&LgkDHyvci>FI2e=QbLH)nG0n1Hnp0wS#USsMv5qe45(QM}63CK8L@}AYhvn*C8Pu=Gz-)Hc?l2dU35V|R zXn^R+oMjwW?9|b2qy8#fvrT{mzv6887CMB!2#?vOx3~>t&9;$mII^`M!XorG;z z7K4Yq+?#}<3(mx9mu72uEW}7r(cOkLV9>W3b4c7|7^sR&}U*}o$gKo8#ToqDkp18*n4EmlCua}h4K>AFL@hAccBt;%D5HH5lm(>GZ6d>$P$T_ zWP*2+Jj5-IL(s#(?3}?smrOdoDXmIZr8Sk@mMqhW6KWhe021S%U*dN1eItGg1QtU@ zN^eq#Exa_p%c~`7psbL-D0@UyTkP4=noOLEh&U=Vv21 z@G;TO<>6d|8lm}t^+6CU2w3s(NIq9^noU-L7$NqCo_>%%%4Y^f@&Qj6Gx0Whq7*oY zaL|Q@BwApRG$egGVqjCfz#GY9w(XTl*Du? zJ37^wIScvNaM15+6CGQmM^bX+;DRibWUofklMhEzG#W`NP6OnyO)QnS=JDcWLJLmP zBqer{9$~eRO>?jfQcUtVu}~+aPErb(N3=TnIF0ludNWX+`1uobtAH<-CNiC@3mnoq z$X=A3GE?+ibZ(Tab6B0?(W5SK%1RCB1n_ID#&gsyh>%0ZOM{Hrj7V78KjU}Oj zC+E8cXM#ST%oVmMbKD{|vyv>CL<|N~1~0p8ZzN_niXjna}(BMMU;d9BQd< zu8zWL@<&aCJ!e#f|LES_x!o3Z2uT02KifAViAkSVH&r*1q-0vQE5%maEA=PojvhcZ z4PrWAlcZBh^=z5x8yx@Qew2)zDPc0=LLmF=x!m2Go_X7fs$9r~ph{^4Nl@Nt{Be@k zS=w%4>p+z?Cj9b8#Mk{4gBq(Fsq08pihtS=bvd_H)=&H^SE2&~kx}j6(;GZR8EqWG zXlc7-ZQ9$Ga>S_mq0&AqqrKjZ?!Ld=Hy1gP!9>uBAV$||`dSO}k@;ho5SSmtsH*o9 znSZ#SyoDH@%#_tJTG4;Sp4Y6q$`PZ>nYMaHJxj-CZ~yDq1eD%BrUvQBeU=&Z)Oap) z&VfFc5u@jrv?@kD))N28SW6tT=^oPon}QgvSUby?`EtjND4B+jlsI8Bh*9J4m%ORJ z&h3q=jAuenWiX=^pHA|w7*q2H#Av!|hKyVoMptA-<#ux-M(qbbvWSef3}H0AM| zbx+krjHZ_>?ZYzq)sAPn&lz#@JaS?z6G0~sqc;|x(Hxp3WFYh3XF_1UoOfUz?RjB( z)^**8%ZSlEOj#YH##f^MJ@sn6t%%X{Oj|vp8qL-0HLv|u3#E6TsX=;jpJhgkf1l18 zSfc+LG5YKw#oVeGt;i|$_ne;J3E8A!I$%=}qpDdmeMbA!f1zZCFeOX|F{Dnl3@a_v^umX&u)5ThR~UwlMlv_lA^zHjPi zqmD(+MU1|<MkC(cK5H;YFE+(e9CWy&cebd_*B~!?hFd4+CDyO?wHK%Y9 zs`6bX1XTtzs@gWuJ8yFTafngZYQ^3VM#);AeS7B*;}N43+dS=R)3rW1gwaOR{*yiO zRPJ@esCT>4J}je6bo+A$w@EsKoVdwE&(z29v+9GTx}hw?UHzMQwrsO8J- z?Y|7Xj~Gp2%IX-^)LXFs=F@H0Ax4KXZS{;+9GjqZ{FQYKrDtbqke=LUnNiKrSZ%{Q z#Vrw|3z@VkMm=AZ`qxd`a0c15gXw@xL5vzV&GDT(dp#N@^Bq&dWDujC*12Ba{LNZa zWhE1WDuWp{E}rN;;x(^CjMm+ygc`!=(#T%fEne

5WvzRpyU(aaD=KhUhsE`IB~ zR}iC~1f_jgM!#6{M(&MuvsNJ|jxrH+0x^2G=&0s}Z+{bz`8Symm@nroe>3%Np!L3y z^%P>XPNLG!>KLtfLhbD^^Ta=h(T+^}I7X*v3sc`6gVO8C)F3^%&oZN`Pv>d}885Uz zjOv-RDn^aJmHMp*Os^rEN|+AV6wGMuY~T1BJ@Qd9%b5};gBUfA$??8<_Sx~M%Dqeo zstjh-V;=9#j-JvRF}glki6(^6A=hl#YepwdMvQKF@3RXcqdh|y?Nuu+`*gR5C5X}D z#Y+3IjLvp7%gs4jYdLaa6B9uv5TiYs_%*|B^qhmtKgEQ=d^zs`qnfkBv~S#9Za|FQ zVanZf}?y9!-j-M_eM*A{p zRg8L;miXVhkWvTPWMn#EQxKySugvwOrBB_9l9|qwFd4+CswB&sK4T$1y;#MBpvquI zD;gJjGd~!07%{qhsbX&kqccXmmHkK8C4R){&Ob(`Jx$m8!6A$;Xi%B8nY%j`F*@^4 zrF~dNFHQD!`~6B&JX&`#5p)7E+NsfAO`jUhqsaX8ObE=E^Oon)@rBy8J&w#ojNWI; z>KLt9wAy=nHOC`HoBXBJ_;^P5p6{VUjHWU*NKfvw%&6yo^>fl@j(P$yI+RJPVpQXs z;{R^(d_0c|Ob2WVV$?W(zOQ`P?yHE=x0n(pgBhKr_4a-*1@CA+VnR@5Fr%LBMc$7J zhs{KcuBue*4Po?h?uzXHuHL#DF}iTZuJ=VoM};st>EzC=l>aPSiWq%;rqVtvqXh%+ zcU!nzzZy9)kBOiYh|xZawrRYho{mC{Zev1VzMQu_kB*qGT{dFOPQ>UbrmT)p&EBKl z2LDd|05N)(X&=w19zw(i);Fr!pa^GMLedd*i(Ozj^Hx zVszMS#h?&IvpVg}E{g5tK#We?+WDTys3C;WO}`G!Dr?jIL&Rw6CrbOUjP`F8)9vA- zXPO}=I3|KlAVxDztkr0Dzh92bpT>m1d^vBK(G9b;t;&Y>M~tp!%IX+ZO{`aTesG7k z5TjnEt)9_}V^g%}Z`{v8>0M-Mke=LUnNj1wp*at3cHM>;{gX+nV$@hZ!QZ;{x&zr1 zxlIWL*c8O5rpdd$VgFg8L&+pDB}@h}s(L-k`)VpT9921(2|<-1jQZW)Pj5N8B1WIw zq4YL{(WQ}xvM;pUcLFi0O?-Yx1YPTGA&kEKiyrrvq$67->mGqBu&{b7LbR`pnrPS;DT$>uT#D-iP z3RKVxgz4VoBF*G6zRk#taX<+IkTVZ(s+zb-+icJpOI)CUvYeCry zQG@Cv1jN>VwHO?67>FNF>d|(i+oLh^Banj-lRtd5@Q zKN?wcJ@P6V2!U5Yuo@SC=$n4Zdl&_j38XL}M62;px;IU~t_P}gBv67%LvVfC>0NbX z;I^26$wdvb9!9+IkX0JO>u_TYZQ3ipev5eB<}Dp3@;Wz^S4%{3*BBJg854x% zRTsHgJ7Ck&O~{T9fC}1yc&*WTw5INcejQK-UZ4aS$id6=>I<8+H?Mwq0r7esNUP(u zqVCGF4;v+!5wCv&@#A?_$8N{xV37xyO&~?N+cK}Jl}B>kS^uCM@fr`bRq^UcFZJK> zyx# zI?~5n3gLCes5;u`H*Yo|UK?N9eNp6faVW1TC0)AakKXkR;`JC1gyq$i9jT3YFe(Sx z@c^ix9f;R=Dzh}#*KT?TWzhTw)zBhc|cnwum79s@9eqxI`V1*5CX4) zcr`w?#&>wq&%01SUjiu%2=S`Xr+MdVwGB|Emw*yf8p3OD!CUX|VV4oF8!xkPLwMae zFiLyv-8)kdum6<$GV9R&z{*fw+pmo1nzXm14Dq`75EF#u^_{wdv@e{#+!@*72P$X> z;x(gBYt5B9*RP=rt^p;;Kn^~@tI_s}*5I!3G2*r6Vb;}Zd0oDZ>O?gUCuX$Y@nHt(~=6Dtv~ZH};nLU<*AzOQZH`RiMV z*TrYYE)#j(6w0e5;(kW^9Mf{dYsNVy2+QmB-*#*JPEY+e;?)dP&<@1w+C@)l3hSrz zLmA8iN|1pZyu5#%zFAvqQ0-X6>oy>*me-kk%7$$DH5c)E9Ecy!>ra1XeTY*08OTA3 za<^q(RokxT?3`SyFXHvd@0i6^@oKb@v;TVuuOhEn10nD#h*#rN>wN!x@t<52P!^EF zfDo@0r!%~ZYZO(YN=E`Es5F>Y&ql%fa)-~75U&j`uu4OCZO|oFYpcKGZ^Y}-9-aOa zdHpn$*CwBz$`~;@?R&)Ql%7lwme)_-UZ_2^u*W;dj;%li?LfS~=RK2rcIDYZl)*`$ z1R2P|2YA&S`dHg=;FYb2*E>L3Ew7WkWfN;|pMiL7(2LdicwXD@`A3CPj0bX%qTFqn zSI_L9b825dG#v5T7ig>ERdr{g-#1x{|00tK2!U5Yys8q{`h-opPoRLN11Ss$^ZIw1 z_sXs__)l6^0VSw3nAeJ9Ht)LmnmUNrL4BCXA-raF(rD88S!DJ7q0Eb5ne z4e{FjIVK3p>sxEiWm}IGKSFkl1uAF<;U zRq?8coa&$QaqVr$t0W)$F|YSce|}yx3TPmZ!hjI36`!VikI(*M3aZosl%UcO zUdIUD^)1J>M7;jfj8z)KEBU*B+TER|E=9cV&u}!aN9WhC!}I#%7hPu|UM*djAS|zI zeu~K%(e1Sj$d08z1?@n*nij_)HE%EGN1&N2J@<^WAoO1run}S zuZC>qQV6eyY7fzVW=lALcs-WcW4y@g#qhkAzn(h>@%nBS6NKgU`lE~4{fkRpM0Oki zDrg7dwadNJNew?{#8I2ccSGKz1Ms{ohDrg7d^;A7RY059n zKR~=z042yk4qoQ9XuEb*?DzSI*Bd}u9j~gN4wkjC=gmdDs=Ki|AJ1#s)c>wPDYgT0 zkfPjenOEb$-*P5K{B{%ZnhUg5@v2HG^^aKg!sp1VF+d2s3g-2|2A|e^c@PR{JdnbG z5U-xgY2I}g9_>MuE(S_aX)v#HfZK!{h( z-gNIq%?vB5vqR23m7%=S z@2#43PG7~m`k5dsuQ%hjW&d_&<|oLG(?A97K)gP6yC@;@$#!p{4DJIZ$UqKWo?ka@ z)~=pA?L6YO=|R@j>Uh=MJyd2qwB{`0wKEVup4Y(@F$pNeSAZO(D0f@tRTFtP=fK>U zeTY{F&{oAO-M@CIcOo8n^$rjMuY!3UzRq_sD*a~^&=w$t0byPX(!Irpf4Ge*JqDDZ z(hy#c+Pvj;R{V{49e;==6vFG2b46P7wTPLhpOwbg8zQfzH9LZSioY44BfnWhe*-M4 z$>^p%hi&=s*VMq@RXB6-Uq42CW@5_r>bB~Z((epNO6gY;;ZJf!-Ptfa=j6T}rJBG) z>TZZ09&rskq^-J*x;c6X{e@`ot6ov}9-huzJ#yXP_JK$2y}Z3=MBRMWkEZG->SqFv z5Pv0?{OX4@inbn<{^HL_)~N~^{V5gkSGA-5JHKi6?=AnCh8*q-WFZcxpPZsyxuJC1 F{{bgXpoahe literal 0 HcmV?d00001 diff --git a/special_tokens_map.json b/special_tokens_map.json new file mode 100644 index 0000000..344c826 --- /dev/null +++ b/special_tokens_map.json @@ -0,0 +1,23 @@ +{ + "bos_token": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/tokenizer.json b/tokenizer.json new file mode 100644 index 0000000..1c1d8d5 --- /dev/null +++ b/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b +size 17209920 diff --git a/tokenizer_config.json b/tokenizer_config.json new file mode 100644 index 0000000..8ae1bc6 --- /dev/null +++ b/tokenizer_config.json @@ -0,0 +1,2064 @@ +{ + "added_tokens_decoder": { + "128000": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128001": { + "content": "<|end_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128002": { + "content": "<|reserved_special_token_0|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128003": { + "content": "<|reserved_special_token_1|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128004": { + "content": "<|finetune_right_pad_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128005": { + "content": "<|reserved_special_token_2|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128006": { + "content": "<|start_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128007": { + "content": "<|end_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128008": { + "content": "<|eom_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128009": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128010": { + "content": "<|python_tag|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128011": { + "content": "<|reserved_special_token_3|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128012": { + "content": "<|reserved_special_token_4|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128013": { + "content": "<|reserved_special_token_5|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128014": { + "content": "<|reserved_special_token_6|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128015": { + "content": "<|reserved_special_token_7|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128016": { + "content": "<|reserved_special_token_8|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128017": { + "content": "<|reserved_special_token_9|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128018": { + "content": "<|reserved_special_token_10|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128019": { + "content": "<|reserved_special_token_11|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128020": { + "content": "<|reserved_special_token_12|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128021": { + "content": "<|reserved_special_token_13|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128022": { + "content": "<|reserved_special_token_14|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128023": { + "content": "<|reserved_special_token_15|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128024": { + "content": "<|reserved_special_token_16|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128025": { + "content": "<|reserved_special_token_17|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128026": { + "content": "<|reserved_special_token_18|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128027": { + "content": "<|reserved_special_token_19|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128028": { + "content": "<|reserved_special_token_20|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128029": { + "content": "<|reserved_special_token_21|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128030": { + "content": "<|reserved_special_token_22|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128031": { + "content": "<|reserved_special_token_23|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128032": { + "content": "<|reserved_special_token_24|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128033": { + "content": "<|reserved_special_token_25|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128034": { + "content": "<|reserved_special_token_26|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128035": { + "content": "<|reserved_special_token_27|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128036": { + "content": "<|reserved_special_token_28|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128037": { + "content": "<|reserved_special_token_29|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128038": { + "content": "<|reserved_special_token_30|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128039": { + "content": "<|reserved_special_token_31|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128040": { + "content": "<|reserved_special_token_32|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128041": { + "content": "<|reserved_special_token_33|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128042": { + "content": "<|reserved_special_token_34|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128043": { + "content": "<|reserved_special_token_35|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128044": { + "content": "<|reserved_special_token_36|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128045": { + "content": "<|reserved_special_token_37|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128046": { + "content": "<|reserved_special_token_38|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128047": { + "content": "<|reserved_special_token_39|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128048": { + "content": "<|reserved_special_token_40|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128049": { + "content": "<|reserved_special_token_41|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128050": { + "content": "<|reserved_special_token_42|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128051": { + "content": "<|reserved_special_token_43|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128052": { + "content": "<|reserved_special_token_44|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128053": { + "content": "<|reserved_special_token_45|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128054": { + "content": "<|reserved_special_token_46|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128055": { + "content": "<|reserved_special_token_47|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128056": { + "content": "<|reserved_special_token_48|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128057": { + "content": "<|reserved_special_token_49|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128058": { + "content": "<|reserved_special_token_50|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128059": { + "content": "<|reserved_special_token_51|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128060": { + "content": "<|reserved_special_token_52|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128061": { + "content": "<|reserved_special_token_53|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128062": { + "content": "<|reserved_special_token_54|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128063": { + "content": "<|reserved_special_token_55|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128064": { + "content": "<|reserved_special_token_56|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128065": { + "content": "<|reserved_special_token_57|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128066": { + "content": "<|reserved_special_token_58|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128067": { + "content": "<|reserved_special_token_59|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128068": { + "content": "<|reserved_special_token_60|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128069": { + "content": "<|reserved_special_token_61|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128070": { + "content": "<|reserved_special_token_62|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128071": { + "content": "<|reserved_special_token_63|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128072": { + "content": "<|reserved_special_token_64|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128073": { + "content": "<|reserved_special_token_65|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128074": { + "content": "<|reserved_special_token_66|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128075": { + "content": "<|reserved_special_token_67|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128076": { + "content": "<|reserved_special_token_68|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128077": { + "content": "<|reserved_special_token_69|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128078": { + "content": "<|reserved_special_token_70|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128079": { + "content": "<|reserved_special_token_71|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128080": { + "content": "<|reserved_special_token_72|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128081": { + "content": "<|reserved_special_token_73|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128082": { + "content": "<|reserved_special_token_74|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128083": { + "content": "<|reserved_special_token_75|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128084": { + "content": "<|reserved_special_token_76|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128085": { + "content": "<|reserved_special_token_77|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128086": { + "content": "<|reserved_special_token_78|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128087": { + "content": "<|reserved_special_token_79|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128088": { + "content": "<|reserved_special_token_80|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128089": { + "content": "<|reserved_special_token_81|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128090": { + "content": "<|reserved_special_token_82|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128091": { + "content": "<|reserved_special_token_83|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128092": { + "content": "<|reserved_special_token_84|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128093": { + "content": "<|reserved_special_token_85|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128094": { + "content": "<|reserved_special_token_86|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128095": { + "content": "<|reserved_special_token_87|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128096": { + "content": "<|reserved_special_token_88|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128097": { + "content": "<|reserved_special_token_89|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128098": { + "content": "<|reserved_special_token_90|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128099": { + "content": "<|reserved_special_token_91|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128100": { + "content": "<|reserved_special_token_92|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128101": { + "content": "<|reserved_special_token_93|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128102": { + "content": "<|reserved_special_token_94|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128103": { + "content": "<|reserved_special_token_95|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128104": { + "content": "<|reserved_special_token_96|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128105": { + "content": "<|reserved_special_token_97|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128106": { + "content": "<|reserved_special_token_98|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128107": { + "content": "<|reserved_special_token_99|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128108": { + "content": "<|reserved_special_token_100|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128109": { + "content": "<|reserved_special_token_101|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128110": { + "content": "<|reserved_special_token_102|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128111": { + "content": "<|reserved_special_token_103|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128112": { + "content": "<|reserved_special_token_104|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128113": { + "content": "<|reserved_special_token_105|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128114": { + "content": "<|reserved_special_token_106|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128115": { + "content": "<|reserved_special_token_107|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128116": { + "content": "<|reserved_special_token_108|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128117": { + "content": "<|reserved_special_token_109|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128118": { + "content": "<|reserved_special_token_110|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128119": { + "content": "<|reserved_special_token_111|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128120": { + "content": "<|reserved_special_token_112|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128121": { + "content": "<|reserved_special_token_113|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128122": { + "content": "<|reserved_special_token_114|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128123": { + "content": "<|reserved_special_token_115|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128124": { + "content": "<|reserved_special_token_116|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128125": { + "content": "<|reserved_special_token_117|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128126": { + "content": "<|reserved_special_token_118|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128127": { + "content": "<|reserved_special_token_119|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128128": { + "content": "<|reserved_special_token_120|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128129": { + "content": "<|reserved_special_token_121|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128130": { + "content": "<|reserved_special_token_122|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128131": { + "content": "<|reserved_special_token_123|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128132": { + "content": "<|reserved_special_token_124|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128133": { + "content": "<|reserved_special_token_125|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128134": { + "content": "<|reserved_special_token_126|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128135": { + "content": "<|reserved_special_token_127|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128136": { + "content": "<|reserved_special_token_128|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128137": { + "content": "<|reserved_special_token_129|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128138": { + "content": "<|reserved_special_token_130|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128139": { + "content": "<|reserved_special_token_131|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128140": { + "content": "<|reserved_special_token_132|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128141": { + "content": "<|reserved_special_token_133|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128142": { + "content": "<|reserved_special_token_134|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128143": { + "content": "<|reserved_special_token_135|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128144": { + "content": "<|reserved_special_token_136|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128145": { + "content": "<|reserved_special_token_137|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128146": { + "content": "<|reserved_special_token_138|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128147": { + "content": "<|reserved_special_token_139|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128148": { + "content": "<|reserved_special_token_140|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128149": { + "content": "<|reserved_special_token_141|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128150": { + "content": "<|reserved_special_token_142|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128151": { + "content": "<|reserved_special_token_143|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128152": { + "content": "<|reserved_special_token_144|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128153": { + "content": "<|reserved_special_token_145|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128154": { + "content": "<|reserved_special_token_146|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128155": { + "content": "<|reserved_special_token_147|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128156": { + "content": "<|reserved_special_token_148|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128157": { + "content": "<|reserved_special_token_149|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128158": { + "content": "<|reserved_special_token_150|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128159": { + "content": "<|reserved_special_token_151|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128160": { + "content": "<|reserved_special_token_152|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128161": { + "content": "<|reserved_special_token_153|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128162": { + "content": "<|reserved_special_token_154|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128163": { + "content": "<|reserved_special_token_155|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128164": { + "content": "<|reserved_special_token_156|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128165": { + "content": "<|reserved_special_token_157|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128166": { + "content": "<|reserved_special_token_158|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128167": { + "content": "<|reserved_special_token_159|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128168": { + "content": "<|reserved_special_token_160|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128169": { + "content": "<|reserved_special_token_161|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128170": { + "content": "<|reserved_special_token_162|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128171": { + "content": "<|reserved_special_token_163|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128172": { + "content": "<|reserved_special_token_164|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128173": { + "content": "<|reserved_special_token_165|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128174": { + "content": "<|reserved_special_token_166|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128175": { + "content": "<|reserved_special_token_167|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128176": { + "content": "<|reserved_special_token_168|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128177": { + "content": "<|reserved_special_token_169|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128178": { + "content": "<|reserved_special_token_170|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128179": { + "content": "<|reserved_special_token_171|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128180": { + "content": "<|reserved_special_token_172|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128181": { + "content": "<|reserved_special_token_173|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128182": { + "content": "<|reserved_special_token_174|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128183": { + "content": "<|reserved_special_token_175|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128184": { + "content": "<|reserved_special_token_176|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128185": { + "content": "<|reserved_special_token_177|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128186": { + "content": "<|reserved_special_token_178|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128187": { + "content": "<|reserved_special_token_179|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128188": { + "content": "<|reserved_special_token_180|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128189": { + "content": "<|reserved_special_token_181|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128190": { + "content": "<|reserved_special_token_182|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128191": { + "content": "<|reserved_special_token_183|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128192": { + "content": "<|reserved_special_token_184|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128193": { + "content": "<|reserved_special_token_185|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128194": { + "content": "<|reserved_special_token_186|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128195": { + "content": "<|reserved_special_token_187|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128196": { + "content": "<|reserved_special_token_188|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128197": { + "content": "<|reserved_special_token_189|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128198": { + "content": "<|reserved_special_token_190|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128199": { + "content": "<|reserved_special_token_191|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128200": { + "content": "<|reserved_special_token_192|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128201": { + "content": "<|reserved_special_token_193|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128202": { + "content": "<|reserved_special_token_194|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128203": { + "content": "<|reserved_special_token_195|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128204": { + "content": "<|reserved_special_token_196|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128205": { + "content": "<|reserved_special_token_197|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128206": { + "content": "<|reserved_special_token_198|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128207": { + "content": "<|reserved_special_token_199|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128208": { + "content": "<|reserved_special_token_200|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128209": { + "content": "<|reserved_special_token_201|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128210": { + "content": "<|reserved_special_token_202|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128211": { + "content": "<|reserved_special_token_203|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128212": { + "content": "<|reserved_special_token_204|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128213": { + "content": "<|reserved_special_token_205|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128214": { + "content": "<|reserved_special_token_206|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128215": { + "content": "<|reserved_special_token_207|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128216": { + "content": "<|reserved_special_token_208|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128217": { + "content": "<|reserved_special_token_209|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128218": { + "content": "<|reserved_special_token_210|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128219": { + "content": "<|reserved_special_token_211|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128220": { + "content": "<|reserved_special_token_212|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128221": { + "content": "<|reserved_special_token_213|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128222": { + "content": "<|reserved_special_token_214|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128223": { + "content": "<|reserved_special_token_215|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128224": { + "content": "<|reserved_special_token_216|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128225": { + "content": "<|reserved_special_token_217|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128226": { + "content": "<|reserved_special_token_218|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128227": { + "content": "<|reserved_special_token_219|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128228": { + "content": "<|reserved_special_token_220|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128229": { + "content": "<|reserved_special_token_221|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128230": { + "content": "<|reserved_special_token_222|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128231": { + "content": "<|reserved_special_token_223|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128232": { + "content": "<|reserved_special_token_224|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128233": { + "content": "<|reserved_special_token_225|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128234": { + "content": "<|reserved_special_token_226|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128235": { + "content": "<|reserved_special_token_227|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128236": { + "content": "<|reserved_special_token_228|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128237": { + "content": "<|reserved_special_token_229|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128238": { + "content": "<|reserved_special_token_230|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128239": { + "content": "<|reserved_special_token_231|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128240": { + "content": "<|reserved_special_token_232|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128241": { + "content": "<|reserved_special_token_233|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128242": { + "content": "<|reserved_special_token_234|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128243": { + "content": "<|reserved_special_token_235|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128244": { + "content": "<|reserved_special_token_236|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128245": { + "content": "<|reserved_special_token_237|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128246": { + "content": "<|reserved_special_token_238|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128247": { + "content": "<|reserved_special_token_239|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128248": { + "content": "<|reserved_special_token_240|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128249": { + "content": "<|reserved_special_token_241|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128250": { + "content": "<|reserved_special_token_242|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128251": { + "content": "<|reserved_special_token_243|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128252": { + "content": "<|reserved_special_token_244|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128253": { + "content": "<|reserved_special_token_245|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128254": { + "content": "<|reserved_special_token_246|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128255": { + "content": "<|reserved_special_token_247|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "bos_token": "<|begin_of_text|>", + "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", + "clean_up_tokenization_spaces": true, + "eos_token": "<|eot_id|>", + "extra_special_tokens": {}, + "model_input_names": [ + "input_ids", + "attention_mask" + ], + "model_max_length": 131072, + "pad_token": "<|eot_id|>", + "tokenizer_class": "PreTrainedTokenizer" +} diff --git a/trainer_state.json b/trainer_state.json new file mode 100644 index 0000000..f24af12 --- /dev/null +++ b/trainer_state.json @@ -0,0 +1,409 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9992709599027947, + "eval_steps": 500, + "global_step": 257, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.038882138517618466, + "grad_norm": 160.0573020755792, + "learning_rate": 3.846153846153846e-08, + "logits/chosen": -0.7651611566543579, + "logits/rejected": -0.9917968511581421, + "logps/chosen": -1039.6500244140625, + "logps/rejected": -269.35626220703125, + "loss": 0.7032, + "rewards/accuracies": 0.17109374701976776, + "rewards/chosen": -0.00718765240162611, + "rewards/margins": -0.006218528840690851, + "rewards/rejected": -0.0009794235229492188, + "step": 10 + }, + { + "epoch": 0.07776427703523693, + "grad_norm": 160.5753020431693, + "learning_rate": 7.692307692307692e-08, + "logits/chosen": -0.7669433355331421, + "logits/rejected": -0.991040050983429, + "logps/chosen": -1027.425048828125, + "logps/rejected": -273.2875061035156, + "loss": 0.6959, + "rewards/accuracies": 0.23593750596046448, + "rewards/chosen": 0.008073806762695312, + "rewards/margins": 0.009277725592255592, + "rewards/rejected": -0.0012153625721111894, + "step": 20 + }, + { + "epoch": 0.1166464155528554, + "grad_norm": 144.69837671419592, + "learning_rate": 9.992603458948281e-08, + "logits/chosen": -0.766284167766571, + "logits/rejected": -0.9904540777206421, + "logps/chosen": -1034.0250244140625, + "logps/rejected": -269.0625, + "loss": 0.6534, + "rewards/accuracies": 0.4437499940395355, + "rewards/chosen": 0.06487135589122772, + "rewards/margins": 0.10399703681468964, + "rewards/rejected": -0.03914298862218857, + "step": 30 + }, + { + "epoch": 0.15552855407047386, + "grad_norm": 114.38616995997658, + "learning_rate": 9.909643486313532e-08, + "logits/chosen": -0.7626953125, + "logits/rejected": -0.987060546875, + "logps/chosen": -1035.2249755859375, + "logps/rejected": -268.70001220703125, + "loss": 0.5462, + "rewards/accuracies": 0.81640625, + "rewards/chosen": 0.22022247314453125, + "rewards/margins": 0.36003416776657104, + "rewards/rejected": -0.1398262083530426, + "step": 40 + }, + { + "epoch": 0.19441069258809235, + "grad_norm": 95.18572051179285, + "learning_rate": 9.73601505145955e-08, + "logits/chosen": -0.7583252191543579, + "logits/rejected": -0.9866698980331421, + "logps/chosen": -1038.300048828125, + "logps/rejected": -277.33123779296875, + "loss": 0.4385, + "rewards/accuracies": 0.983593761920929, + "rewards/chosen": 0.379730224609375, + "rewards/margins": 0.663256824016571, + "rewards/rejected": -0.283303827047348, + "step": 50 + }, + { + "epoch": 0.2332928311057108, + "grad_norm": 58.648166089179284, + "learning_rate": 9.474924624108548e-08, + "logits/chosen": -0.7514404058456421, + "logits/rejected": -0.9781738519668579, + "logps/chosen": -1039.574951171875, + "logps/rejected": -274.51873779296875, + "loss": 0.3168, + "rewards/accuracies": 0.999218761920929, + "rewards/chosen": 0.5596984624862671, + "rewards/margins": 1.0992431640625, + "rewards/rejected": -0.539135754108429, + "step": 60 + }, + { + "epoch": 0.27217496962332927, + "grad_norm": 38.61380385007199, + "learning_rate": 9.131193871579974e-08, + "logits/chosen": -0.740234375, + "logits/rejected": -0.972973644733429, + "logps/chosen": -1028.800048828125, + "logps/rejected": -280.6625061035156, + "loss": 0.1797, + "rewards/accuracies": 1.0, + "rewards/chosen": 0.835314929485321, + "rewards/margins": 1.780029296875, + "rewards/rejected": -0.9450439214706421, + "step": 70 + }, + { + "epoch": 0.3110571081409477, + "grad_norm": 24.436947703379207, + "learning_rate": 8.71117061502135e-08, + "logits/chosen": -0.7481689453125, + "logits/rejected": -0.9714599847793579, + "logps/chosen": -1023.0499877929688, + "logps/rejected": -284.95623779296875, + "loss": 0.1221, + "rewards/accuracies": 1.0, + "rewards/chosen": 0.9628051519393921, + "rewards/margins": 2.2330079078674316, + "rewards/rejected": -1.270166039466858, + "step": 80 + }, + { + "epoch": 0.34993924665856624, + "grad_norm": 31.163290960443195, + "learning_rate": 8.222611601511082e-08, + "logits/chosen": -0.7407470941543579, + "logits/rejected": -0.968579113483429, + "logps/chosen": -1046.9000244140625, + "logps/rejected": -283.3500061035156, + "loss": 0.0878, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.033911108970642, + "rewards/margins": 2.625439405441284, + "rewards/rejected": -1.591650366783142, + "step": 90 + }, + { + "epoch": 0.3888213851761847, + "grad_norm": 10.132641150484275, + "learning_rate": 7.674539256930362e-08, + "logits/chosen": -0.73974609375, + "logits/rejected": -0.960034191608429, + "logps/chosen": -1007.0999755859375, + "logps/rejected": -294.1812438964844, + "loss": 0.0515, + "rewards/accuracies": 0.999218761920929, + "rewards/chosen": 1.075415015220642, + "rewards/margins": 3.2860350608825684, + "rewards/rejected": -2.210693359375, + "step": 100 + }, + { + "epoch": 0.42770352369380316, + "grad_norm": 3.381766003891569, + "learning_rate": 7.077075065009432e-08, + "logits/chosen": -0.735546886920929, + "logits/rejected": -0.957446277141571, + "logps/chosen": -1014.7249755859375, + "logps/rejected": -305.45623779296875, + "loss": 0.0199, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.205664038658142, + "rewards/margins": 4.345410346984863, + "rewards/rejected": -3.1407227516174316, + "step": 110 + }, + { + "epoch": 0.4665856622114216, + "grad_norm": 2.9204173229145636, + "learning_rate": 6.441252649607854e-08, + "logits/chosen": -0.729931652545929, + "logits/rejected": -0.951171875, + "logps/chosen": -1013.6749877929688, + "logps/rejected": -311.09375, + "loss": 0.0101, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.242578148841858, + "rewards/margins": 5.182031154632568, + "rewards/rejected": -3.9403319358825684, + "step": 120 + }, + { + "epoch": 0.5054678007290401, + "grad_norm": 1.4356612196732965, + "learning_rate": 5.7788140121173145e-08, + "logits/chosen": -0.7265380620956421, + "logits/rejected": -0.9493408203125, + "logps/chosen": -1042.8499755859375, + "logps/rejected": -315.17498779296875, + "loss": 0.0053, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.3319091796875, + "rewards/margins": 5.879687309265137, + "rewards/rejected": -4.5478515625, + "step": 130 + }, + { + "epoch": 0.5443499392466585, + "grad_norm": 1.344341217145263, + "learning_rate": 5.1019926869570276e-08, + "logits/chosen": -0.7250732183456421, + "logits/rejected": -0.9361572265625, + "logps/chosen": -1033.800048828125, + "logps/rejected": -322.75, + "loss": 0.0038, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.329248070716858, + "rewards/margins": 6.33203125, + "rewards/rejected": -5.001562595367432, + "step": 140 + }, + { + "epoch": 0.583232077764277, + "grad_norm": 0.5751713296126671, + "learning_rate": 4.4232878197208657e-08, + "logits/chosen": -0.720263659954071, + "logits/rejected": -0.9346923828125, + "logps/chosen": -1047.300048828125, + "logps/rejected": -326.04998779296875, + "loss": 0.0028, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.3618042469024658, + "rewards/margins": 6.713671684265137, + "rewards/rejected": -5.349413871765137, + "step": 150 + }, + { + "epoch": 0.6221142162818954, + "grad_norm": 0.8087854397697334, + "learning_rate": 3.75523334017065e-08, + "logits/chosen": -0.7223144769668579, + "logits/rejected": -0.9326171875, + "logps/chosen": -1028.300048828125, + "logps/rejected": -328.9125061035156, + "loss": 0.0021, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.3705565929412842, + "rewards/margins": 6.984765529632568, + "rewards/rejected": -5.613085746765137, + "step": 160 + }, + { + "epoch": 0.660996354799514, + "grad_norm": 1.4449125849513227, + "learning_rate": 3.1101664928554675e-08, + "logits/chosen": -0.729809582233429, + "logits/rejected": -0.9361816644668579, + "logps/chosen": -1021.7249755859375, + "logps/rejected": -330.04998779296875, + "loss": 0.0019, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.3477783203125, + "rewards/margins": 7.177734375, + "rewards/rejected": -5.828906059265137, + "step": 170 + }, + { + "epoch": 0.6998784933171325, + "grad_norm": 1.1450098797738104, + "learning_rate": 2.500000000000001e-08, + "logits/chosen": -0.722607433795929, + "logits/rejected": -0.932812511920929, + "logps/chosen": -1011.75, + "logps/rejected": -328.375, + "loss": 0.0016, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.342016577720642, + "rewards/margins": 7.329297065734863, + "rewards/rejected": -5.984179496765137, + "step": 180 + }, + { + "epoch": 0.7387606318347509, + "grad_norm": 0.43432731079566683, + "learning_rate": 1.9360020642261153e-08, + "logits/chosen": -0.7257324457168579, + "logits/rejected": -0.9291747808456421, + "logps/chosen": -1048.449951171875, + "logps/rejected": -332.2875061035156, + "loss": 0.0021, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.404748558998108, + "rewards/margins": 7.463281154632568, + "rewards/rejected": -6.056640625, + "step": 190 + }, + { + "epoch": 0.7776427703523694, + "grad_norm": 0.42312272714084886, + "learning_rate": 1.428588273890482e-08, + "logits/chosen": -0.725415050983429, + "logits/rejected": -0.9350341558456421, + "logps/chosen": -1040.5999755859375, + "logps/rejected": -332.82501220703125, + "loss": 0.0014, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.384423851966858, + "rewards/margins": 7.4599609375, + "rewards/rejected": -6.073828220367432, + "step": 200 + }, + { + "epoch": 0.8165249088699879, + "grad_norm": 2.545449065189939, + "learning_rate": 9.871292540103377e-09, + "logits/chosen": -0.722412109375, + "logits/rejected": -0.929858386516571, + "logps/chosen": -1036.6500244140625, + "logps/rejected": -329.54998779296875, + "loss": 0.0014, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.4049561023712158, + "rewards/margins": 7.583984375, + "rewards/rejected": -6.179296970367432, + "step": 210 + }, + { + "epoch": 0.8554070473876063, + "grad_norm": 0.43163218586837404, + "learning_rate": 6.1977761496908695e-09, + "logits/chosen": -0.7271728515625, + "logits/rejected": -0.9333740472793579, + "logps/chosen": -1029.324951171875, + "logps/rejected": -330.95001220703125, + "loss": 0.0014, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.387841820716858, + "rewards/margins": 7.583984375, + "rewards/rejected": -6.195703029632568, + "step": 220 + }, + { + "epoch": 0.8942891859052248, + "grad_norm": 1.1334297545543626, + "learning_rate": 3.3331739481316624e-09, + "logits/chosen": -0.7221924066543579, + "logits/rejected": -0.9290771484375, + "logps/chosen": -1033.625, + "logps/rejected": -335.51251220703125, + "loss": 0.0015, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.386499047279358, + "rewards/margins": 7.572070121765137, + "rewards/rejected": -6.1845703125, + "step": 230 + }, + { + "epoch": 0.9331713244228432, + "grad_norm": 0.2245653138567375, + "learning_rate": 1.3303877555292442e-09, + "logits/chosen": -0.725512683391571, + "logits/rejected": -0.934619128704071, + "logps/chosen": -1013.4249877929688, + "logps/rejected": -333.98748779296875, + "loss": 0.0012, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.3474853038787842, + "rewards/margins": 7.607226371765137, + "rewards/rejected": -6.259961128234863, + "step": 240 + }, + { + "epoch": 0.9720534629404617, + "grad_norm": 0.34410802938606294, + "learning_rate": 2.2640387134577055e-10, + "logits/chosen": -0.724291980266571, + "logits/rejected": -0.932788074016571, + "logps/chosen": -1036.6500244140625, + "logps/rejected": -330.67498779296875, + "loss": 0.0014, + "rewards/accuracies": 1.0, + "rewards/chosen": 1.348962426185608, + "rewards/margins": 7.596093654632568, + "rewards/rejected": -6.248437404632568, + "step": 250 + } + ], + "logging_steps": 10, + "max_steps": 257, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 0.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +} diff --git a/training_args.bin b/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..845dc761a8317a2aa13d0436daf473424c75c1fd GIT binary patch literal 7800 zcmcIpd7K?tReo6}$-LRO%s#*{FrAs{<)zb|1e}S4Bps$FZ`#hF&(T=atGe%1b=P+8 zt?JGTG|VUs{bW%QR7ON_#|2c}P~3L~6c-S87#$T+5K%T|{LZ;m@4fD1fPbv~y3=*< zx#ymHmhYT%YhixJ&RT8Hp4xfO-rA+LT)SZsCSAwXT{G#sxjT_9M<}wvV;y*d{IVtjHZd)U>1ZA`DE1iOE8( zCIjD}Ts7lucjn;!#_ZJifrFhx;|DyqF@E^a0X6Q=wmo0956}4R_N3dGni`MYURzD} ziXct8=Yq7D)T&+Nxf@NM2orTm`45~LPaCwebuI6^qV%c`bLEUka8bX-XZKS_jzGKm zt=;xrt+)~G?u*En&b0DFFQ+~I)@2~zS;t^auDU(C7)o#x(@Ax#w5hb=6nPk#`U{d= zX^@{Tsq0j4Z3l}arAyZ7(nY#V#49%l)5M8VW1KXJdUScebu&mar5#_bgr2fuI&C-i z0>^~s6kRdDt##988WC=U`ZdRPn#uH#6kYB5p-OVc^}Hf3 zA~$DYO|CLS*UVojEM(~g=C5svB4yfFk=vu|=4-MdPc@FB;wvwx6B>+1O&j-x)Q&VA}+P6veXL<-3qH*j0s+4HFVqkPvU7^g~3QW1_$+v2zJeT znWCqCF$X$b(LMyyIz7F$3p~8kz{Tkq>va1fJ+pPqR<_Hl9J6%C8a>OZ%GGg^InZL5 z(zEA|yz7npJT^LK<-8MiVacPk>qKfrMKmTMk-5f%D~k5nBUH1FR#_B!F5M+m*p;nC zcZ*t>2V6wKtWcY;=$?(%C6;%WVTCbZy!ASa{C{J>q%IOOdCKkWKkQ5$8lRdT-`{Y6 zR#W5Shi49X9lyQ5p=KK6eo8AdX?QkRritnZBI+R4Cxd$HB1mktHMLHSMcUuGdgN4{ zP@7z=PSX}GpofwP$I5~EdTZ|pec~ogm~@;@7y&70W<&hVxGNkACBn>pQ3nC(5M2%i z5bDq(9TsOV+3>#4IrXV+awZweQ{bc%g>jhEbDP6)<91Z85|e5}U$G&^7Wz$gytm)F zf_1<$GEg}&tQ*mN{nnir*2*BN6~Z}++OD^ZV<32q)6r?{z?r?}EDe*Kj##iwkTVM3 zrLVNFg~7_Kpy%~lmxCF<#+fPckB;_RNLV$aluyr}ze`+|6O$t&$1G?pg}F%NFy7E{ zi}B1T%+y&rVOg|0jai!85O$pmf1k7)^50IH<)!z(K*0LJ2Hh`43*4x~$-oN@qaPnn zB|gmy8@I?)dO*Zx?oOwOA}hnG5#Gbmrj;b`Oz)=$EkZ^qw1zNEmT3Wz$kF2a*iV2ibecJlYjXN( zo9P6gt!sZxQw}^r%{#89dKo#6S`d!Y&d4;h0@p#PcU%wlr?0gq0&wK;Zy%a}*kT}O zT*RZ3xUr(I3w8&gxTER!P6BO7wqLXCk}BY=qPiw8yixk{bbHoMbufN?wB~p7UBTJJ8$-f&4(33C2xR zOph$Xkp$PlSiT`n%F)UpEk0q~ID>>#YzTeAfrmg@=1jBFZ{30oFI=3IS#Y&kv@KF7 z0Aa@1aklgraUv6^X&!J!DqXiX2bS_k009lcn9f~5DZGXgZT&$eK|Q9MaB5Sh7mEqG zpjd^zQR2``)@f~#UMi?83}qp`tSOR#vfTR-i>+OqB44+AfT5(D2lS}*0Q*+Pf{mbW zl!Bln`zGOrq@3y;8`I0r(E6CjVnMx~y4t5#$gx)4#Bb9pl?`M#a0g*pMVmM=G67*qA1*V-V;3KOuy2SuB{r7wsE z$Sh)YUN-tx>x>{Q(IfQf+lFIz>1@l3u?~Z&;-7 z5EJn}d5Yvg-)Vt{2QnS50EW1zO6j|rl4N<@jaJ>BariLwMth2Z!w(ZGhr?r5u@|0o z%6Rzhrf4rryr}Sz6%#aMNY9A>GJnvUtm-R@(W&n_LvI#6+G1WBJ`YG+1@QOUAeA&) zrN_4f4~QvwNm2C;Woh%)rtN=*h8pVs`4#)L&ehi8M>?e+wm>Br+UG?;>CnXVBi5NDko_L=R!VQPajRX})=N>t z|EQDzl1zVWkhcK}3hgHS_^Ca+1sH_|!Z*&sNYPK&%(Q86kn7)W;d%2UCpLOVGce6S zH(NL2Je7^ancySvMWc{zeEP}01X)GEsYy(2bz~S{6Z$D@=|pw%B2g&QL&TIDWr0gS zZ7pNLDhJFV^07bK7DMl}!dcykCseNUGb0@`fEu-*&go~ZVkKU~E<{8Tk)UGyIqS^A zfG4|d9$+?YFDW$JKW`6!(9v!er2&VFLHnt>LADLCpD}Pxh_-@9@UkZ|6vf;?f0LsJhcQ*e$yVGC)nL0CGC2dCVtD-W;|Jr{I*4#9EBO8nPF61?RVsUgoTz; z>USjpSMqhVQuKR$8>6U+u}vZ9)xjM`m*r_2`u)DRDc8&+^CLm`2TjRFd74Is{?LY% zT!Bby{2r-pOxTT6ocO&GWZc%%`-V*{H`N_cp`DrERJI zSb$nHQp2vx5xsw%{$!EXr>rs6gbdXefC?dBieZa7{XL4u53kWj#>BkR{ylEH z!779NgUlqJ7tudTy<@|lp5Ru)3+nc@PX8oy(og>LeC-wVFN^fi;GseNIS@=dzz2vfQ#A*l z+gt#llUFD=7&6R+U+r5z*?EX>kCi@Fv93CW&x7^nT7v>_L-VTB_y#v3; z1ashvNJcw(*$!Efb~@5TmaCN37s--cyd-$SI$?@a?Z$w3L$ZQe5^YVO_h6`5h!HC` zQbg;Eg++ZUi%T%DR4&>FaYdNvDg`!X+el>=9QZc4%wb$d>PxY5qY#0Ig)5idK|_`ZCG&GyuLsC~| zt-(3UaMCw}WJxS79}r645TjNkw=fA?Qm)9)=QMPWTbYROuF$?JeH*`VBv(r;*CF}S zn6Xf?3HemhKoPh^Cvv;`>6o=xY-i*q&bo}0H2N7J*pS4Fb61}x;6u1u!wmozI!50P znz92IyE`#A+WMKon|GD>;a!-%12e(>TL8x1Z0Kk4+>;U#z@!M}8~o_m7~GIZ^Ks+d z(Rbqg%S4Z@u%XAWcHg6VAO2g@cj5O~@P##fH~+CT4E!Y<;`hj~4eTWw!sD{!39)Ax zc1t!gC(4=(>jY^<|Xq2@lj*flV zqD;+B&mNdLcyRjQ;RCY=rW#W-jp-?Q8Z!rHW*dj5ry7T5XBsT!JMTSy%NQi&pN08t zEbQRR=crZqZhrG4(7{E}rthFzd5l_sBYXI~6t>>XV_?m#Oh#47SD?e^1LrX*O}CZB z=sE8guE$@CxGCE#$AvAhFu!ZpJ>|;Vc8m{rFnaM*{m<~k&u91s4iDDx*VEcw&eeE$ RdSU(|c1f(+Sz~%j`+wM=yhs25 literal 0 HcmV?d00001