From 521b6fffe27f2e0a440af432d7c046debe4f4ade Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Thu, 23 Apr 2026 18:15:00 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: Neelectric/Llama-3.1-8B-Instruct_SDFT_mathv00.09 Source: Original Platform --- .gitattributes | 36 + README.md | 69 + all_results.json | 9 + chat_template.jinja | 109 + config.json | 36 + generation_config.json | 9 + model-00001-of-00004.safetensors | 3 + model-00002-of-00004.safetensors | 3 + model-00003-of-00004.safetensors | 3 + model-00004-of-00004.safetensors | 3 + model.safetensors.index.json | 299 + special_tokens_map.json | 17 + tokenizer.json | 3 + tokenizer_config.json | 2063 +++++++ train_results.json | 9 + trainer_state.json | 9262 ++++++++++++++++++++++++++++++ training_args.bin | 3 + 17 files changed, 11936 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 all_results.json create mode 100644 chat_template.jinja create mode 100644 config.json create mode 100644 generation_config.json create mode 100644 model-00001-of-00004.safetensors create mode 100644 model-00002-of-00004.safetensors create mode 100644 model-00003-of-00004.safetensors create mode 100644 model-00004-of-00004.safetensors create mode 100644 model.safetensors.index.json create mode 100644 special_tokens_map.json create mode 100644 tokenizer.json create mode 100644 tokenizer_config.json create mode 100644 train_results.json create mode 100644 trainer_state.json create mode 100644 training_args.bin diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..52373fe --- /dev/null +++ b/.gitattributes @@ -0,0 +1,36 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +tokenizer.json filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..b647fa0 --- /dev/null +++ b/README.md @@ -0,0 +1,69 @@ +--- +base_model: meta-llama/Llama-3.1-8B-Instruct +datasets: Neelectric/OpenR1-Math-220k_all_SDFT_nr +library_name: transformers +model_name: Llama-3.1-8B-Instruct_SDFT_mathv00.09 +tags: +- generated_from_trainer +- open-r1 +- sdft +- trl +licence: license +--- + +# Model Card for Llama-3.1-8B-Instruct_SDFT_mathv00.09 + +This model is a fine-tuned version of [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) on the [Neelectric/OpenR1-Math-220k_all_SDFT_nr](https://huggingface.co/datasets/Neelectric/OpenR1-Math-220k_all_SDFT_nr) dataset. +It has been trained using [TRL](https://github.com/huggingface/trl). + +## Quick start + +```python +from transformers import pipeline + +question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" +generator = pipeline("text-generation", model="Neelectric/Llama-3.1-8B-Instruct_SDFT_mathv00.09", device="cuda") +output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] +print(output["generated_text"]) +``` + +## Training procedure + +[Visualize in Weights & Biases](https://wandb.ai/neelectric/open-r1_math/runs/iq80gxyh) + + + +This model was trained with SDFT, a method introduced in [Self-Training with On-Policy Self-Distillation for Language Model Alignment](https://huggingface.co/papers/2601.19897). + +### Framework versions + +- TRL: 1.1.0.dev0 +- Transformers: 4.57.6 +- Pytorch: 2.9.0 +- Datasets: 4.8.4 +- Tokenizers: 0.22.2 + +## Citations + +Cite SDFT as: + +```bibtex +@article{hubotter2026selftraining, + title = {{Self-Training with On-Policy Self-Distillation for Language Model Alignment}}, + author = {Jonas H\"ubotter and Frederike L\"ubeck and Lejs Behric and Anton Baumann and Marco Bagatella and Daniel Marta and Ido Hakimi and Idan Shenfeld and Thomas Kleine Buening and Carlos Guestrin and Andreas Krause}, + year = 2026, + eprint = {arXiv:2601.19897} +} +``` + +Cite TRL as: + +```bibtex +@software{vonwerra2020trl, + title = {{TRL: Transformers Reinforcement Learning}}, + author = {von Werra, Leandro and Belkada, Younes and Tunstall, Lewis and Beeching, Edward and Thrush, Tristan and Lambert, Nathan and Huang, Shengyi and Rasul, Kashif and Gallouédec, Quentin}, + license = {Apache-2.0}, + url = {https://github.com/huggingface/trl}, + year = {2020} +} +``` \ No newline at end of file diff --git a/all_results.json b/all_results.json new file mode 100644 index 0000000..75fcb09 --- /dev/null +++ b/all_results.json @@ -0,0 +1,9 @@ +{ + "epoch": 3.0, + "total_flos": 0.0, + "train_loss": 0.06867336523107444, + "train_runtime": 41172.7067, + "train_samples": 225129, + "train_samples_per_second": 16.404, + "train_steps_per_second": 0.032 +} \ No newline at end of file diff --git a/chat_template.jinja b/chat_template.jinja new file mode 100644 index 0000000..33089ac --- /dev/null +++ b/chat_template.jinja @@ -0,0 +1,109 @@ +{{- bos_token }} +{%- if custom_tools is defined %} + {%- set tools = custom_tools %} +{%- endif %} +{%- if not tools_in_user_message is defined %} + {%- set tools_in_user_message = true %} +{%- endif %} +{%- if not date_string is defined %} + {%- set date_string = "26 Jul 2024" %} +{%- endif %} +{%- if not tools is defined %} + {%- set tools = none %} +{%- endif %} + +{#- This block extracts the system message, so we can slot it into the right place. #} +{%- if messages[0]['role'] == 'system' %} + {%- set system_message = messages[0]['content']|trim %} + {%- set messages = messages[1:] %} +{%- else %} + {%- set system_message = "" %} +{%- endif %} + +{#- System message + builtin tools #} +{{- "<|start_header_id|>system<|end_header_id|>\n\n" }} +{%- if builtin_tools is defined or tools is not none %} + {{- "Environment: ipython\n" }} +{%- endif %} +{%- if builtin_tools is defined %} + {{- "Tools: " + builtin_tools | reject('equalto', 'code_interpreter') | join(", ") + "\n\n"}} +{%- endif %} +{{- "Cutting Knowledge Date: December 2023\n" }} +{{- "Today Date: " + date_string + "\n\n" }} +{%- if tools is not none and not tools_in_user_message %} + {{- "You have access to the following functions. To call a function, please respond with JSON for a function call." }} + {{- 'Respond in the format {"name": function name, "parameters": dictionary of argument name and its value}.' }} + {{- "Do not use variables.\n\n" }} + {%- for t in tools %} + {{- t | tojson(indent=4) }} + {{- "\n\n" }} + {%- endfor %} +{%- endif %} +{{- system_message }} +{{- "<|eot_id|>" }} + +{#- Custom tools are passed in a user message with some extra guidance #} +{%- if tools_in_user_message and not tools is none %} + {#- Extract the first user message so we can plug it in here #} + {%- if messages | length != 0 %} + {%- set first_user_message = messages[0]['content']|trim %} + {%- set messages = messages[1:] %} + {%- else %} + {{- raise_exception("Cannot put tools in the first user message when there's no first user message!") }} +{%- endif %} + {{- '<|start_header_id|>user<|end_header_id|>\n\n' -}} + {{- "Given the following functions, please respond with a JSON for a function call " }} + {{- "with its proper arguments that best answers the given prompt.\n\n" }} + {{- 'Respond in the format {"name": function name, "parameters": dictionary of argument name and its value}.' }} + {{- "Do not use variables.\n\n" }} + {%- for t in tools %} + {{- t | tojson(indent=4) }} + {{- "\n\n" }} + {%- endfor %} + {{- first_user_message + "<|eot_id|>"}} +{%- endif %} + +{%- for message in messages %} + {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %} + {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' }} + {%- elif 'tool_calls' in message %} + {%- if not message.tool_calls|length == 1 %} + {{- raise_exception("This model only supports single tool-calls at once!") }} + {%- endif %} + {%- set tool_call = message.tool_calls[0].function %} + {%- if builtin_tools is defined and tool_call.name in builtin_tools %} + {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}} + {{- "<|python_tag|>" + tool_call.name + ".call(" }} + {%- for arg_name, arg_val in tool_call.arguments | items %} + {{- arg_name + '="' + arg_val + '"' }} + {%- if not loop.last %} + {{- ", " }} + {%- endif %} + {%- endfor %} + {{- ")" }} + {%- else %} + {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}} + {{- '{"name": "' + tool_call.name + '", ' }} + {{- '"parameters": ' }} + {{- tool_call.arguments | tojson }} + {{- "}" }} + {%- endif %} + {%- if builtin_tools is defined %} + {#- This means we're in ipython mode #} + {{- "<|eom_id|>" }} + {%- else %} + {{- "<|eot_id|>" }} + {%- endif %} + {%- elif message.role == "tool" or message.role == "ipython" %} + {{- "<|start_header_id|>ipython<|end_header_id|>\n\n" }} + {%- if message.content is mapping or message.content is iterable %} + {{- message.content | tojson }} + {%- else %} + {{- message.content }} + {%- endif %} + {{- "<|eot_id|>" }} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' }} +{%- endif %} diff --git a/config.json b/config.json new file mode 100644 index 0000000..06df27b --- /dev/null +++ b/config.json @@ -0,0 +1,36 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 128000, + "dtype": "bfloat16", + "eos_token_id": 128009, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 4096, + "initializer_range": 0.02, + "intermediate_size": 14336, + "max_position_embeddings": 131072, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 32, + "num_hidden_layers": 32, + "num_key_value_heads": 8, + "pad_token_id": 128009, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": { + "factor": 8.0, + "high_freq_factor": 4.0, + "low_freq_factor": 1.0, + "original_max_position_embeddings": 8192, + "rope_type": "llama3" + }, + "rope_theta": 500000.0, + "tie_word_embeddings": false, + "transformers_version": "4.57.6", + "use_cache": true, + "vocab_size": 128256 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000..50f6077 --- /dev/null +++ b/generation_config.json @@ -0,0 +1,9 @@ +{ + "bos_token_id": 128000, + "do_sample": true, + "eos_token_id": 128009, + "pad_token_id": 128009, + "temperature": 0.6, + "top_p": 0.9, + "transformers_version": "4.57.6" +} diff --git a/model-00001-of-00004.safetensors b/model-00001-of-00004.safetensors new file mode 100644 index 0000000..30fadad --- /dev/null +++ b/model-00001-of-00004.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:026d6e58d9b3bd58a89c67fe7f4a2de3587751f3a6fd40cbbfd002e5bd59bf4c +size 4976698672 diff --git a/model-00002-of-00004.safetensors b/model-00002-of-00004.safetensors new file mode 100644 index 0000000..b1c2bd3 --- /dev/null +++ b/model-00002-of-00004.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5146df6a061e993f679c95371b356e0ad034375910aefc32dc1bfabe35ba321c +size 4999802720 diff --git a/model-00003-of-00004.safetensors b/model-00003-of-00004.safetensors new file mode 100644 index 0000000..45b570e --- /dev/null +++ b/model-00003-of-00004.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:22be0a94dbab241cde57a11553656c12bd4f8e43b52b7d63071205083d0e70d5 +size 4915916176 diff --git a/model-00004-of-00004.safetensors b/model-00004-of-00004.safetensors new file mode 100644 index 0000000..42d2831 --- /dev/null +++ b/model-00004-of-00004.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:49a5ad32881984cfa2e272b252f1dcfca5a999b7c1d03e919cb25b009e998524 +size 1168138808 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000..5c64f1e --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,299 @@ +{ + "metadata": { + "total_parameters": 8030261248, + "total_size": 16060522496 + }, + "weight_map": { + "lm_head.weight": "model-00004-of-00004.safetensors", + "model.embed_tokens.weight": "model-00001-of-00004.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.1.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.10.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.11.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.12.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.13.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.14.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.15.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.16.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.17.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.18.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.19.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.2.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.20.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.21.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.22.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.23.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.24.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.25.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.26.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.27.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.28.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.29.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.3.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.30.input_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00003-of-00004.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.31.input_layernorm.weight": "model-00004-of-00004.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00004-of-00004.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00004-of-00004.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00003-of-00004.safetensors", + "model.layers.4.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.5.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.6.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.7.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.8.input_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00004.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00004.safetensors", + "model.layers.9.input_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00002-of-00004.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00002-of-00004.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00002-of-00004.safetensors", + "model.norm.weight": "model-00004-of-00004.safetensors" + } +} diff --git a/special_tokens_map.json b/special_tokens_map.json new file mode 100644 index 0000000..b43be96 --- /dev/null +++ b/special_tokens_map.json @@ -0,0 +1,17 @@ +{ + "bos_token": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": "<|eot_id|>" +} diff --git a/tokenizer.json b/tokenizer.json new file mode 100644 index 0000000..356db7b --- /dev/null +++ b/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dce9b61086126b8dd5c77f984b948929028bb0f4429feed6eb1ddcb791362b38 +size 17210183 diff --git a/tokenizer_config.json b/tokenizer_config.json new file mode 100644 index 0000000..3beeacc --- /dev/null +++ b/tokenizer_config.json @@ -0,0 +1,2063 @@ +{ + "added_tokens_decoder": { + "128000": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128001": { + "content": "<|end_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128002": { + "content": "<|reserved_special_token_0|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128003": { + "content": "<|reserved_special_token_1|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128004": { + "content": "<|finetune_right_pad_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128005": { + "content": "<|reserved_special_token_2|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128006": { + "content": "<|start_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128007": { + "content": "<|end_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128008": { + "content": "<|eom_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128009": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128010": { + "content": "<|python_tag|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128011": { + "content": "<|reserved_special_token_3|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128012": { + "content": "<|reserved_special_token_4|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128013": { + "content": "<|reserved_special_token_5|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128014": { + "content": "<|reserved_special_token_6|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128015": { + "content": "<|reserved_special_token_7|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128016": { + "content": "<|reserved_special_token_8|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128017": { + "content": "<|reserved_special_token_9|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128018": { + "content": "<|reserved_special_token_10|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128019": { + "content": "<|reserved_special_token_11|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128020": { + "content": "<|reserved_special_token_12|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128021": { + "content": "<|reserved_special_token_13|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128022": { + "content": "<|reserved_special_token_14|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128023": { + "content": "<|reserved_special_token_15|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128024": { + "content": "<|reserved_special_token_16|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128025": { + "content": "<|reserved_special_token_17|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128026": { + "content": "<|reserved_special_token_18|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128027": { + "content": "<|reserved_special_token_19|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128028": { + "content": "<|reserved_special_token_20|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128029": { + "content": "<|reserved_special_token_21|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128030": { + "content": "<|reserved_special_token_22|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128031": { + "content": "<|reserved_special_token_23|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128032": { + "content": "<|reserved_special_token_24|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128033": { + "content": "<|reserved_special_token_25|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128034": { + "content": "<|reserved_special_token_26|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128035": { + "content": "<|reserved_special_token_27|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128036": { + "content": "<|reserved_special_token_28|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128037": { + "content": "<|reserved_special_token_29|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128038": { + "content": "<|reserved_special_token_30|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128039": { + "content": "<|reserved_special_token_31|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128040": { + "content": "<|reserved_special_token_32|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128041": { + "content": "<|reserved_special_token_33|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128042": { + "content": "<|reserved_special_token_34|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128043": { + "content": "<|reserved_special_token_35|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128044": { + "content": "<|reserved_special_token_36|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128045": { + "content": "<|reserved_special_token_37|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128046": { + "content": "<|reserved_special_token_38|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128047": { + "content": "<|reserved_special_token_39|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128048": { + "content": "<|reserved_special_token_40|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128049": { + "content": "<|reserved_special_token_41|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128050": { + "content": "<|reserved_special_token_42|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128051": { + "content": "<|reserved_special_token_43|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128052": { + "content": "<|reserved_special_token_44|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128053": { + "content": "<|reserved_special_token_45|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128054": { + "content": "<|reserved_special_token_46|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128055": { + "content": "<|reserved_special_token_47|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128056": { + "content": "<|reserved_special_token_48|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128057": { + "content": "<|reserved_special_token_49|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128058": { + "content": "<|reserved_special_token_50|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128059": { + "content": "<|reserved_special_token_51|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128060": { + "content": "<|reserved_special_token_52|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128061": { + "content": "<|reserved_special_token_53|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128062": { + "content": "<|reserved_special_token_54|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128063": { + "content": "<|reserved_special_token_55|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128064": { + "content": "<|reserved_special_token_56|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128065": { + "content": "<|reserved_special_token_57|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128066": { + "content": "<|reserved_special_token_58|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128067": { + "content": "<|reserved_special_token_59|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128068": { + "content": "<|reserved_special_token_60|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128069": { + "content": "<|reserved_special_token_61|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128070": { + "content": "<|reserved_special_token_62|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128071": { + "content": "<|reserved_special_token_63|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128072": { + "content": "<|reserved_special_token_64|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128073": { + "content": "<|reserved_special_token_65|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128074": { + "content": "<|reserved_special_token_66|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128075": { + "content": "<|reserved_special_token_67|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128076": { + "content": "<|reserved_special_token_68|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128077": { + "content": "<|reserved_special_token_69|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128078": { + "content": "<|reserved_special_token_70|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128079": { + "content": "<|reserved_special_token_71|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128080": { + "content": "<|reserved_special_token_72|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128081": { + "content": "<|reserved_special_token_73|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128082": { + "content": "<|reserved_special_token_74|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128083": { + "content": "<|reserved_special_token_75|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128084": { + "content": "<|reserved_special_token_76|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128085": { + "content": "<|reserved_special_token_77|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128086": { + "content": "<|reserved_special_token_78|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128087": { + "content": "<|reserved_special_token_79|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128088": { + "content": "<|reserved_special_token_80|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128089": { + "content": "<|reserved_special_token_81|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128090": { + "content": "<|reserved_special_token_82|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128091": { + "content": "<|reserved_special_token_83|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128092": { + "content": "<|reserved_special_token_84|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128093": { + "content": "<|reserved_special_token_85|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128094": { + "content": "<|reserved_special_token_86|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128095": { + "content": "<|reserved_special_token_87|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128096": { + "content": "<|reserved_special_token_88|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128097": { + "content": "<|reserved_special_token_89|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128098": { + "content": "<|reserved_special_token_90|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128099": { + "content": "<|reserved_special_token_91|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128100": { + "content": "<|reserved_special_token_92|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128101": { + "content": "<|reserved_special_token_93|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128102": { + "content": "<|reserved_special_token_94|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128103": { + "content": "<|reserved_special_token_95|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128104": { + "content": "<|reserved_special_token_96|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128105": { + "content": "<|reserved_special_token_97|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128106": { + "content": "<|reserved_special_token_98|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128107": { + "content": "<|reserved_special_token_99|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128108": { + "content": "<|reserved_special_token_100|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128109": { + "content": "<|reserved_special_token_101|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128110": { + "content": "<|reserved_special_token_102|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128111": { + "content": "<|reserved_special_token_103|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128112": { + "content": "<|reserved_special_token_104|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128113": { + "content": "<|reserved_special_token_105|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128114": { + "content": "<|reserved_special_token_106|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128115": { + "content": "<|reserved_special_token_107|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128116": { + "content": "<|reserved_special_token_108|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128117": { + "content": "<|reserved_special_token_109|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128118": { + "content": "<|reserved_special_token_110|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128119": { + "content": "<|reserved_special_token_111|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128120": { + "content": "<|reserved_special_token_112|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128121": { + "content": "<|reserved_special_token_113|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128122": { + "content": "<|reserved_special_token_114|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128123": { + "content": "<|reserved_special_token_115|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128124": { + "content": "<|reserved_special_token_116|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128125": { + "content": "<|reserved_special_token_117|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128126": { + "content": "<|reserved_special_token_118|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128127": { + "content": "<|reserved_special_token_119|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128128": { + "content": "<|reserved_special_token_120|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128129": { + "content": "<|reserved_special_token_121|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128130": { + "content": "<|reserved_special_token_122|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128131": { + "content": "<|reserved_special_token_123|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128132": { + "content": "<|reserved_special_token_124|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128133": { + "content": "<|reserved_special_token_125|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128134": { + "content": "<|reserved_special_token_126|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128135": { + "content": "<|reserved_special_token_127|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128136": { + "content": "<|reserved_special_token_128|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128137": { + "content": "<|reserved_special_token_129|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128138": { + "content": "<|reserved_special_token_130|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128139": { + "content": "<|reserved_special_token_131|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128140": { + "content": "<|reserved_special_token_132|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128141": { + "content": "<|reserved_special_token_133|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128142": { + "content": "<|reserved_special_token_134|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128143": { + "content": "<|reserved_special_token_135|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128144": { + "content": "<|reserved_special_token_136|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128145": { + "content": "<|reserved_special_token_137|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128146": { + "content": "<|reserved_special_token_138|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128147": { + "content": "<|reserved_special_token_139|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128148": { + "content": "<|reserved_special_token_140|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128149": { + "content": "<|reserved_special_token_141|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128150": { + "content": "<|reserved_special_token_142|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128151": { + "content": "<|reserved_special_token_143|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128152": { + "content": "<|reserved_special_token_144|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128153": { + "content": "<|reserved_special_token_145|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128154": { + "content": "<|reserved_special_token_146|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128155": { + "content": "<|reserved_special_token_147|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128156": { + "content": "<|reserved_special_token_148|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128157": { + "content": "<|reserved_special_token_149|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128158": { + "content": "<|reserved_special_token_150|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128159": { + "content": "<|reserved_special_token_151|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128160": { + "content": "<|reserved_special_token_152|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128161": { + "content": "<|reserved_special_token_153|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128162": { + "content": "<|reserved_special_token_154|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128163": { + "content": "<|reserved_special_token_155|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128164": { + "content": "<|reserved_special_token_156|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128165": { + "content": "<|reserved_special_token_157|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128166": { + "content": "<|reserved_special_token_158|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128167": { + "content": "<|reserved_special_token_159|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128168": { + "content": "<|reserved_special_token_160|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128169": { + "content": "<|reserved_special_token_161|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128170": { + "content": "<|reserved_special_token_162|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128171": { + "content": "<|reserved_special_token_163|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128172": { + "content": "<|reserved_special_token_164|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128173": { + "content": "<|reserved_special_token_165|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128174": { + "content": "<|reserved_special_token_166|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128175": { + "content": "<|reserved_special_token_167|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128176": { + "content": "<|reserved_special_token_168|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128177": { + "content": "<|reserved_special_token_169|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128178": { + "content": "<|reserved_special_token_170|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128179": { + "content": "<|reserved_special_token_171|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128180": { + "content": "<|reserved_special_token_172|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128181": { + "content": "<|reserved_special_token_173|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128182": { + "content": "<|reserved_special_token_174|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128183": { + "content": "<|reserved_special_token_175|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128184": { + "content": "<|reserved_special_token_176|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128185": { + "content": "<|reserved_special_token_177|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128186": { + "content": "<|reserved_special_token_178|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128187": { + "content": "<|reserved_special_token_179|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128188": { + "content": "<|reserved_special_token_180|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128189": { + "content": "<|reserved_special_token_181|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128190": { + "content": "<|reserved_special_token_182|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128191": { + "content": "<|reserved_special_token_183|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128192": { + "content": "<|reserved_special_token_184|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128193": { + "content": "<|reserved_special_token_185|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128194": { + "content": "<|reserved_special_token_186|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128195": { + "content": "<|reserved_special_token_187|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128196": { + "content": "<|reserved_special_token_188|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128197": { + "content": "<|reserved_special_token_189|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128198": { + "content": "<|reserved_special_token_190|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128199": { + "content": "<|reserved_special_token_191|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128200": { + "content": "<|reserved_special_token_192|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128201": { + "content": "<|reserved_special_token_193|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128202": { + "content": "<|reserved_special_token_194|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128203": { + "content": "<|reserved_special_token_195|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128204": { + "content": "<|reserved_special_token_196|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128205": { + "content": "<|reserved_special_token_197|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128206": { + "content": "<|reserved_special_token_198|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128207": { + "content": "<|reserved_special_token_199|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128208": { + "content": "<|reserved_special_token_200|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128209": { + "content": "<|reserved_special_token_201|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128210": { + "content": "<|reserved_special_token_202|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128211": { + "content": "<|reserved_special_token_203|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128212": { + "content": "<|reserved_special_token_204|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128213": { + "content": "<|reserved_special_token_205|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128214": { + "content": "<|reserved_special_token_206|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128215": { + "content": "<|reserved_special_token_207|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128216": { + "content": "<|reserved_special_token_208|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128217": { + "content": "<|reserved_special_token_209|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128218": { + "content": "<|reserved_special_token_210|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128219": { + "content": "<|reserved_special_token_211|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128220": { + "content": "<|reserved_special_token_212|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128221": { + "content": "<|reserved_special_token_213|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128222": { + "content": "<|reserved_special_token_214|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128223": { + "content": "<|reserved_special_token_215|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128224": { + "content": "<|reserved_special_token_216|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128225": { + "content": "<|reserved_special_token_217|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128226": { + "content": "<|reserved_special_token_218|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128227": { + "content": "<|reserved_special_token_219|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128228": { + "content": "<|reserved_special_token_220|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128229": { + "content": "<|reserved_special_token_221|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128230": { + "content": "<|reserved_special_token_222|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128231": { + "content": "<|reserved_special_token_223|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128232": { + "content": "<|reserved_special_token_224|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128233": { + "content": "<|reserved_special_token_225|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128234": { + "content": "<|reserved_special_token_226|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128235": { + "content": "<|reserved_special_token_227|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128236": { + "content": "<|reserved_special_token_228|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128237": { + "content": "<|reserved_special_token_229|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128238": { + "content": "<|reserved_special_token_230|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128239": { + "content": "<|reserved_special_token_231|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128240": { + "content": "<|reserved_special_token_232|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128241": { + "content": "<|reserved_special_token_233|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128242": { + "content": "<|reserved_special_token_234|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128243": { + "content": "<|reserved_special_token_235|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128244": { + "content": "<|reserved_special_token_236|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128245": { + "content": "<|reserved_special_token_237|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128246": { + "content": "<|reserved_special_token_238|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128247": { + "content": "<|reserved_special_token_239|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128248": { + "content": "<|reserved_special_token_240|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128249": { + "content": "<|reserved_special_token_241|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128250": { + "content": "<|reserved_special_token_242|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128251": { + "content": "<|reserved_special_token_243|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128252": { + "content": "<|reserved_special_token_244|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128253": { + "content": "<|reserved_special_token_245|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128254": { + "content": "<|reserved_special_token_246|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128255": { + "content": "<|reserved_special_token_247|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "bos_token": "<|begin_of_text|>", + "clean_up_tokenization_spaces": true, + "eos_token": "<|eot_id|>", + "extra_special_tokens": {}, + "model_input_names": [ + "input_ids", + "attention_mask" + ], + "model_max_length": 131072, + "pad_token": "<|eot_id|>", + "tokenizer_class": "PreTrainedTokenizerFast" +} diff --git a/train_results.json b/train_results.json new file mode 100644 index 0000000..75fcb09 --- /dev/null +++ b/train_results.json @@ -0,0 +1,9 @@ +{ + "epoch": 3.0, + "total_flos": 0.0, + "train_loss": 0.06867336523107444, + "train_runtime": 41172.7067, + "train_samples": 225129, + "train_samples_per_second": 16.404, + "train_steps_per_second": 0.032 +} \ No newline at end of file diff --git a/trainer_state.json b/trainer_state.json new file mode 100644 index 0000000..b99152e --- /dev/null +++ b/trainer_state.json @@ -0,0 +1,9262 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 3.0, + "eval_steps": 500, + "global_step": 1317, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.002277904328018223, + "grad_norm": 3.7515621185302734, + "learning_rate": 0.0, + "loss": 0.1547, + "step": 1 + }, + { + "epoch": 0.004555808656036446, + "grad_norm": 3.7196226119995117, + "learning_rate": 3.787878787878788e-08, + "loss": 0.1577, + "step": 2 + }, + { + "epoch": 0.00683371298405467, + "grad_norm": 3.765834093093872, + "learning_rate": 7.575757575757576e-08, + "loss": 0.1601, + "step": 3 + }, + { + "epoch": 0.009111617312072893, + "grad_norm": 3.8851237297058105, + "learning_rate": 1.1363636363636364e-07, + "loss": 0.1575, + "step": 4 + }, + { + "epoch": 0.011389521640091117, + "grad_norm": 3.452960252761841, + "learning_rate": 1.5151515151515152e-07, + "loss": 0.1564, + "step": 5 + }, + { + "epoch": 0.01366742596810934, + "grad_norm": 3.7792060375213623, + "learning_rate": 1.893939393939394e-07, + "loss": 0.1606, + "step": 6 + }, + { + "epoch": 0.015945330296127564, + "grad_norm": 3.6438887119293213, + "learning_rate": 2.2727272727272729e-07, + "loss": 0.1575, + "step": 7 + }, + { + "epoch": 0.018223234624145785, + "grad_norm": 3.7777984142303467, + "learning_rate": 2.651515151515152e-07, + "loss": 0.1546, + "step": 8 + }, + { + "epoch": 0.02050113895216401, + "grad_norm": 3.470759868621826, + "learning_rate": 3.0303030303030305e-07, + "loss": 0.1534, + "step": 9 + }, + { + "epoch": 0.022779043280182234, + "grad_norm": 3.5244765281677246, + "learning_rate": 3.409090909090909e-07, + "loss": 0.1567, + "step": 10 + }, + { + "epoch": 0.025056947608200455, + "grad_norm": 3.4608511924743652, + "learning_rate": 3.787878787878788e-07, + "loss": 0.1528, + "step": 11 + }, + { + "epoch": 0.02733485193621868, + "grad_norm": 3.4218990802764893, + "learning_rate": 4.1666666666666667e-07, + "loss": 0.1503, + "step": 12 + }, + { + "epoch": 0.029612756264236904, + "grad_norm": 2.852553367614746, + "learning_rate": 4.5454545454545457e-07, + "loss": 0.1461, + "step": 13 + }, + { + "epoch": 0.03189066059225513, + "grad_norm": 2.992276191711426, + "learning_rate": 4.924242424242425e-07, + "loss": 0.1475, + "step": 14 + }, + { + "epoch": 0.03416856492027335, + "grad_norm": 2.432781457901001, + "learning_rate": 5.303030303030304e-07, + "loss": 0.1398, + "step": 15 + }, + { + "epoch": 0.03644646924829157, + "grad_norm": 2.0288331508636475, + "learning_rate": 5.681818181818182e-07, + "loss": 0.1386, + "step": 16 + }, + { + "epoch": 0.0387243735763098, + "grad_norm": 1.6633920669555664, + "learning_rate": 6.060606060606061e-07, + "loss": 0.1294, + "step": 17 + }, + { + "epoch": 0.04100227790432802, + "grad_norm": 1.4245625734329224, + "learning_rate": 6.43939393939394e-07, + "loss": 0.1309, + "step": 18 + }, + { + "epoch": 0.04328018223234624, + "grad_norm": 1.3093671798706055, + "learning_rate": 6.818181818181818e-07, + "loss": 0.1271, + "step": 19 + }, + { + "epoch": 0.04555808656036447, + "grad_norm": 1.2491767406463623, + "learning_rate": 7.196969696969698e-07, + "loss": 0.1257, + "step": 20 + }, + { + "epoch": 0.04783599088838269, + "grad_norm": 1.123982548713684, + "learning_rate": 7.575757575757576e-07, + "loss": 0.1213, + "step": 21 + }, + { + "epoch": 0.05011389521640091, + "grad_norm": 1.0984852313995361, + "learning_rate": 7.954545454545455e-07, + "loss": 0.1186, + "step": 22 + }, + { + "epoch": 0.05239179954441914, + "grad_norm": 1.37436842918396, + "learning_rate": 8.333333333333333e-07, + "loss": 0.1116, + "step": 23 + }, + { + "epoch": 0.05466970387243736, + "grad_norm": 1.382936954498291, + "learning_rate": 8.712121212121213e-07, + "loss": 0.109, + "step": 24 + }, + { + "epoch": 0.05694760820045558, + "grad_norm": 1.3406814336776733, + "learning_rate": 9.090909090909091e-07, + "loss": 0.1085, + "step": 25 + }, + { + "epoch": 0.05922551252847381, + "grad_norm": 1.279089331626892, + "learning_rate": 9.469696969696971e-07, + "loss": 0.1063, + "step": 26 + }, + { + "epoch": 0.06150341685649203, + "grad_norm": 0.9895230531692505, + "learning_rate": 9.84848484848485e-07, + "loss": 0.1056, + "step": 27 + }, + { + "epoch": 0.06378132118451026, + "grad_norm": 0.8956998586654663, + "learning_rate": 1.0227272727272729e-06, + "loss": 0.1091, + "step": 28 + }, + { + "epoch": 0.06605922551252848, + "grad_norm": 0.815238893032074, + "learning_rate": 1.0606060606060608e-06, + "loss": 0.105, + "step": 29 + }, + { + "epoch": 0.0683371298405467, + "grad_norm": 0.9920597076416016, + "learning_rate": 1.0984848484848485e-06, + "loss": 0.1084, + "step": 30 + }, + { + "epoch": 0.07061503416856492, + "grad_norm": 1.0040704011917114, + "learning_rate": 1.1363636363636364e-06, + "loss": 0.1062, + "step": 31 + }, + { + "epoch": 0.07289293849658314, + "grad_norm": 1.0493673086166382, + "learning_rate": 1.1742424242424245e-06, + "loss": 0.1086, + "step": 32 + }, + { + "epoch": 0.07517084282460136, + "grad_norm": 1.0451900959014893, + "learning_rate": 1.2121212121212122e-06, + "loss": 0.1028, + "step": 33 + }, + { + "epoch": 0.0774487471526196, + "grad_norm": 0.7993922829627991, + "learning_rate": 1.25e-06, + "loss": 0.1013, + "step": 34 + }, + { + "epoch": 0.07972665148063782, + "grad_norm": 0.7440354824066162, + "learning_rate": 1.287878787878788e-06, + "loss": 0.1009, + "step": 35 + }, + { + "epoch": 0.08200455580865604, + "grad_norm": 0.710520327091217, + "learning_rate": 1.3257575757575757e-06, + "loss": 0.0962, + "step": 36 + }, + { + "epoch": 0.08428246013667426, + "grad_norm": 0.7249789834022522, + "learning_rate": 1.3636363636363636e-06, + "loss": 0.0948, + "step": 37 + }, + { + "epoch": 0.08656036446469248, + "grad_norm": 0.7503432631492615, + "learning_rate": 1.4015151515151515e-06, + "loss": 0.0933, + "step": 38 + }, + { + "epoch": 0.0888382687927107, + "grad_norm": 0.6732667684555054, + "learning_rate": 1.4393939393939396e-06, + "loss": 0.094, + "step": 39 + }, + { + "epoch": 0.09111617312072894, + "grad_norm": 0.5606656074523926, + "learning_rate": 1.4772727272727275e-06, + "loss": 0.0951, + "step": 40 + }, + { + "epoch": 0.09339407744874716, + "grad_norm": 0.5518288612365723, + "learning_rate": 1.5151515151515152e-06, + "loss": 0.0924, + "step": 41 + }, + { + "epoch": 0.09567198177676538, + "grad_norm": 0.5367372632026672, + "learning_rate": 1.5530303030303032e-06, + "loss": 0.0938, + "step": 42 + }, + { + "epoch": 0.0979498861047836, + "grad_norm": 0.5601660013198853, + "learning_rate": 1.590909090909091e-06, + "loss": 0.0926, + "step": 43 + }, + { + "epoch": 0.10022779043280182, + "grad_norm": 0.5742217302322388, + "learning_rate": 1.628787878787879e-06, + "loss": 0.0894, + "step": 44 + }, + { + "epoch": 0.10250569476082004, + "grad_norm": 0.4841398596763611, + "learning_rate": 1.6666666666666667e-06, + "loss": 0.0908, + "step": 45 + }, + { + "epoch": 0.10478359908883828, + "grad_norm": 0.4976852238178253, + "learning_rate": 1.7045454545454546e-06, + "loss": 0.092, + "step": 46 + }, + { + "epoch": 0.1070615034168565, + "grad_norm": 0.49381938576698303, + "learning_rate": 1.7424242424242427e-06, + "loss": 0.0889, + "step": 47 + }, + { + "epoch": 0.10933940774487472, + "grad_norm": 0.5361044406890869, + "learning_rate": 1.7803030303030306e-06, + "loss": 0.0912, + "step": 48 + }, + { + "epoch": 0.11161731207289294, + "grad_norm": 0.44863954186439514, + "learning_rate": 1.8181818181818183e-06, + "loss": 0.0878, + "step": 49 + }, + { + "epoch": 0.11389521640091116, + "grad_norm": 0.4702779948711395, + "learning_rate": 1.8560606060606062e-06, + "loss": 0.0895, + "step": 50 + }, + { + "epoch": 0.11617312072892938, + "grad_norm": 0.4760969877243042, + "learning_rate": 1.8939393939393941e-06, + "loss": 0.0869, + "step": 51 + }, + { + "epoch": 0.11845102505694761, + "grad_norm": 0.5038303136825562, + "learning_rate": 1.931818181818182e-06, + "loss": 0.0864, + "step": 52 + }, + { + "epoch": 0.12072892938496584, + "grad_norm": 0.39018872380256653, + "learning_rate": 1.96969696969697e-06, + "loss": 0.0832, + "step": 53 + }, + { + "epoch": 0.12300683371298406, + "grad_norm": 0.4925285577774048, + "learning_rate": 2.0075757575757576e-06, + "loss": 0.0825, + "step": 54 + }, + { + "epoch": 0.1252847380410023, + "grad_norm": 0.521104097366333, + "learning_rate": 2.0454545454545457e-06, + "loss": 0.0839, + "step": 55 + }, + { + "epoch": 0.1275626423690205, + "grad_norm": 0.46636050939559937, + "learning_rate": 2.0833333333333334e-06, + "loss": 0.0822, + "step": 56 + }, + { + "epoch": 0.12984054669703873, + "grad_norm": 0.4480585753917694, + "learning_rate": 2.1212121212121216e-06, + "loss": 0.0831, + "step": 57 + }, + { + "epoch": 0.13211845102505695, + "grad_norm": 0.4655265212059021, + "learning_rate": 2.1590909090909092e-06, + "loss": 0.0813, + "step": 58 + }, + { + "epoch": 0.13439635535307518, + "grad_norm": 0.6499490141868591, + "learning_rate": 2.196969696969697e-06, + "loss": 0.0806, + "step": 59 + }, + { + "epoch": 0.1366742596810934, + "grad_norm": 0.42999497056007385, + "learning_rate": 2.234848484848485e-06, + "loss": 0.082, + "step": 60 + }, + { + "epoch": 0.13895216400911162, + "grad_norm": 0.38357794284820557, + "learning_rate": 2.2727272727272728e-06, + "loss": 0.0794, + "step": 61 + }, + { + "epoch": 0.14123006833712984, + "grad_norm": 0.49868491291999817, + "learning_rate": 2.310606060606061e-06, + "loss": 0.0826, + "step": 62 + }, + { + "epoch": 0.14350797266514806, + "grad_norm": 0.43505406379699707, + "learning_rate": 2.348484848484849e-06, + "loss": 0.0809, + "step": 63 + }, + { + "epoch": 0.14578587699316628, + "grad_norm": 0.3699239492416382, + "learning_rate": 2.3863636363636367e-06, + "loss": 0.0794, + "step": 64 + }, + { + "epoch": 0.1480637813211845, + "grad_norm": 0.37682032585144043, + "learning_rate": 2.4242424242424244e-06, + "loss": 0.0805, + "step": 65 + }, + { + "epoch": 0.15034168564920272, + "grad_norm": 0.42901331186294556, + "learning_rate": 2.4621212121212125e-06, + "loss": 0.0804, + "step": 66 + }, + { + "epoch": 0.15261958997722094, + "grad_norm": 0.40458548069000244, + "learning_rate": 2.5e-06, + "loss": 0.0784, + "step": 67 + }, + { + "epoch": 0.1548974943052392, + "grad_norm": 0.4400278329849243, + "learning_rate": 2.537878787878788e-06, + "loss": 0.08, + "step": 68 + }, + { + "epoch": 0.1571753986332574, + "grad_norm": 0.4129616916179657, + "learning_rate": 2.575757575757576e-06, + "loss": 0.0773, + "step": 69 + }, + { + "epoch": 0.15945330296127563, + "grad_norm": 0.4329078793525696, + "learning_rate": 2.6136363636363637e-06, + "loss": 0.0771, + "step": 70 + }, + { + "epoch": 0.16173120728929385, + "grad_norm": 0.3519822657108307, + "learning_rate": 2.6515151515151514e-06, + "loss": 0.0754, + "step": 71 + }, + { + "epoch": 0.16400911161731208, + "grad_norm": 0.3959544897079468, + "learning_rate": 2.6893939393939395e-06, + "loss": 0.0769, + "step": 72 + }, + { + "epoch": 0.1662870159453303, + "grad_norm": 0.3731265366077423, + "learning_rate": 2.7272727272727272e-06, + "loss": 0.077, + "step": 73 + }, + { + "epoch": 0.16856492027334852, + "grad_norm": 0.43425431847572327, + "learning_rate": 2.7651515151515153e-06, + "loss": 0.0752, + "step": 74 + }, + { + "epoch": 0.17084282460136674, + "grad_norm": 0.3866608738899231, + "learning_rate": 2.803030303030303e-06, + "loss": 0.0768, + "step": 75 + }, + { + "epoch": 0.17312072892938496, + "grad_norm": 0.40310585498809814, + "learning_rate": 2.8409090909090916e-06, + "loss": 0.0741, + "step": 76 + }, + { + "epoch": 0.17539863325740318, + "grad_norm": 0.37993788719177246, + "learning_rate": 2.8787878787878793e-06, + "loss": 0.0761, + "step": 77 + }, + { + "epoch": 0.1776765375854214, + "grad_norm": 0.5747309327125549, + "learning_rate": 2.916666666666667e-06, + "loss": 0.0752, + "step": 78 + }, + { + "epoch": 0.17995444191343962, + "grad_norm": 0.3601266145706177, + "learning_rate": 2.954545454545455e-06, + "loss": 0.0732, + "step": 79 + }, + { + "epoch": 0.18223234624145787, + "grad_norm": 0.3578413128852844, + "learning_rate": 2.992424242424243e-06, + "loss": 0.0777, + "step": 80 + }, + { + "epoch": 0.1845102505694761, + "grad_norm": 0.3445342183113098, + "learning_rate": 3.0303030303030305e-06, + "loss": 0.077, + "step": 81 + }, + { + "epoch": 0.1867881548974943, + "grad_norm": 0.46273601055145264, + "learning_rate": 3.0681818181818186e-06, + "loss": 0.0743, + "step": 82 + }, + { + "epoch": 0.18906605922551253, + "grad_norm": 0.3583069443702698, + "learning_rate": 3.1060606060606063e-06, + "loss": 0.0735, + "step": 83 + }, + { + "epoch": 0.19134396355353075, + "grad_norm": 0.35333251953125, + "learning_rate": 3.143939393939394e-06, + "loss": 0.0705, + "step": 84 + }, + { + "epoch": 0.19362186788154898, + "grad_norm": 0.42324912548065186, + "learning_rate": 3.181818181818182e-06, + "loss": 0.0768, + "step": 85 + }, + { + "epoch": 0.1958997722095672, + "grad_norm": 0.38333430886268616, + "learning_rate": 3.21969696969697e-06, + "loss": 0.074, + "step": 86 + }, + { + "epoch": 0.19817767653758542, + "grad_norm": 0.382974773645401, + "learning_rate": 3.257575757575758e-06, + "loss": 0.0763, + "step": 87 + }, + { + "epoch": 0.20045558086560364, + "grad_norm": 0.3540959656238556, + "learning_rate": 3.2954545454545456e-06, + "loss": 0.0752, + "step": 88 + }, + { + "epoch": 0.20273348519362186, + "grad_norm": 0.44744303822517395, + "learning_rate": 3.3333333333333333e-06, + "loss": 0.0774, + "step": 89 + }, + { + "epoch": 0.20501138952164008, + "grad_norm": 0.38871949911117554, + "learning_rate": 3.3712121212121214e-06, + "loss": 0.074, + "step": 90 + }, + { + "epoch": 0.2072892938496583, + "grad_norm": 0.3690223693847656, + "learning_rate": 3.409090909090909e-06, + "loss": 0.0716, + "step": 91 + }, + { + "epoch": 0.20956719817767655, + "grad_norm": 0.4073256552219391, + "learning_rate": 3.4469696969696977e-06, + "loss": 0.0724, + "step": 92 + }, + { + "epoch": 0.21184510250569477, + "grad_norm": 0.4237968623638153, + "learning_rate": 3.4848484848484854e-06, + "loss": 0.0753, + "step": 93 + }, + { + "epoch": 0.214123006833713, + "grad_norm": 0.4609670341014862, + "learning_rate": 3.522727272727273e-06, + "loss": 0.0762, + "step": 94 + }, + { + "epoch": 0.2164009111617312, + "grad_norm": 0.3941766023635864, + "learning_rate": 3.560606060606061e-06, + "loss": 0.0743, + "step": 95 + }, + { + "epoch": 0.21867881548974943, + "grad_norm": 0.4086122512817383, + "learning_rate": 3.598484848484849e-06, + "loss": 0.0731, + "step": 96 + }, + { + "epoch": 0.22095671981776766, + "grad_norm": 0.42508289217948914, + "learning_rate": 3.6363636363636366e-06, + "loss": 0.0747, + "step": 97 + }, + { + "epoch": 0.22323462414578588, + "grad_norm": 0.3469661772251129, + "learning_rate": 3.6742424242424247e-06, + "loss": 0.0758, + "step": 98 + }, + { + "epoch": 0.2255125284738041, + "grad_norm": 2.2205488681793213, + "learning_rate": 3.7121212121212124e-06, + "loss": 0.0744, + "step": 99 + }, + { + "epoch": 0.22779043280182232, + "grad_norm": 0.5607733130455017, + "learning_rate": 3.7500000000000005e-06, + "loss": 0.0727, + "step": 100 + }, + { + "epoch": 0.23006833712984054, + "grad_norm": 0.3631141483783722, + "learning_rate": 3.7878787878787882e-06, + "loss": 0.0729, + "step": 101 + }, + { + "epoch": 0.23234624145785876, + "grad_norm": 0.41111883521080017, + "learning_rate": 3.825757575757576e-06, + "loss": 0.0707, + "step": 102 + }, + { + "epoch": 0.23462414578587698, + "grad_norm": 0.42842748761177063, + "learning_rate": 3.863636363636364e-06, + "loss": 0.0741, + "step": 103 + }, + { + "epoch": 0.23690205011389523, + "grad_norm": 0.5235817432403564, + "learning_rate": 3.901515151515151e-06, + "loss": 0.076, + "step": 104 + }, + { + "epoch": 0.23917995444191345, + "grad_norm": 0.47007620334625244, + "learning_rate": 3.93939393939394e-06, + "loss": 0.0738, + "step": 105 + }, + { + "epoch": 0.24145785876993167, + "grad_norm": 0.4875897765159607, + "learning_rate": 3.9772727272727275e-06, + "loss": 0.075, + "step": 106 + }, + { + "epoch": 0.2437357630979499, + "grad_norm": 0.482083797454834, + "learning_rate": 4.015151515151515e-06, + "loss": 0.073, + "step": 107 + }, + { + "epoch": 0.2460136674259681, + "grad_norm": 0.3963218033313751, + "learning_rate": 4.053030303030303e-06, + "loss": 0.0711, + "step": 108 + }, + { + "epoch": 0.24829157175398633, + "grad_norm": 0.45080500841140747, + "learning_rate": 4.0909090909090915e-06, + "loss": 0.0736, + "step": 109 + }, + { + "epoch": 0.2505694760820046, + "grad_norm": 0.4224866032600403, + "learning_rate": 4.128787878787879e-06, + "loss": 0.0729, + "step": 110 + }, + { + "epoch": 0.2528473804100228, + "grad_norm": 0.44626861810684204, + "learning_rate": 4.166666666666667e-06, + "loss": 0.0739, + "step": 111 + }, + { + "epoch": 0.255125284738041, + "grad_norm": 0.4718227684497833, + "learning_rate": 4.204545454545455e-06, + "loss": 0.0734, + "step": 112 + }, + { + "epoch": 0.25740318906605925, + "grad_norm": 0.43946653604507446, + "learning_rate": 4.242424242424243e-06, + "loss": 0.0738, + "step": 113 + }, + { + "epoch": 0.25968109339407747, + "grad_norm": 0.4009428918361664, + "learning_rate": 4.280303030303031e-06, + "loss": 0.0735, + "step": 114 + }, + { + "epoch": 0.2619589977220957, + "grad_norm": 0.4430725574493408, + "learning_rate": 4.3181818181818185e-06, + "loss": 0.0713, + "step": 115 + }, + { + "epoch": 0.2642369020501139, + "grad_norm": 0.42131391167640686, + "learning_rate": 4.356060606060606e-06, + "loss": 0.0715, + "step": 116 + }, + { + "epoch": 0.26651480637813213, + "grad_norm": 0.4587585926055908, + "learning_rate": 4.393939393939394e-06, + "loss": 0.0739, + "step": 117 + }, + { + "epoch": 0.26879271070615035, + "grad_norm": 0.42365217208862305, + "learning_rate": 4.4318181818181824e-06, + "loss": 0.0766, + "step": 118 + }, + { + "epoch": 0.27107061503416857, + "grad_norm": 0.48815441131591797, + "learning_rate": 4.46969696969697e-06, + "loss": 0.072, + "step": 119 + }, + { + "epoch": 0.2733485193621868, + "grad_norm": 0.4345132112503052, + "learning_rate": 4.507575757575758e-06, + "loss": 0.0746, + "step": 120 + }, + { + "epoch": 0.275626423690205, + "grad_norm": 0.40394189953804016, + "learning_rate": 4.5454545454545455e-06, + "loss": 0.0735, + "step": 121 + }, + { + "epoch": 0.27790432801822323, + "grad_norm": 0.40433070063591003, + "learning_rate": 4.583333333333333e-06, + "loss": 0.0715, + "step": 122 + }, + { + "epoch": 0.28018223234624146, + "grad_norm": 0.3981536626815796, + "learning_rate": 4.621212121212122e-06, + "loss": 0.0729, + "step": 123 + }, + { + "epoch": 0.2824601366742597, + "grad_norm": 0.42909497022628784, + "learning_rate": 4.6590909090909095e-06, + "loss": 0.0748, + "step": 124 + }, + { + "epoch": 0.2847380410022779, + "grad_norm": 0.3979114890098572, + "learning_rate": 4.696969696969698e-06, + "loss": 0.0706, + "step": 125 + }, + { + "epoch": 0.2870159453302961, + "grad_norm": 0.4175008237361908, + "learning_rate": 4.734848484848486e-06, + "loss": 0.0723, + "step": 126 + }, + { + "epoch": 0.28929384965831434, + "grad_norm": 0.4061114490032196, + "learning_rate": 4.772727272727273e-06, + "loss": 0.0741, + "step": 127 + }, + { + "epoch": 0.29157175398633256, + "grad_norm": 0.4005403518676758, + "learning_rate": 4.810606060606061e-06, + "loss": 0.074, + "step": 128 + }, + { + "epoch": 0.2938496583143508, + "grad_norm": 0.40054985880851746, + "learning_rate": 4.848484848484849e-06, + "loss": 0.0719, + "step": 129 + }, + { + "epoch": 0.296127562642369, + "grad_norm": 0.4266146421432495, + "learning_rate": 4.8863636363636365e-06, + "loss": 0.0736, + "step": 130 + }, + { + "epoch": 0.2984054669703872, + "grad_norm": 0.3816811442375183, + "learning_rate": 4.924242424242425e-06, + "loss": 0.0698, + "step": 131 + }, + { + "epoch": 0.30068337129840544, + "grad_norm": 0.3987654745578766, + "learning_rate": 4.962121212121213e-06, + "loss": 0.0703, + "step": 132 + }, + { + "epoch": 0.30296127562642367, + "grad_norm": 0.41395506262779236, + "learning_rate": 5e-06, + "loss": 0.0743, + "step": 133 + }, + { + "epoch": 0.3052391799544419, + "grad_norm": 0.41543126106262207, + "learning_rate": 5e-06, + "loss": 0.0734, + "step": 134 + }, + { + "epoch": 0.30751708428246016, + "grad_norm": 0.4777982532978058, + "learning_rate": 5e-06, + "loss": 0.0745, + "step": 135 + }, + { + "epoch": 0.3097949886104784, + "grad_norm": 0.47125476598739624, + "learning_rate": 5e-06, + "loss": 0.0732, + "step": 136 + }, + { + "epoch": 0.3120728929384966, + "grad_norm": 0.3936716616153717, + "learning_rate": 5e-06, + "loss": 0.0746, + "step": 137 + }, + { + "epoch": 0.3143507972665148, + "grad_norm": 0.4368305504322052, + "learning_rate": 5e-06, + "loss": 0.0731, + "step": 138 + }, + { + "epoch": 0.31662870159453305, + "grad_norm": 0.4058026671409607, + "learning_rate": 5e-06, + "loss": 0.0738, + "step": 139 + }, + { + "epoch": 0.31890660592255127, + "grad_norm": 0.4541865587234497, + "learning_rate": 5e-06, + "loss": 0.0735, + "step": 140 + }, + { + "epoch": 0.3211845102505695, + "grad_norm": 0.3929808437824249, + "learning_rate": 5e-06, + "loss": 0.0729, + "step": 141 + }, + { + "epoch": 0.3234624145785877, + "grad_norm": 0.4054822623729706, + "learning_rate": 5e-06, + "loss": 0.0746, + "step": 142 + }, + { + "epoch": 0.32574031890660593, + "grad_norm": 0.38358375430107117, + "learning_rate": 5e-06, + "loss": 0.0722, + "step": 143 + }, + { + "epoch": 0.32801822323462415, + "grad_norm": 0.41057437658309937, + "learning_rate": 5e-06, + "loss": 0.0697, + "step": 144 + }, + { + "epoch": 0.33029612756264237, + "grad_norm": 0.4309552013874054, + "learning_rate": 5e-06, + "loss": 0.0739, + "step": 145 + }, + { + "epoch": 0.3325740318906606, + "grad_norm": 0.38301971554756165, + "learning_rate": 5e-06, + "loss": 0.0713, + "step": 146 + }, + { + "epoch": 0.3348519362186788, + "grad_norm": 0.435709148645401, + "learning_rate": 5e-06, + "loss": 0.0705, + "step": 147 + }, + { + "epoch": 0.33712984054669703, + "grad_norm": 0.4956047832965851, + "learning_rate": 5e-06, + "loss": 0.0715, + "step": 148 + }, + { + "epoch": 0.33940774487471526, + "grad_norm": 0.4198216199874878, + "learning_rate": 5e-06, + "loss": 0.0724, + "step": 149 + }, + { + "epoch": 0.3416856492027335, + "grad_norm": 0.42688027024269104, + "learning_rate": 5e-06, + "loss": 0.0703, + "step": 150 + }, + { + "epoch": 0.3439635535307517, + "grad_norm": 0.4963231086730957, + "learning_rate": 5e-06, + "loss": 0.0708, + "step": 151 + }, + { + "epoch": 0.3462414578587699, + "grad_norm": 0.44305217266082764, + "learning_rate": 5e-06, + "loss": 0.0717, + "step": 152 + }, + { + "epoch": 0.34851936218678814, + "grad_norm": 0.5162036418914795, + "learning_rate": 5e-06, + "loss": 0.0735, + "step": 153 + }, + { + "epoch": 0.35079726651480636, + "grad_norm": 0.4345501959323883, + "learning_rate": 5e-06, + "loss": 0.0732, + "step": 154 + }, + { + "epoch": 0.3530751708428246, + "grad_norm": 0.3963910937309265, + "learning_rate": 5e-06, + "loss": 0.0722, + "step": 155 + }, + { + "epoch": 0.3553530751708428, + "grad_norm": 0.42344456911087036, + "learning_rate": 5e-06, + "loss": 0.0727, + "step": 156 + }, + { + "epoch": 0.357630979498861, + "grad_norm": 0.40760302543640137, + "learning_rate": 5e-06, + "loss": 0.0718, + "step": 157 + }, + { + "epoch": 0.35990888382687924, + "grad_norm": 0.5510509610176086, + "learning_rate": 5e-06, + "loss": 0.0731, + "step": 158 + }, + { + "epoch": 0.3621867881548975, + "grad_norm": 0.4162476062774658, + "learning_rate": 5e-06, + "loss": 0.0709, + "step": 159 + }, + { + "epoch": 0.36446469248291574, + "grad_norm": 0.5093443989753723, + "learning_rate": 5e-06, + "loss": 0.0692, + "step": 160 + }, + { + "epoch": 0.36674259681093396, + "grad_norm": 0.4493374228477478, + "learning_rate": 5e-06, + "loss": 0.071, + "step": 161 + }, + { + "epoch": 0.3690205011389522, + "grad_norm": 0.44762253761291504, + "learning_rate": 5e-06, + "loss": 0.072, + "step": 162 + }, + { + "epoch": 0.3712984054669704, + "grad_norm": 0.375567764043808, + "learning_rate": 5e-06, + "loss": 0.0745, + "step": 163 + }, + { + "epoch": 0.3735763097949886, + "grad_norm": 0.4188103973865509, + "learning_rate": 5e-06, + "loss": 0.0699, + "step": 164 + }, + { + "epoch": 0.37585421412300685, + "grad_norm": 0.37547120451927185, + "learning_rate": 5e-06, + "loss": 0.0721, + "step": 165 + }, + { + "epoch": 0.37813211845102507, + "grad_norm": 0.45365750789642334, + "learning_rate": 5e-06, + "loss": 0.0721, + "step": 166 + }, + { + "epoch": 0.3804100227790433, + "grad_norm": 0.4168195128440857, + "learning_rate": 5e-06, + "loss": 0.0717, + "step": 167 + }, + { + "epoch": 0.3826879271070615, + "grad_norm": 0.4386206269264221, + "learning_rate": 5e-06, + "loss": 0.0697, + "step": 168 + }, + { + "epoch": 0.38496583143507973, + "grad_norm": 0.46875110268592834, + "learning_rate": 5e-06, + "loss": 0.0712, + "step": 169 + }, + { + "epoch": 0.38724373576309795, + "grad_norm": 0.3938765823841095, + "learning_rate": 5e-06, + "loss": 0.0693, + "step": 170 + }, + { + "epoch": 0.3895216400911162, + "grad_norm": 0.4123467803001404, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 171 + }, + { + "epoch": 0.3917995444191344, + "grad_norm": 0.4382531940937042, + "learning_rate": 5e-06, + "loss": 0.0712, + "step": 172 + }, + { + "epoch": 0.3940774487471526, + "grad_norm": 0.4919157922267914, + "learning_rate": 5e-06, + "loss": 0.0729, + "step": 173 + }, + { + "epoch": 0.39635535307517084, + "grad_norm": 0.5063402056694031, + "learning_rate": 5e-06, + "loss": 0.0707, + "step": 174 + }, + { + "epoch": 0.39863325740318906, + "grad_norm": 0.6215523481369019, + "learning_rate": 5e-06, + "loss": 0.0699, + "step": 175 + }, + { + "epoch": 0.4009111617312073, + "grad_norm": 0.4623260200023651, + "learning_rate": 5e-06, + "loss": 0.0721, + "step": 176 + }, + { + "epoch": 0.4031890660592255, + "grad_norm": 0.5183009505271912, + "learning_rate": 5e-06, + "loss": 0.0731, + "step": 177 + }, + { + "epoch": 0.4054669703872437, + "grad_norm": 0.4948502480983734, + "learning_rate": 5e-06, + "loss": 0.0703, + "step": 178 + }, + { + "epoch": 0.40774487471526194, + "grad_norm": 0.4933330714702606, + "learning_rate": 5e-06, + "loss": 0.0711, + "step": 179 + }, + { + "epoch": 0.41002277904328016, + "grad_norm": 0.4790695607662201, + "learning_rate": 5e-06, + "loss": 0.0722, + "step": 180 + }, + { + "epoch": 0.4123006833712984, + "grad_norm": 0.4662487506866455, + "learning_rate": 5e-06, + "loss": 0.0708, + "step": 181 + }, + { + "epoch": 0.4145785876993166, + "grad_norm": 0.3859008252620697, + "learning_rate": 5e-06, + "loss": 0.0697, + "step": 182 + }, + { + "epoch": 0.4168564920273349, + "grad_norm": 0.6519958972930908, + "learning_rate": 5e-06, + "loss": 0.0701, + "step": 183 + }, + { + "epoch": 0.4191343963553531, + "grad_norm": 0.494468092918396, + "learning_rate": 5e-06, + "loss": 0.072, + "step": 184 + }, + { + "epoch": 0.4214123006833713, + "grad_norm": 0.5084004998207092, + "learning_rate": 5e-06, + "loss": 0.0713, + "step": 185 + }, + { + "epoch": 0.42369020501138954, + "grad_norm": 0.4281460642814636, + "learning_rate": 5e-06, + "loss": 0.0727, + "step": 186 + }, + { + "epoch": 0.42596810933940776, + "grad_norm": 0.4384775161743164, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 187 + }, + { + "epoch": 0.428246013667426, + "grad_norm": 0.4647282660007477, + "learning_rate": 5e-06, + "loss": 0.0714, + "step": 188 + }, + { + "epoch": 0.4305239179954442, + "grad_norm": 0.4845445156097412, + "learning_rate": 5e-06, + "loss": 0.071, + "step": 189 + }, + { + "epoch": 0.4328018223234624, + "grad_norm": 0.40431490540504456, + "learning_rate": 5e-06, + "loss": 0.0694, + "step": 190 + }, + { + "epoch": 0.43507972665148065, + "grad_norm": 0.4188959002494812, + "learning_rate": 5e-06, + "loss": 0.0674, + "step": 191 + }, + { + "epoch": 0.43735763097949887, + "grad_norm": 0.5054275989532471, + "learning_rate": 5e-06, + "loss": 0.0738, + "step": 192 + }, + { + "epoch": 0.4396355353075171, + "grad_norm": 0.438518226146698, + "learning_rate": 5e-06, + "loss": 0.0687, + "step": 193 + }, + { + "epoch": 0.4419134396355353, + "grad_norm": 0.5088305473327637, + "learning_rate": 5e-06, + "loss": 0.0681, + "step": 194 + }, + { + "epoch": 0.44419134396355353, + "grad_norm": 0.44068580865859985, + "learning_rate": 5e-06, + "loss": 0.0708, + "step": 195 + }, + { + "epoch": 0.44646924829157175, + "grad_norm": 0.4711444675922394, + "learning_rate": 5e-06, + "loss": 0.0749, + "step": 196 + }, + { + "epoch": 0.44874715261959, + "grad_norm": 0.42442336678504944, + "learning_rate": 5e-06, + "loss": 0.0666, + "step": 197 + }, + { + "epoch": 0.4510250569476082, + "grad_norm": 0.45719027519226074, + "learning_rate": 5e-06, + "loss": 0.0685, + "step": 198 + }, + { + "epoch": 0.4533029612756264, + "grad_norm": 0.4533730745315552, + "learning_rate": 5e-06, + "loss": 0.073, + "step": 199 + }, + { + "epoch": 0.45558086560364464, + "grad_norm": 0.3968890607357025, + "learning_rate": 5e-06, + "loss": 0.0729, + "step": 200 + }, + { + "epoch": 0.45785876993166286, + "grad_norm": 0.4367174804210663, + "learning_rate": 5e-06, + "loss": 0.068, + "step": 201 + }, + { + "epoch": 0.4601366742596811, + "grad_norm": 0.37111589312553406, + "learning_rate": 5e-06, + "loss": 0.0685, + "step": 202 + }, + { + "epoch": 0.4624145785876993, + "grad_norm": 0.45682379603385925, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 203 + }, + { + "epoch": 0.4646924829157175, + "grad_norm": 0.36972588300704956, + "learning_rate": 5e-06, + "loss": 0.0693, + "step": 204 + }, + { + "epoch": 0.46697038724373574, + "grad_norm": 0.36360761523246765, + "learning_rate": 5e-06, + "loss": 0.0695, + "step": 205 + }, + { + "epoch": 0.46924829157175396, + "grad_norm": 0.3721333146095276, + "learning_rate": 5e-06, + "loss": 0.0698, + "step": 206 + }, + { + "epoch": 0.4715261958997722, + "grad_norm": 0.40355461835861206, + "learning_rate": 5e-06, + "loss": 0.0741, + "step": 207 + }, + { + "epoch": 0.47380410022779046, + "grad_norm": 0.4323730766773224, + "learning_rate": 5e-06, + "loss": 0.0746, + "step": 208 + }, + { + "epoch": 0.4760820045558087, + "grad_norm": 0.38074442744255066, + "learning_rate": 5e-06, + "loss": 0.0699, + "step": 209 + }, + { + "epoch": 0.4783599088838269, + "grad_norm": 0.43709224462509155, + "learning_rate": 5e-06, + "loss": 0.0686, + "step": 210 + }, + { + "epoch": 0.4806378132118451, + "grad_norm": 0.3957338333129883, + "learning_rate": 5e-06, + "loss": 0.0703, + "step": 211 + }, + { + "epoch": 0.48291571753986334, + "grad_norm": 0.5257817506790161, + "learning_rate": 5e-06, + "loss": 0.0697, + "step": 212 + }, + { + "epoch": 0.48519362186788156, + "grad_norm": 0.3774060904979706, + "learning_rate": 5e-06, + "loss": 0.0679, + "step": 213 + }, + { + "epoch": 0.4874715261958998, + "grad_norm": 0.5191784501075745, + "learning_rate": 5e-06, + "loss": 0.0732, + "step": 214 + }, + { + "epoch": 0.489749430523918, + "grad_norm": 0.40021416544914246, + "learning_rate": 5e-06, + "loss": 0.0664, + "step": 215 + }, + { + "epoch": 0.4920273348519362, + "grad_norm": 0.44555339217185974, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 216 + }, + { + "epoch": 0.49430523917995445, + "grad_norm": 0.5284780859947205, + "learning_rate": 5e-06, + "loss": 0.0726, + "step": 217 + }, + { + "epoch": 0.49658314350797267, + "grad_norm": 0.4448579251766205, + "learning_rate": 5e-06, + "loss": 0.0711, + "step": 218 + }, + { + "epoch": 0.4988610478359909, + "grad_norm": 0.4625869393348694, + "learning_rate": 5e-06, + "loss": 0.0711, + "step": 219 + }, + { + "epoch": 0.5011389521640092, + "grad_norm": 0.46510323882102966, + "learning_rate": 5e-06, + "loss": 0.0736, + "step": 220 + }, + { + "epoch": 0.5034168564920274, + "grad_norm": 0.480757474899292, + "learning_rate": 5e-06, + "loss": 0.0696, + "step": 221 + }, + { + "epoch": 0.5056947608200456, + "grad_norm": 0.4454173445701599, + "learning_rate": 5e-06, + "loss": 0.0696, + "step": 222 + }, + { + "epoch": 0.5079726651480638, + "grad_norm": 0.44396576285362244, + "learning_rate": 5e-06, + "loss": 0.0696, + "step": 223 + }, + { + "epoch": 0.510250569476082, + "grad_norm": 0.4392300844192505, + "learning_rate": 5e-06, + "loss": 0.0695, + "step": 224 + }, + { + "epoch": 0.5125284738041003, + "grad_norm": 0.4110094904899597, + "learning_rate": 5e-06, + "loss": 0.0717, + "step": 225 + }, + { + "epoch": 0.5148063781321185, + "grad_norm": 0.41785213351249695, + "learning_rate": 5e-06, + "loss": 0.0728, + "step": 226 + }, + { + "epoch": 0.5170842824601367, + "grad_norm": 0.4551200568675995, + "learning_rate": 5e-06, + "loss": 0.0677, + "step": 227 + }, + { + "epoch": 0.5193621867881549, + "grad_norm": 0.41336214542388916, + "learning_rate": 5e-06, + "loss": 0.0697, + "step": 228 + }, + { + "epoch": 0.5216400911161732, + "grad_norm": 0.420012503862381, + "learning_rate": 5e-06, + "loss": 0.0703, + "step": 229 + }, + { + "epoch": 0.5239179954441914, + "grad_norm": 0.3874534070491791, + "learning_rate": 5e-06, + "loss": 0.0688, + "step": 230 + }, + { + "epoch": 0.5261958997722096, + "grad_norm": 0.44382932782173157, + "learning_rate": 5e-06, + "loss": 0.0709, + "step": 231 + }, + { + "epoch": 0.5284738041002278, + "grad_norm": 0.4071871042251587, + "learning_rate": 5e-06, + "loss": 0.0705, + "step": 232 + }, + { + "epoch": 0.530751708428246, + "grad_norm": 0.4357655644416809, + "learning_rate": 5e-06, + "loss": 0.072, + "step": 233 + }, + { + "epoch": 0.5330296127562643, + "grad_norm": 0.4081815183162689, + "learning_rate": 5e-06, + "loss": 0.0695, + "step": 234 + }, + { + "epoch": 0.5353075170842825, + "grad_norm": 0.38522443175315857, + "learning_rate": 5e-06, + "loss": 0.0685, + "step": 235 + }, + { + "epoch": 0.5375854214123007, + "grad_norm": 0.4070127308368683, + "learning_rate": 5e-06, + "loss": 0.0698, + "step": 236 + }, + { + "epoch": 0.5398633257403189, + "grad_norm": 0.43571823835372925, + "learning_rate": 5e-06, + "loss": 0.0714, + "step": 237 + }, + { + "epoch": 0.5421412300683371, + "grad_norm": 0.43545764684677124, + "learning_rate": 5e-06, + "loss": 0.0682, + "step": 238 + }, + { + "epoch": 0.5444191343963554, + "grad_norm": 0.37378817796707153, + "learning_rate": 5e-06, + "loss": 0.0681, + "step": 239 + }, + { + "epoch": 0.5466970387243736, + "grad_norm": 0.40321189165115356, + "learning_rate": 5e-06, + "loss": 0.0688, + "step": 240 + }, + { + "epoch": 0.5489749430523918, + "grad_norm": 0.4021066725254059, + "learning_rate": 5e-06, + "loss": 0.0683, + "step": 241 + }, + { + "epoch": 0.55125284738041, + "grad_norm": 0.3792532682418823, + "learning_rate": 5e-06, + "loss": 0.0679, + "step": 242 + }, + { + "epoch": 0.5535307517084282, + "grad_norm": 0.351995050907135, + "learning_rate": 5e-06, + "loss": 0.0701, + "step": 243 + }, + { + "epoch": 0.5558086560364465, + "grad_norm": 0.4324621558189392, + "learning_rate": 5e-06, + "loss": 0.0719, + "step": 244 + }, + { + "epoch": 0.5580865603644647, + "grad_norm": 0.4158404767513275, + "learning_rate": 5e-06, + "loss": 0.0689, + "step": 245 + }, + { + "epoch": 0.5603644646924829, + "grad_norm": 0.39611494541168213, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 246 + }, + { + "epoch": 0.5626423690205011, + "grad_norm": 0.4095311462879181, + "learning_rate": 5e-06, + "loss": 0.0712, + "step": 247 + }, + { + "epoch": 0.5649202733485194, + "grad_norm": 0.425353467464447, + "learning_rate": 5e-06, + "loss": 0.0708, + "step": 248 + }, + { + "epoch": 0.5671981776765376, + "grad_norm": 0.4120287299156189, + "learning_rate": 5e-06, + "loss": 0.0699, + "step": 249 + }, + { + "epoch": 0.5694760820045558, + "grad_norm": 0.42984098196029663, + "learning_rate": 5e-06, + "loss": 0.0695, + "step": 250 + }, + { + "epoch": 0.571753986332574, + "grad_norm": 0.41558778285980225, + "learning_rate": 5e-06, + "loss": 0.07, + "step": 251 + }, + { + "epoch": 0.5740318906605922, + "grad_norm": 0.39627233147621155, + "learning_rate": 5e-06, + "loss": 0.068, + "step": 252 + }, + { + "epoch": 0.5763097949886105, + "grad_norm": 0.4116947650909424, + "learning_rate": 5e-06, + "loss": 0.0675, + "step": 253 + }, + { + "epoch": 0.5785876993166287, + "grad_norm": 0.4483950436115265, + "learning_rate": 5e-06, + "loss": 0.0699, + "step": 254 + }, + { + "epoch": 0.5808656036446469, + "grad_norm": 0.40655770897865295, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 255 + }, + { + "epoch": 0.5831435079726651, + "grad_norm": 0.4380018413066864, + "learning_rate": 5e-06, + "loss": 0.0692, + "step": 256 + }, + { + "epoch": 0.5854214123006833, + "grad_norm": 0.37291720509529114, + "learning_rate": 5e-06, + "loss": 0.0697, + "step": 257 + }, + { + "epoch": 0.5876993166287016, + "grad_norm": 0.4471172094345093, + "learning_rate": 5e-06, + "loss": 0.0706, + "step": 258 + }, + { + "epoch": 0.5899772209567198, + "grad_norm": 0.33719295263290405, + "learning_rate": 5e-06, + "loss": 0.0704, + "step": 259 + }, + { + "epoch": 0.592255125284738, + "grad_norm": 0.4563951790332794, + "learning_rate": 5e-06, + "loss": 0.0704, + "step": 260 + }, + { + "epoch": 0.5945330296127562, + "grad_norm": 0.479675829410553, + "learning_rate": 5e-06, + "loss": 0.068, + "step": 261 + }, + { + "epoch": 0.5968109339407744, + "grad_norm": 0.48119309544563293, + "learning_rate": 5e-06, + "loss": 0.0704, + "step": 262 + }, + { + "epoch": 0.5990888382687927, + "grad_norm": 0.4188333749771118, + "learning_rate": 5e-06, + "loss": 0.0689, + "step": 263 + }, + { + "epoch": 0.6013667425968109, + "grad_norm": 0.4289758801460266, + "learning_rate": 5e-06, + "loss": 0.0698, + "step": 264 + }, + { + "epoch": 0.6036446469248291, + "grad_norm": 0.4562152922153473, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 265 + }, + { + "epoch": 0.6059225512528473, + "grad_norm": 0.3772522807121277, + "learning_rate": 5e-06, + "loss": 0.0688, + "step": 266 + }, + { + "epoch": 0.6082004555808656, + "grad_norm": 0.4561619460582733, + "learning_rate": 5e-06, + "loss": 0.0668, + "step": 267 + }, + { + "epoch": 0.6104783599088838, + "grad_norm": 0.3349689245223999, + "learning_rate": 5e-06, + "loss": 0.0686, + "step": 268 + }, + { + "epoch": 0.6127562642369021, + "grad_norm": 0.37530314922332764, + "learning_rate": 5e-06, + "loss": 0.0679, + "step": 269 + }, + { + "epoch": 0.6150341685649203, + "grad_norm": 0.4140493869781494, + "learning_rate": 5e-06, + "loss": 0.0664, + "step": 270 + }, + { + "epoch": 0.6173120728929385, + "grad_norm": 0.3943673074245453, + "learning_rate": 5e-06, + "loss": 0.0666, + "step": 271 + }, + { + "epoch": 0.6195899772209568, + "grad_norm": 0.4929119348526001, + "learning_rate": 5e-06, + "loss": 0.0694, + "step": 272 + }, + { + "epoch": 0.621867881548975, + "grad_norm": 0.3644123375415802, + "learning_rate": 5e-06, + "loss": 0.0673, + "step": 273 + }, + { + "epoch": 0.6241457858769932, + "grad_norm": 0.3780081570148468, + "learning_rate": 5e-06, + "loss": 0.0691, + "step": 274 + }, + { + "epoch": 0.6264236902050114, + "grad_norm": 0.3895011246204376, + "learning_rate": 5e-06, + "loss": 0.0673, + "step": 275 + }, + { + "epoch": 0.6287015945330297, + "grad_norm": 0.39944520592689514, + "learning_rate": 5e-06, + "loss": 0.0698, + "step": 276 + }, + { + "epoch": 0.6309794988610479, + "grad_norm": 0.5009689927101135, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 277 + }, + { + "epoch": 0.6332574031890661, + "grad_norm": 0.3824428915977478, + "learning_rate": 5e-06, + "loss": 0.0686, + "step": 278 + }, + { + "epoch": 0.6355353075170843, + "grad_norm": 0.40774327516555786, + "learning_rate": 5e-06, + "loss": 0.0683, + "step": 279 + }, + { + "epoch": 0.6378132118451025, + "grad_norm": 0.4349845051765442, + "learning_rate": 5e-06, + "loss": 0.0677, + "step": 280 + }, + { + "epoch": 0.6400911161731208, + "grad_norm": 0.3697284758090973, + "learning_rate": 5e-06, + "loss": 0.0691, + "step": 281 + }, + { + "epoch": 0.642369020501139, + "grad_norm": 0.42937129735946655, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 282 + }, + { + "epoch": 0.6446469248291572, + "grad_norm": 0.3947838544845581, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 283 + }, + { + "epoch": 0.6469248291571754, + "grad_norm": 0.435682088136673, + "learning_rate": 5e-06, + "loss": 0.0681, + "step": 284 + }, + { + "epoch": 0.6492027334851936, + "grad_norm": 0.4589371979236603, + "learning_rate": 5e-06, + "loss": 0.0688, + "step": 285 + }, + { + "epoch": 0.6514806378132119, + "grad_norm": 0.43639275431632996, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 286 + }, + { + "epoch": 0.6537585421412301, + "grad_norm": 0.4098499119281769, + "learning_rate": 5e-06, + "loss": 0.0669, + "step": 287 + }, + { + "epoch": 0.6560364464692483, + "grad_norm": 0.35208365321159363, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 288 + }, + { + "epoch": 0.6583143507972665, + "grad_norm": 0.34454038739204407, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 289 + }, + { + "epoch": 0.6605922551252847, + "grad_norm": 0.41582560539245605, + "learning_rate": 5e-06, + "loss": 0.0707, + "step": 290 + }, + { + "epoch": 0.662870159453303, + "grad_norm": 0.4000478684902191, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 291 + }, + { + "epoch": 0.6651480637813212, + "grad_norm": 0.3716413676738739, + "learning_rate": 5e-06, + "loss": 0.0707, + "step": 292 + }, + { + "epoch": 0.6674259681093394, + "grad_norm": 0.4950391352176666, + "learning_rate": 5e-06, + "loss": 0.0668, + "step": 293 + }, + { + "epoch": 0.6697038724373576, + "grad_norm": 0.40866413712501526, + "learning_rate": 5e-06, + "loss": 0.0697, + "step": 294 + }, + { + "epoch": 0.6719817767653758, + "grad_norm": 0.35762685537338257, + "learning_rate": 5e-06, + "loss": 0.0683, + "step": 295 + }, + { + "epoch": 0.6742596810933941, + "grad_norm": 0.37755048274993896, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 296 + }, + { + "epoch": 0.6765375854214123, + "grad_norm": 0.4419291913509369, + "learning_rate": 5e-06, + "loss": 0.0746, + "step": 297 + }, + { + "epoch": 0.6788154897494305, + "grad_norm": 0.42603951692581177, + "learning_rate": 5e-06, + "loss": 0.0705, + "step": 298 + }, + { + "epoch": 0.6810933940774487, + "grad_norm": 0.4297736585140228, + "learning_rate": 5e-06, + "loss": 0.0696, + "step": 299 + }, + { + "epoch": 0.683371298405467, + "grad_norm": 0.4051309823989868, + "learning_rate": 5e-06, + "loss": 0.069, + "step": 300 + }, + { + "epoch": 0.6856492027334852, + "grad_norm": 0.43650469183921814, + "learning_rate": 5e-06, + "loss": 0.0705, + "step": 301 + }, + { + "epoch": 0.6879271070615034, + "grad_norm": 0.4065714180469513, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 302 + }, + { + "epoch": 0.6902050113895216, + "grad_norm": 0.3967154026031494, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 303 + }, + { + "epoch": 0.6924829157175398, + "grad_norm": 0.4712061882019043, + "learning_rate": 5e-06, + "loss": 0.0685, + "step": 304 + }, + { + "epoch": 0.6947608200455581, + "grad_norm": 0.42462876439094543, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 305 + }, + { + "epoch": 0.6970387243735763, + "grad_norm": 0.41642263531684875, + "learning_rate": 5e-06, + "loss": 0.0701, + "step": 306 + }, + { + "epoch": 0.6993166287015945, + "grad_norm": 0.4268721044063568, + "learning_rate": 5e-06, + "loss": 0.0683, + "step": 307 + }, + { + "epoch": 0.7015945330296127, + "grad_norm": 0.4870758652687073, + "learning_rate": 5e-06, + "loss": 0.0701, + "step": 308 + }, + { + "epoch": 0.7038724373576309, + "grad_norm": 0.38001054525375366, + "learning_rate": 5e-06, + "loss": 0.0693, + "step": 309 + }, + { + "epoch": 0.7061503416856492, + "grad_norm": 0.44683194160461426, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 310 + }, + { + "epoch": 0.7084282460136674, + "grad_norm": 0.3841770589351654, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 311 + }, + { + "epoch": 0.7107061503416856, + "grad_norm": 0.41161251068115234, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 312 + }, + { + "epoch": 0.7129840546697038, + "grad_norm": 0.38489028811454773, + "learning_rate": 5e-06, + "loss": 0.07, + "step": 313 + }, + { + "epoch": 0.715261958997722, + "grad_norm": 0.3507677912712097, + "learning_rate": 5e-06, + "loss": 0.0682, + "step": 314 + }, + { + "epoch": 0.7175398633257403, + "grad_norm": 0.38104042410850525, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 315 + }, + { + "epoch": 0.7198177676537585, + "grad_norm": 0.3669570982456207, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 316 + }, + { + "epoch": 0.7220956719817767, + "grad_norm": 0.42886313796043396, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 317 + }, + { + "epoch": 0.724373576309795, + "grad_norm": 0.3710843324661255, + "learning_rate": 5e-06, + "loss": 0.0666, + "step": 318 + }, + { + "epoch": 0.7266514806378133, + "grad_norm": 0.44205132126808167, + "learning_rate": 5e-06, + "loss": 0.0683, + "step": 319 + }, + { + "epoch": 0.7289293849658315, + "grad_norm": 0.41439348459243774, + "learning_rate": 5e-06, + "loss": 0.0679, + "step": 320 + }, + { + "epoch": 0.7312072892938497, + "grad_norm": 0.3940696716308594, + "learning_rate": 5e-06, + "loss": 0.0677, + "step": 321 + }, + { + "epoch": 0.7334851936218679, + "grad_norm": 0.41583356261253357, + "learning_rate": 5e-06, + "loss": 0.0686, + "step": 322 + }, + { + "epoch": 0.7357630979498861, + "grad_norm": 0.39446648955345154, + "learning_rate": 5e-06, + "loss": 0.0703, + "step": 323 + }, + { + "epoch": 0.7380410022779044, + "grad_norm": 0.4365374445915222, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 324 + }, + { + "epoch": 0.7403189066059226, + "grad_norm": 0.36670970916748047, + "learning_rate": 5e-06, + "loss": 0.0701, + "step": 325 + }, + { + "epoch": 0.7425968109339408, + "grad_norm": 0.43367478251457214, + "learning_rate": 5e-06, + "loss": 0.0688, + "step": 326 + }, + { + "epoch": 0.744874715261959, + "grad_norm": 0.3726787567138672, + "learning_rate": 5e-06, + "loss": 0.0701, + "step": 327 + }, + { + "epoch": 0.7471526195899773, + "grad_norm": 0.4093555808067322, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 328 + }, + { + "epoch": 0.7494305239179955, + "grad_norm": 0.49851593375205994, + "learning_rate": 5e-06, + "loss": 0.0717, + "step": 329 + }, + { + "epoch": 0.7517084282460137, + "grad_norm": 0.38292035460472107, + "learning_rate": 5e-06, + "loss": 0.0693, + "step": 330 + }, + { + "epoch": 0.7539863325740319, + "grad_norm": 0.41123974323272705, + "learning_rate": 5e-06, + "loss": 0.07, + "step": 331 + }, + { + "epoch": 0.7562642369020501, + "grad_norm": 0.3536129295825958, + "learning_rate": 5e-06, + "loss": 0.0682, + "step": 332 + }, + { + "epoch": 0.7585421412300684, + "grad_norm": 0.38178107142448425, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 333 + }, + { + "epoch": 0.7608200455580866, + "grad_norm": 0.36010780930519104, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 334 + }, + { + "epoch": 0.7630979498861048, + "grad_norm": 0.4523884356021881, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 335 + }, + { + "epoch": 0.765375854214123, + "grad_norm": 0.39606744050979614, + "learning_rate": 5e-06, + "loss": 0.0685, + "step": 336 + }, + { + "epoch": 0.7676537585421412, + "grad_norm": 0.41211503744125366, + "learning_rate": 5e-06, + "loss": 0.0683, + "step": 337 + }, + { + "epoch": 0.7699316628701595, + "grad_norm": 0.41744863986968994, + "learning_rate": 5e-06, + "loss": 0.0681, + "step": 338 + }, + { + "epoch": 0.7722095671981777, + "grad_norm": 0.3365219235420227, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 339 + }, + { + "epoch": 0.7744874715261959, + "grad_norm": 0.4278775453567505, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 340 + }, + { + "epoch": 0.7767653758542141, + "grad_norm": 0.3654761016368866, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 341 + }, + { + "epoch": 0.7790432801822323, + "grad_norm": 0.4179307520389557, + "learning_rate": 5e-06, + "loss": 0.0695, + "step": 342 + }, + { + "epoch": 0.7813211845102506, + "grad_norm": 0.3700203597545624, + "learning_rate": 5e-06, + "loss": 0.0691, + "step": 343 + }, + { + "epoch": 0.7835990888382688, + "grad_norm": 0.41979873180389404, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 344 + }, + { + "epoch": 0.785876993166287, + "grad_norm": 0.3833925127983093, + "learning_rate": 5e-06, + "loss": 0.0685, + "step": 345 + }, + { + "epoch": 0.7881548974943052, + "grad_norm": 0.38170167803764343, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 346 + }, + { + "epoch": 0.7904328018223234, + "grad_norm": 0.4435146749019623, + "learning_rate": 5e-06, + "loss": 0.0686, + "step": 347 + }, + { + "epoch": 0.7927107061503417, + "grad_norm": 0.3727116584777832, + "learning_rate": 5e-06, + "loss": 0.0695, + "step": 348 + }, + { + "epoch": 0.7949886104783599, + "grad_norm": 0.3438148498535156, + "learning_rate": 5e-06, + "loss": 0.0675, + "step": 349 + }, + { + "epoch": 0.7972665148063781, + "grad_norm": 1.1323906183242798, + "learning_rate": 5e-06, + "loss": 0.0693, + "step": 350 + }, + { + "epoch": 0.7995444191343963, + "grad_norm": 0.49652302265167236, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 351 + }, + { + "epoch": 0.8018223234624146, + "grad_norm": 0.3763923943042755, + "learning_rate": 5e-06, + "loss": 0.0687, + "step": 352 + }, + { + "epoch": 0.8041002277904328, + "grad_norm": 0.3898470997810364, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 353 + }, + { + "epoch": 0.806378132118451, + "grad_norm": 0.43604037165641785, + "learning_rate": 5e-06, + "loss": 0.0699, + "step": 354 + }, + { + "epoch": 0.8086560364464692, + "grad_norm": 0.40902185440063477, + "learning_rate": 5e-06, + "loss": 0.0708, + "step": 355 + }, + { + "epoch": 0.8109339407744874, + "grad_norm": 0.4095119535923004, + "learning_rate": 5e-06, + "loss": 0.0703, + "step": 356 + }, + { + "epoch": 0.8132118451025057, + "grad_norm": 0.38673949241638184, + "learning_rate": 5e-06, + "loss": 0.0673, + "step": 357 + }, + { + "epoch": 0.8154897494305239, + "grad_norm": 0.3702470362186432, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 358 + }, + { + "epoch": 0.8177676537585421, + "grad_norm": 0.38289743661880493, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 359 + }, + { + "epoch": 0.8200455580865603, + "grad_norm": 0.398942232131958, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 360 + }, + { + "epoch": 0.8223234624145785, + "grad_norm": 0.4990338683128357, + "learning_rate": 5e-06, + "loss": 0.0689, + "step": 361 + }, + { + "epoch": 0.8246013667425968, + "grad_norm": 0.4140332341194153, + "learning_rate": 5e-06, + "loss": 0.0702, + "step": 362 + }, + { + "epoch": 0.826879271070615, + "grad_norm": 0.38353094458580017, + "learning_rate": 5e-06, + "loss": 0.068, + "step": 363 + }, + { + "epoch": 0.8291571753986332, + "grad_norm": 0.40966200828552246, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 364 + }, + { + "epoch": 0.8314350797266514, + "grad_norm": 0.40145012736320496, + "learning_rate": 5e-06, + "loss": 0.0674, + "step": 365 + }, + { + "epoch": 0.8337129840546698, + "grad_norm": 0.3553916811943054, + "learning_rate": 5e-06, + "loss": 0.0693, + "step": 366 + }, + { + "epoch": 0.835990888382688, + "grad_norm": 0.36585506796836853, + "learning_rate": 5e-06, + "loss": 0.0682, + "step": 367 + }, + { + "epoch": 0.8382687927107062, + "grad_norm": 0.40545061230659485, + "learning_rate": 5e-06, + "loss": 0.0698, + "step": 368 + }, + { + "epoch": 0.8405466970387244, + "grad_norm": 0.43122202157974243, + "learning_rate": 5e-06, + "loss": 0.07, + "step": 369 + }, + { + "epoch": 0.8428246013667426, + "grad_norm": 0.38822853565216064, + "learning_rate": 5e-06, + "loss": 0.0682, + "step": 370 + }, + { + "epoch": 0.8451025056947609, + "grad_norm": 0.36941057443618774, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 371 + }, + { + "epoch": 0.8473804100227791, + "grad_norm": 0.3812001049518585, + "learning_rate": 5e-06, + "loss": 0.0679, + "step": 372 + }, + { + "epoch": 0.8496583143507973, + "grad_norm": 0.37349483370780945, + "learning_rate": 5e-06, + "loss": 0.0687, + "step": 373 + }, + { + "epoch": 0.8519362186788155, + "grad_norm": 0.36606496572494507, + "learning_rate": 5e-06, + "loss": 0.0664, + "step": 374 + }, + { + "epoch": 0.8542141230068337, + "grad_norm": 0.4783948063850403, + "learning_rate": 5e-06, + "loss": 0.0697, + "step": 375 + }, + { + "epoch": 0.856492027334852, + "grad_norm": 0.42428654432296753, + "learning_rate": 5e-06, + "loss": 0.0673, + "step": 376 + }, + { + "epoch": 0.8587699316628702, + "grad_norm": 0.43024080991744995, + "learning_rate": 5e-06, + "loss": 0.0667, + "step": 377 + }, + { + "epoch": 0.8610478359908884, + "grad_norm": 0.45276060700416565, + "learning_rate": 5e-06, + "loss": 0.068, + "step": 378 + }, + { + "epoch": 0.8633257403189066, + "grad_norm": 0.5208924412727356, + "learning_rate": 5e-06, + "loss": 0.0671, + "step": 379 + }, + { + "epoch": 0.8656036446469249, + "grad_norm": 0.44442853331565857, + "learning_rate": 5e-06, + "loss": 0.0693, + "step": 380 + }, + { + "epoch": 0.8678815489749431, + "grad_norm": 0.4459463059902191, + "learning_rate": 5e-06, + "loss": 0.0708, + "step": 381 + }, + { + "epoch": 0.8701594533029613, + "grad_norm": 0.4089488387107849, + "learning_rate": 5e-06, + "loss": 0.0687, + "step": 382 + }, + { + "epoch": 0.8724373576309795, + "grad_norm": 0.4216194450855255, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 383 + }, + { + "epoch": 0.8747152619589977, + "grad_norm": 0.4460054337978363, + "learning_rate": 5e-06, + "loss": 0.0673, + "step": 384 + }, + { + "epoch": 0.876993166287016, + "grad_norm": 0.3685241937637329, + "learning_rate": 5e-06, + "loss": 0.0669, + "step": 385 + }, + { + "epoch": 0.8792710706150342, + "grad_norm": 0.5342898964881897, + "learning_rate": 5e-06, + "loss": 0.0619, + "step": 386 + }, + { + "epoch": 0.8815489749430524, + "grad_norm": 0.41996005177497864, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 387 + }, + { + "epoch": 0.8838268792710706, + "grad_norm": 0.5424495339393616, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 388 + }, + { + "epoch": 0.8861047835990888, + "grad_norm": 0.5055522918701172, + "learning_rate": 5e-06, + "loss": 0.0674, + "step": 389 + }, + { + "epoch": 0.8883826879271071, + "grad_norm": 0.48916786909103394, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 390 + }, + { + "epoch": 0.8906605922551253, + "grad_norm": 0.4740985333919525, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 391 + }, + { + "epoch": 0.8929384965831435, + "grad_norm": 0.5370116829872131, + "learning_rate": 5e-06, + "loss": 0.0666, + "step": 392 + }, + { + "epoch": 0.8952164009111617, + "grad_norm": 0.45419377088546753, + "learning_rate": 5e-06, + "loss": 0.0688, + "step": 393 + }, + { + "epoch": 0.89749430523918, + "grad_norm": 0.4498198330402374, + "learning_rate": 5e-06, + "loss": 0.0666, + "step": 394 + }, + { + "epoch": 0.8997722095671982, + "grad_norm": 0.47344449162483215, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 395 + }, + { + "epoch": 0.9020501138952164, + "grad_norm": 0.49295154213905334, + "learning_rate": 5e-06, + "loss": 0.0695, + "step": 396 + }, + { + "epoch": 0.9043280182232346, + "grad_norm": 0.49694356322288513, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 397 + }, + { + "epoch": 0.9066059225512528, + "grad_norm": 0.4109736680984497, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 398 + }, + { + "epoch": 0.908883826879271, + "grad_norm": 0.386333703994751, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 399 + }, + { + "epoch": 0.9111617312072893, + "grad_norm": 0.4430575966835022, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 400 + }, + { + "epoch": 0.9134396355353075, + "grad_norm": 0.34944966435432434, + "learning_rate": 5e-06, + "loss": 0.0706, + "step": 401 + }, + { + "epoch": 0.9157175398633257, + "grad_norm": 0.4361039400100708, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 402 + }, + { + "epoch": 0.9179954441913439, + "grad_norm": 0.41876447200775146, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 403 + }, + { + "epoch": 0.9202733485193622, + "grad_norm": 0.46392524242401123, + "learning_rate": 5e-06, + "loss": 0.0698, + "step": 404 + }, + { + "epoch": 0.9225512528473804, + "grad_norm": 0.36828112602233887, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 405 + }, + { + "epoch": 0.9248291571753986, + "grad_norm": 0.445583313703537, + "learning_rate": 5e-06, + "loss": 0.0686, + "step": 406 + }, + { + "epoch": 0.9271070615034168, + "grad_norm": 0.4569609463214874, + "learning_rate": 5e-06, + "loss": 0.0669, + "step": 407 + }, + { + "epoch": 0.929384965831435, + "grad_norm": 0.4351450800895691, + "learning_rate": 5e-06, + "loss": 0.0667, + "step": 408 + }, + { + "epoch": 0.9316628701594533, + "grad_norm": 0.4705968499183655, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 409 + }, + { + "epoch": 0.9339407744874715, + "grad_norm": 0.3710295855998993, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 410 + }, + { + "epoch": 0.9362186788154897, + "grad_norm": 0.46435174345970154, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 411 + }, + { + "epoch": 0.9384965831435079, + "grad_norm": 0.35733482241630554, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 412 + }, + { + "epoch": 0.9407744874715261, + "grad_norm": 0.4195706844329834, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 413 + }, + { + "epoch": 0.9430523917995444, + "grad_norm": 0.3877922594547272, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 414 + }, + { + "epoch": 0.9453302961275627, + "grad_norm": 0.3729753792285919, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 415 + }, + { + "epoch": 0.9476082004555809, + "grad_norm": 0.38233113288879395, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 416 + }, + { + "epoch": 0.9498861047835991, + "grad_norm": 0.34748148918151855, + "learning_rate": 5e-06, + "loss": 0.0692, + "step": 417 + }, + { + "epoch": 0.9521640091116174, + "grad_norm": 0.38823238015174866, + "learning_rate": 5e-06, + "loss": 0.068, + "step": 418 + }, + { + "epoch": 0.9544419134396356, + "grad_norm": 0.38967302441596985, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 419 + }, + { + "epoch": 0.9567198177676538, + "grad_norm": 0.3796670734882355, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 420 + }, + { + "epoch": 0.958997722095672, + "grad_norm": 0.375175803899765, + "learning_rate": 5e-06, + "loss": 0.0687, + "step": 421 + }, + { + "epoch": 0.9612756264236902, + "grad_norm": 0.39777424931526184, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 422 + }, + { + "epoch": 0.9635535307517085, + "grad_norm": 0.3519769310951233, + "learning_rate": 5e-06, + "loss": 0.0654, + "step": 423 + }, + { + "epoch": 0.9658314350797267, + "grad_norm": 0.39732250571250916, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 424 + }, + { + "epoch": 0.9681093394077449, + "grad_norm": 0.3979857861995697, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 425 + }, + { + "epoch": 0.9703872437357631, + "grad_norm": 0.4009624421596527, + "learning_rate": 5e-06, + "loss": 0.0677, + "step": 426 + }, + { + "epoch": 0.9726651480637813, + "grad_norm": 0.37647587060928345, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 427 + }, + { + "epoch": 0.9749430523917996, + "grad_norm": 0.4034980833530426, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 428 + }, + { + "epoch": 0.9772209567198178, + "grad_norm": 0.3745870292186737, + "learning_rate": 5e-06, + "loss": 0.0692, + "step": 429 + }, + { + "epoch": 0.979498861047836, + "grad_norm": 0.3924770653247833, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 430 + }, + { + "epoch": 0.9817767653758542, + "grad_norm": 0.3894212543964386, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 431 + }, + { + "epoch": 0.9840546697038725, + "grad_norm": 0.4379626214504242, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 432 + }, + { + "epoch": 0.9863325740318907, + "grad_norm": 0.41487187147140503, + "learning_rate": 5e-06, + "loss": 0.0679, + "step": 433 + }, + { + "epoch": 0.9886104783599089, + "grad_norm": 0.3902760446071625, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 434 + }, + { + "epoch": 0.9908883826879271, + "grad_norm": 0.4039607644081116, + "learning_rate": 5e-06, + "loss": 0.0704, + "step": 435 + }, + { + "epoch": 0.9931662870159453, + "grad_norm": 0.46864795684814453, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 436 + }, + { + "epoch": 0.9954441913439636, + "grad_norm": 0.4453410804271698, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 437 + }, + { + "epoch": 0.9977220956719818, + "grad_norm": 0.41671499609947205, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 438 + }, + { + "epoch": 1.0, + "grad_norm": 0.43622660636901855, + "learning_rate": 5e-06, + "loss": 0.0679, + "step": 439 + }, + { + "epoch": 1.0022779043280183, + "grad_norm": 0.4500711262226105, + "learning_rate": 5e-06, + "loss": 0.0694, + "step": 440 + }, + { + "epoch": 1.0045558086560364, + "grad_norm": 0.4106467664241791, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 441 + }, + { + "epoch": 1.0068337129840548, + "grad_norm": 0.3629504144191742, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 442 + }, + { + "epoch": 1.0091116173120729, + "grad_norm": 0.3577280640602112, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 443 + }, + { + "epoch": 1.0113895216400912, + "grad_norm": 0.4010712206363678, + "learning_rate": 5e-06, + "loss": 0.0698, + "step": 444 + }, + { + "epoch": 1.0136674259681093, + "grad_norm": 0.4344230890274048, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 445 + }, + { + "epoch": 1.0159453302961277, + "grad_norm": 0.36966174840927124, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 446 + }, + { + "epoch": 1.0182232346241458, + "grad_norm": 0.40414756536483765, + "learning_rate": 5e-06, + "loss": 0.0673, + "step": 447 + }, + { + "epoch": 1.020501138952164, + "grad_norm": 0.4263600707054138, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 448 + }, + { + "epoch": 1.0227790432801822, + "grad_norm": 0.3861621618270874, + "learning_rate": 5e-06, + "loss": 0.0689, + "step": 449 + }, + { + "epoch": 1.0250569476082005, + "grad_norm": 0.4103226363658905, + "learning_rate": 5e-06, + "loss": 0.0633, + "step": 450 + }, + { + "epoch": 1.0273348519362187, + "grad_norm": 0.5162451863288879, + "learning_rate": 5e-06, + "loss": 0.0677, + "step": 451 + }, + { + "epoch": 1.029612756264237, + "grad_norm": 0.3653396964073181, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 452 + }, + { + "epoch": 1.031890660592255, + "grad_norm": 0.40646302700042725, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 453 + }, + { + "epoch": 1.0341685649202734, + "grad_norm": 0.351541668176651, + "learning_rate": 5e-06, + "loss": 0.0685, + "step": 454 + }, + { + "epoch": 1.0364464692482915, + "grad_norm": 0.45820608735084534, + "learning_rate": 5e-06, + "loss": 0.0671, + "step": 455 + }, + { + "epoch": 1.0387243735763099, + "grad_norm": 0.36796748638153076, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 456 + }, + { + "epoch": 1.041002277904328, + "grad_norm": 0.41809725761413574, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 457 + }, + { + "epoch": 1.0432801822323463, + "grad_norm": 0.4381527006626129, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 458 + }, + { + "epoch": 1.0455580865603644, + "grad_norm": 0.34027010202407837, + "learning_rate": 5e-06, + "loss": 0.0673, + "step": 459 + }, + { + "epoch": 1.0478359908883828, + "grad_norm": 0.3854738175868988, + "learning_rate": 5e-06, + "loss": 0.0667, + "step": 460 + }, + { + "epoch": 1.0501138952164009, + "grad_norm": 0.33832982182502747, + "learning_rate": 5e-06, + "loss": 0.0623, + "step": 461 + }, + { + "epoch": 1.0523917995444192, + "grad_norm": 0.3581103980541229, + "learning_rate": 5e-06, + "loss": 0.0681, + "step": 462 + }, + { + "epoch": 1.0546697038724373, + "grad_norm": 0.3820573687553406, + "learning_rate": 5e-06, + "loss": 0.0667, + "step": 463 + }, + { + "epoch": 1.0569476082004556, + "grad_norm": 0.38710519671440125, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 464 + }, + { + "epoch": 1.0592255125284737, + "grad_norm": 0.45680227875709534, + "learning_rate": 5e-06, + "loss": 0.0675, + "step": 465 + }, + { + "epoch": 1.061503416856492, + "grad_norm": 0.4713359773159027, + "learning_rate": 5e-06, + "loss": 0.0672, + "step": 466 + }, + { + "epoch": 1.0637813211845102, + "grad_norm": 0.4612472355365753, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 467 + }, + { + "epoch": 1.0660592255125285, + "grad_norm": 0.3581363558769226, + "learning_rate": 5e-06, + "loss": 0.0664, + "step": 468 + }, + { + "epoch": 1.0683371298405466, + "grad_norm": 0.6403517723083496, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 469 + }, + { + "epoch": 1.070615034168565, + "grad_norm": 0.6075398921966553, + "learning_rate": 5e-06, + "loss": 0.0667, + "step": 470 + }, + { + "epoch": 1.072892938496583, + "grad_norm": 0.37106701731681824, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 471 + }, + { + "epoch": 1.0751708428246014, + "grad_norm": 0.4145383834838867, + "learning_rate": 5e-06, + "loss": 0.0674, + "step": 472 + }, + { + "epoch": 1.0774487471526195, + "grad_norm": 0.4051216244697571, + "learning_rate": 5e-06, + "loss": 0.0686, + "step": 473 + }, + { + "epoch": 1.0797266514806378, + "grad_norm": 0.4450092017650604, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 474 + }, + { + "epoch": 1.082004555808656, + "grad_norm": 0.6834749579429626, + "learning_rate": 5e-06, + "loss": 0.0683, + "step": 475 + }, + { + "epoch": 1.0842824601366743, + "grad_norm": 0.4596191346645355, + "learning_rate": 5e-06, + "loss": 0.068, + "step": 476 + }, + { + "epoch": 1.0865603644646924, + "grad_norm": 0.35578453540802, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 477 + }, + { + "epoch": 1.0888382687927107, + "grad_norm": 0.38176682591438293, + "learning_rate": 5e-06, + "loss": 0.0696, + "step": 478 + }, + { + "epoch": 1.0911161731207288, + "grad_norm": 0.4199659526348114, + "learning_rate": 5e-06, + "loss": 0.0681, + "step": 479 + }, + { + "epoch": 1.0933940774487472, + "grad_norm": 0.39762362837791443, + "learning_rate": 5e-06, + "loss": 0.0707, + "step": 480 + }, + { + "epoch": 1.0956719817767653, + "grad_norm": 0.3889635503292084, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 481 + }, + { + "epoch": 1.0979498861047836, + "grad_norm": 0.3593704402446747, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 482 + }, + { + "epoch": 1.1002277904328017, + "grad_norm": 0.36741095781326294, + "learning_rate": 5e-06, + "loss": 0.0691, + "step": 483 + }, + { + "epoch": 1.10250569476082, + "grad_norm": 0.3881998062133789, + "learning_rate": 5e-06, + "loss": 0.0669, + "step": 484 + }, + { + "epoch": 1.1047835990888384, + "grad_norm": 0.4464372992515564, + "learning_rate": 5e-06, + "loss": 0.0654, + "step": 485 + }, + { + "epoch": 1.1070615034168565, + "grad_norm": 0.431099534034729, + "learning_rate": 5e-06, + "loss": 0.0671, + "step": 486 + }, + { + "epoch": 1.1093394077448746, + "grad_norm": 0.3904043436050415, + "learning_rate": 5e-06, + "loss": 0.0679, + "step": 487 + }, + { + "epoch": 1.111617312072893, + "grad_norm": 0.3737926781177521, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 488 + }, + { + "epoch": 1.1138952164009113, + "grad_norm": 0.3733069896697998, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 489 + }, + { + "epoch": 1.1161731207289294, + "grad_norm": 0.4064606726169586, + "learning_rate": 5e-06, + "loss": 0.0689, + "step": 490 + }, + { + "epoch": 1.1184510250569477, + "grad_norm": 0.3892678916454315, + "learning_rate": 5e-06, + "loss": 0.0638, + "step": 491 + }, + { + "epoch": 1.1207289293849658, + "grad_norm": 0.40777847170829773, + "learning_rate": 5e-06, + "loss": 0.0679, + "step": 492 + }, + { + "epoch": 1.1230068337129842, + "grad_norm": 0.3872988522052765, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 493 + }, + { + "epoch": 1.1252847380410023, + "grad_norm": 0.42812952399253845, + "learning_rate": 5e-06, + "loss": 0.0664, + "step": 494 + }, + { + "epoch": 1.1275626423690206, + "grad_norm": 0.33735978603363037, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 495 + }, + { + "epoch": 1.1298405466970387, + "grad_norm": 0.40792304277420044, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 496 + }, + { + "epoch": 1.132118451025057, + "grad_norm": 0.3538738191127777, + "learning_rate": 5e-06, + "loss": 0.0671, + "step": 497 + }, + { + "epoch": 1.1343963553530751, + "grad_norm": 0.40359169244766235, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 498 + }, + { + "epoch": 1.1366742596810935, + "grad_norm": 0.4550337791442871, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 499 + }, + { + "epoch": 1.1389521640091116, + "grad_norm": 0.37589704990386963, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 500 + }, + { + "epoch": 1.14123006833713, + "grad_norm": 0.5502781867980957, + "learning_rate": 5e-06, + "loss": 0.0685, + "step": 501 + }, + { + "epoch": 1.143507972665148, + "grad_norm": 0.3509303629398346, + "learning_rate": 5e-06, + "loss": 0.0679, + "step": 502 + }, + { + "epoch": 1.1457858769931664, + "grad_norm": 0.36012589931488037, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 503 + }, + { + "epoch": 1.1480637813211845, + "grad_norm": 0.40766096115112305, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 504 + }, + { + "epoch": 1.1503416856492028, + "grad_norm": 0.35230013728141785, + "learning_rate": 5e-06, + "loss": 0.0671, + "step": 505 + }, + { + "epoch": 1.152619589977221, + "grad_norm": 0.3493717610836029, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 506 + }, + { + "epoch": 1.1548974943052392, + "grad_norm": 0.34316879510879517, + "learning_rate": 5e-06, + "loss": 0.0664, + "step": 507 + }, + { + "epoch": 1.1571753986332574, + "grad_norm": 0.37918758392333984, + "learning_rate": 5e-06, + "loss": 0.07, + "step": 508 + }, + { + "epoch": 1.1594533029612757, + "grad_norm": 0.3596925437450409, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 509 + }, + { + "epoch": 1.1617312072892938, + "grad_norm": 0.36058440804481506, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 510 + }, + { + "epoch": 1.1640091116173121, + "grad_norm": 0.37516024708747864, + "learning_rate": 5e-06, + "loss": 0.0677, + "step": 511 + }, + { + "epoch": 1.1662870159453302, + "grad_norm": 0.39049404859542847, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 512 + }, + { + "epoch": 1.1685649202733486, + "grad_norm": 0.3826717734336853, + "learning_rate": 5e-06, + "loss": 0.0668, + "step": 513 + }, + { + "epoch": 1.1708428246013667, + "grad_norm": 0.3979628384113312, + "learning_rate": 5e-06, + "loss": 0.0677, + "step": 514 + }, + { + "epoch": 1.173120728929385, + "grad_norm": 0.36135396361351013, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 515 + }, + { + "epoch": 1.1753986332574031, + "grad_norm": 0.3812471032142639, + "learning_rate": 5e-06, + "loss": 0.0683, + "step": 516 + }, + { + "epoch": 1.1776765375854215, + "grad_norm": 0.381881982088089, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 517 + }, + { + "epoch": 1.1799544419134396, + "grad_norm": 0.40275490283966064, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 518 + }, + { + "epoch": 1.182232346241458, + "grad_norm": 0.35566964745521545, + "learning_rate": 5e-06, + "loss": 0.0671, + "step": 519 + }, + { + "epoch": 1.184510250569476, + "grad_norm": 0.4010874927043915, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 520 + }, + { + "epoch": 1.1867881548974943, + "grad_norm": 0.35122379660606384, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 521 + }, + { + "epoch": 1.1890660592255125, + "grad_norm": 0.38552436232566833, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 522 + }, + { + "epoch": 1.1913439635535308, + "grad_norm": 0.36752209067344666, + "learning_rate": 5e-06, + "loss": 0.0651, + "step": 523 + }, + { + "epoch": 1.193621867881549, + "grad_norm": 0.3583572208881378, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 524 + }, + { + "epoch": 1.1958997722095672, + "grad_norm": 0.34054288268089294, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 525 + }, + { + "epoch": 1.1981776765375853, + "grad_norm": 0.4100807309150696, + "learning_rate": 5e-06, + "loss": 0.0668, + "step": 526 + }, + { + "epoch": 1.2004555808656037, + "grad_norm": 0.3639017641544342, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 527 + }, + { + "epoch": 1.2027334851936218, + "grad_norm": 0.47511011362075806, + "learning_rate": 5e-06, + "loss": 0.0672, + "step": 528 + }, + { + "epoch": 1.20501138952164, + "grad_norm": 0.4454311728477478, + "learning_rate": 5e-06, + "loss": 0.0691, + "step": 529 + }, + { + "epoch": 1.2072892938496582, + "grad_norm": 0.38880598545074463, + "learning_rate": 5e-06, + "loss": 0.069, + "step": 530 + }, + { + "epoch": 1.2095671981776766, + "grad_norm": 0.3639177978038788, + "learning_rate": 5e-06, + "loss": 0.0668, + "step": 531 + }, + { + "epoch": 1.2118451025056949, + "grad_norm": 0.36371222138404846, + "learning_rate": 5e-06, + "loss": 0.0664, + "step": 532 + }, + { + "epoch": 1.214123006833713, + "grad_norm": 0.3781519830226898, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 533 + }, + { + "epoch": 1.216400911161731, + "grad_norm": 0.4092641770839691, + "learning_rate": 5e-06, + "loss": 0.0683, + "step": 534 + }, + { + "epoch": 1.2186788154897494, + "grad_norm": 0.4048423171043396, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 535 + }, + { + "epoch": 1.2209567198177678, + "grad_norm": 0.40468040108680725, + "learning_rate": 5e-06, + "loss": 0.068, + "step": 536 + }, + { + "epoch": 1.2232346241457859, + "grad_norm": 0.38839131593704224, + "learning_rate": 5e-06, + "loss": 0.0667, + "step": 537 + }, + { + "epoch": 1.225512528473804, + "grad_norm": 0.35276639461517334, + "learning_rate": 5e-06, + "loss": 0.0633, + "step": 538 + }, + { + "epoch": 1.2277904328018223, + "grad_norm": 0.43698349595069885, + "learning_rate": 5e-06, + "loss": 0.0693, + "step": 539 + }, + { + "epoch": 1.2300683371298406, + "grad_norm": 0.4079034626483917, + "learning_rate": 5e-06, + "loss": 0.0672, + "step": 540 + }, + { + "epoch": 1.2323462414578588, + "grad_norm": 0.40131327509880066, + "learning_rate": 5e-06, + "loss": 0.0691, + "step": 541 + }, + { + "epoch": 1.2346241457858769, + "grad_norm": 0.3789280652999878, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 542 + }, + { + "epoch": 1.2369020501138952, + "grad_norm": 0.31505241990089417, + "learning_rate": 5e-06, + "loss": 0.0667, + "step": 543 + }, + { + "epoch": 1.2391799544419135, + "grad_norm": 0.40908434987068176, + "learning_rate": 5e-06, + "loss": 0.0705, + "step": 544 + }, + { + "epoch": 1.2414578587699316, + "grad_norm": 0.39963164925575256, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 545 + }, + { + "epoch": 1.24373576309795, + "grad_norm": 0.41629716753959656, + "learning_rate": 5e-06, + "loss": 0.0677, + "step": 546 + }, + { + "epoch": 1.246013667425968, + "grad_norm": 0.384478896856308, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 547 + }, + { + "epoch": 1.2482915717539864, + "grad_norm": 0.3786999583244324, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 548 + }, + { + "epoch": 1.2505694760820045, + "grad_norm": 0.40297600626945496, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 549 + }, + { + "epoch": 1.2528473804100229, + "grad_norm": 0.40436241030693054, + "learning_rate": 5e-06, + "loss": 0.068, + "step": 550 + }, + { + "epoch": 1.255125284738041, + "grad_norm": 0.42469269037246704, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 551 + }, + { + "epoch": 1.2574031890660593, + "grad_norm": 0.3578304052352905, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 552 + }, + { + "epoch": 1.2596810933940774, + "grad_norm": 0.45142272114753723, + "learning_rate": 5e-06, + "loss": 0.0688, + "step": 553 + }, + { + "epoch": 1.2619589977220957, + "grad_norm": 0.36942797899246216, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 554 + }, + { + "epoch": 1.2642369020501139, + "grad_norm": 0.4410291314125061, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 555 + }, + { + "epoch": 1.2665148063781322, + "grad_norm": 0.376847505569458, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 556 + }, + { + "epoch": 1.2687927107061503, + "grad_norm": 0.4184603989124298, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 557 + }, + { + "epoch": 1.2710706150341686, + "grad_norm": 0.35660773515701294, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 558 + }, + { + "epoch": 1.2733485193621867, + "grad_norm": 0.41945064067840576, + "learning_rate": 5e-06, + "loss": 0.0666, + "step": 559 + }, + { + "epoch": 1.275626423690205, + "grad_norm": 0.3595651686191559, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 560 + }, + { + "epoch": 1.2779043280182232, + "grad_norm": 0.43288135528564453, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 561 + }, + { + "epoch": 1.2801822323462415, + "grad_norm": 0.38666829466819763, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 562 + }, + { + "epoch": 1.2824601366742596, + "grad_norm": 0.40617242455482483, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 563 + }, + { + "epoch": 1.284738041002278, + "grad_norm": 0.46813103556632996, + "learning_rate": 5e-06, + "loss": 0.0693, + "step": 564 + }, + { + "epoch": 1.287015945330296, + "grad_norm": 0.36550068855285645, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 565 + }, + { + "epoch": 1.2892938496583144, + "grad_norm": 0.4070315659046173, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 566 + }, + { + "epoch": 1.2915717539863325, + "grad_norm": 0.35947465896606445, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 567 + }, + { + "epoch": 1.2938496583143508, + "grad_norm": 0.36455580592155457, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 568 + }, + { + "epoch": 1.296127562642369, + "grad_norm": 0.38757917284965515, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 569 + }, + { + "epoch": 1.2984054669703873, + "grad_norm": 0.3300890326499939, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 570 + }, + { + "epoch": 1.3006833712984054, + "grad_norm": 0.3766978085041046, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 571 + }, + { + "epoch": 1.3029612756264237, + "grad_norm": 0.37804752588272095, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 572 + }, + { + "epoch": 1.3052391799544418, + "grad_norm": 0.3786337077617645, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 573 + }, + { + "epoch": 1.3075170842824602, + "grad_norm": 0.3882148265838623, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 574 + }, + { + "epoch": 1.3097949886104785, + "grad_norm": 3.220677137374878, + "learning_rate": 5e-06, + "loss": 0.0848, + "step": 575 + }, + { + "epoch": 1.3120728929384966, + "grad_norm": 0.5298126339912415, + "learning_rate": 5e-06, + "loss": 0.0679, + "step": 576 + }, + { + "epoch": 1.3143507972665147, + "grad_norm": 0.3402281701564789, + "learning_rate": 5e-06, + "loss": 0.0666, + "step": 577 + }, + { + "epoch": 1.316628701594533, + "grad_norm": 0.5069661140441895, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 578 + }, + { + "epoch": 1.3189066059225514, + "grad_norm": 0.3968346416950226, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 579 + }, + { + "epoch": 1.3211845102505695, + "grad_norm": 0.5330082774162292, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 580 + }, + { + "epoch": 1.3234624145785876, + "grad_norm": 0.4296810030937195, + "learning_rate": 5e-06, + "loss": 0.0674, + "step": 581 + }, + { + "epoch": 1.325740318906606, + "grad_norm": 0.39723390340805054, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 582 + }, + { + "epoch": 1.3280182232346243, + "grad_norm": 0.432210773229599, + "learning_rate": 5e-06, + "loss": 0.0664, + "step": 583 + }, + { + "epoch": 1.3302961275626424, + "grad_norm": 0.4127228558063507, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 584 + }, + { + "epoch": 1.3325740318906605, + "grad_norm": 0.43230441212654114, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 585 + }, + { + "epoch": 1.3348519362186788, + "grad_norm": 0.4305982291698456, + "learning_rate": 5e-06, + "loss": 0.0664, + "step": 586 + }, + { + "epoch": 1.3371298405466971, + "grad_norm": 0.3959517180919647, + "learning_rate": 5e-06, + "loss": 0.0667, + "step": 587 + }, + { + "epoch": 1.3394077448747153, + "grad_norm": 0.3866342306137085, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 588 + }, + { + "epoch": 1.3416856492027334, + "grad_norm": 0.43050622940063477, + "learning_rate": 5e-06, + "loss": 0.0631, + "step": 589 + }, + { + "epoch": 1.3439635535307517, + "grad_norm": 0.35574769973754883, + "learning_rate": 5e-06, + "loss": 0.0672, + "step": 590 + }, + { + "epoch": 1.34624145785877, + "grad_norm": 0.4432176351547241, + "learning_rate": 5e-06, + "loss": 0.0692, + "step": 591 + }, + { + "epoch": 1.3485193621867881, + "grad_norm": 0.3702254593372345, + "learning_rate": 5e-06, + "loss": 0.0687, + "step": 592 + }, + { + "epoch": 1.3507972665148062, + "grad_norm": 0.39732804894447327, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 593 + }, + { + "epoch": 1.3530751708428246, + "grad_norm": 0.39644646644592285, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 594 + }, + { + "epoch": 1.355353075170843, + "grad_norm": 0.3526952564716339, + "learning_rate": 5e-06, + "loss": 0.0668, + "step": 595 + }, + { + "epoch": 1.357630979498861, + "grad_norm": 0.4213428199291229, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 596 + }, + { + "epoch": 1.3599088838268791, + "grad_norm": 0.37950488924980164, + "learning_rate": 5e-06, + "loss": 0.0628, + "step": 597 + }, + { + "epoch": 1.3621867881548975, + "grad_norm": 0.42429694533348083, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 598 + }, + { + "epoch": 1.3644646924829158, + "grad_norm": 0.36994752287864685, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 599 + }, + { + "epoch": 1.366742596810934, + "grad_norm": 0.43309274315834045, + "learning_rate": 5e-06, + "loss": 0.0674, + "step": 600 + }, + { + "epoch": 1.3690205011389522, + "grad_norm": 0.39048635959625244, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 601 + }, + { + "epoch": 1.3712984054669703, + "grad_norm": 0.42135119438171387, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 602 + }, + { + "epoch": 1.3735763097949887, + "grad_norm": 0.36672961711883545, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 603 + }, + { + "epoch": 1.3758542141230068, + "grad_norm": 0.3776116371154785, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 604 + }, + { + "epoch": 1.3781321184510251, + "grad_norm": 0.34916606545448303, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 605 + }, + { + "epoch": 1.3804100227790432, + "grad_norm": 0.36569520831108093, + "learning_rate": 5e-06, + "loss": 0.0683, + "step": 606 + }, + { + "epoch": 1.3826879271070616, + "grad_norm": 0.3608874976634979, + "learning_rate": 5e-06, + "loss": 0.0673, + "step": 607 + }, + { + "epoch": 1.3849658314350797, + "grad_norm": 0.350178062915802, + "learning_rate": 5e-06, + "loss": 0.0654, + "step": 608 + }, + { + "epoch": 1.387243735763098, + "grad_norm": 0.37865710258483887, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 609 + }, + { + "epoch": 1.3895216400911161, + "grad_norm": 0.34312868118286133, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 610 + }, + { + "epoch": 1.3917995444191344, + "grad_norm": 0.35455793142318726, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 611 + }, + { + "epoch": 1.3940774487471526, + "grad_norm": 0.39417535066604614, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 612 + }, + { + "epoch": 1.396355353075171, + "grad_norm": 0.3636500835418701, + "learning_rate": 5e-06, + "loss": 0.0671, + "step": 613 + }, + { + "epoch": 1.398633257403189, + "grad_norm": 0.34348249435424805, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 614 + }, + { + "epoch": 1.4009111617312073, + "grad_norm": 0.3714514672756195, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 615 + }, + { + "epoch": 1.4031890660592254, + "grad_norm": 0.3414510488510132, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 616 + }, + { + "epoch": 1.4054669703872438, + "grad_norm": 0.38482949137687683, + "learning_rate": 5e-06, + "loss": 0.0686, + "step": 617 + }, + { + "epoch": 1.4077448747152619, + "grad_norm": 0.36240047216415405, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 618 + }, + { + "epoch": 1.4100227790432802, + "grad_norm": 0.39666205644607544, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 619 + }, + { + "epoch": 1.4123006833712983, + "grad_norm": 0.40982839465141296, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 620 + }, + { + "epoch": 1.4145785876993167, + "grad_norm": 0.38487517833709717, + "learning_rate": 5e-06, + "loss": 0.0682, + "step": 621 + }, + { + "epoch": 1.416856492027335, + "grad_norm": 0.35109570622444153, + "learning_rate": 5e-06, + "loss": 0.0664, + "step": 622 + }, + { + "epoch": 1.419134396355353, + "grad_norm": 0.4212898015975952, + "learning_rate": 5e-06, + "loss": 0.0651, + "step": 623 + }, + { + "epoch": 1.4214123006833712, + "grad_norm": 0.3583887815475464, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 624 + }, + { + "epoch": 1.4236902050113895, + "grad_norm": 0.4153973162174225, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 625 + }, + { + "epoch": 1.4259681093394079, + "grad_norm": 0.3374607264995575, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 626 + }, + { + "epoch": 1.428246013667426, + "grad_norm": 0.4959464371204376, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 627 + }, + { + "epoch": 1.430523917995444, + "grad_norm": 0.36800554394721985, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 628 + }, + { + "epoch": 1.4328018223234624, + "grad_norm": 0.3965972065925598, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 629 + }, + { + "epoch": 1.4350797266514808, + "grad_norm": 0.39917150139808655, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 630 + }, + { + "epoch": 1.4373576309794989, + "grad_norm": 0.41251757740974426, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 631 + }, + { + "epoch": 1.439635535307517, + "grad_norm": 0.43166816234588623, + "learning_rate": 5e-06, + "loss": 0.0633, + "step": 632 + }, + { + "epoch": 1.4419134396355353, + "grad_norm": 0.368181437253952, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 633 + }, + { + "epoch": 1.4441913439635536, + "grad_norm": 0.37732475996017456, + "learning_rate": 5e-06, + "loss": 0.0688, + "step": 634 + }, + { + "epoch": 1.4464692482915718, + "grad_norm": 0.3955388069152832, + "learning_rate": 5e-06, + "loss": 0.0654, + "step": 635 + }, + { + "epoch": 1.4487471526195899, + "grad_norm": 0.37553057074546814, + "learning_rate": 5e-06, + "loss": 0.063, + "step": 636 + }, + { + "epoch": 1.4510250569476082, + "grad_norm": 0.43095558881759644, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 637 + }, + { + "epoch": 1.4533029612756265, + "grad_norm": 0.38773244619369507, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 638 + }, + { + "epoch": 1.4555808656036446, + "grad_norm": 0.39239251613616943, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 639 + }, + { + "epoch": 1.4578587699316627, + "grad_norm": 0.3327813148498535, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 640 + }, + { + "epoch": 1.460136674259681, + "grad_norm": 0.40319737792015076, + "learning_rate": 5e-06, + "loss": 0.069, + "step": 641 + }, + { + "epoch": 1.4624145785876994, + "grad_norm": 0.34045639634132385, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 642 + }, + { + "epoch": 1.4646924829157175, + "grad_norm": 0.3521750867366791, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 643 + }, + { + "epoch": 1.4669703872437356, + "grad_norm": 0.45059895515441895, + "learning_rate": 5e-06, + "loss": 0.0666, + "step": 644 + }, + { + "epoch": 1.469248291571754, + "grad_norm": 0.3266962468624115, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 645 + }, + { + "epoch": 1.4715261958997723, + "grad_norm": 0.3925187885761261, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 646 + }, + { + "epoch": 1.4738041002277904, + "grad_norm": 0.36219561100006104, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 647 + }, + { + "epoch": 1.4760820045558087, + "grad_norm": 0.4496672749519348, + "learning_rate": 5e-06, + "loss": 0.0696, + "step": 648 + }, + { + "epoch": 1.4783599088838268, + "grad_norm": 0.37146294116973877, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 649 + }, + { + "epoch": 1.4806378132118452, + "grad_norm": 0.3783736824989319, + "learning_rate": 5e-06, + "loss": 0.0695, + "step": 650 + }, + { + "epoch": 1.4829157175398633, + "grad_norm": 0.3593132495880127, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 651 + }, + { + "epoch": 1.4851936218678816, + "grad_norm": 0.3631625771522522, + "learning_rate": 5e-06, + "loss": 0.0666, + "step": 652 + }, + { + "epoch": 1.4874715261958997, + "grad_norm": 0.37941277027130127, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 653 + }, + { + "epoch": 1.489749430523918, + "grad_norm": 0.37117141485214233, + "learning_rate": 5e-06, + "loss": 0.0687, + "step": 654 + }, + { + "epoch": 1.4920273348519362, + "grad_norm": 0.3478919267654419, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 655 + }, + { + "epoch": 1.4943052391799545, + "grad_norm": 0.341808557510376, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 656 + }, + { + "epoch": 1.4965831435079726, + "grad_norm": 0.3450130224227905, + "learning_rate": 5e-06, + "loss": 0.0651, + "step": 657 + }, + { + "epoch": 1.498861047835991, + "grad_norm": 0.3433770537376404, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 658 + }, + { + "epoch": 1.501138952164009, + "grad_norm": 0.33789244294166565, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 659 + }, + { + "epoch": 1.5034168564920274, + "grad_norm": 0.36195793747901917, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 660 + }, + { + "epoch": 1.5056947608200457, + "grad_norm": 0.33813396096229553, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 661 + }, + { + "epoch": 1.5079726651480638, + "grad_norm": 0.343996524810791, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 662 + }, + { + "epoch": 1.510250569476082, + "grad_norm": 0.360070139169693, + "learning_rate": 5e-06, + "loss": 0.0686, + "step": 663 + }, + { + "epoch": 1.5125284738041003, + "grad_norm": 0.3460543751716614, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 664 + }, + { + "epoch": 1.5148063781321186, + "grad_norm": 0.34158506989479065, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 665 + }, + { + "epoch": 1.5170842824601367, + "grad_norm": 0.35960525274276733, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 666 + }, + { + "epoch": 1.5193621867881548, + "grad_norm": 0.4116029143333435, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 667 + }, + { + "epoch": 1.5216400911161732, + "grad_norm": 0.33416780829429626, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 668 + }, + { + "epoch": 1.5239179954441915, + "grad_norm": 0.37547293305397034, + "learning_rate": 5e-06, + "loss": 0.0621, + "step": 669 + }, + { + "epoch": 1.5261958997722096, + "grad_norm": 0.3883625268936157, + "learning_rate": 5e-06, + "loss": 0.063, + "step": 670 + }, + { + "epoch": 1.5284738041002277, + "grad_norm": 0.394598126411438, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 671 + }, + { + "epoch": 1.530751708428246, + "grad_norm": 0.388380229473114, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 672 + }, + { + "epoch": 1.5330296127562644, + "grad_norm": 0.4153890609741211, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 673 + }, + { + "epoch": 1.5353075170842825, + "grad_norm": 0.33149614930152893, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 674 + }, + { + "epoch": 1.5375854214123006, + "grad_norm": 0.35577189922332764, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 675 + }, + { + "epoch": 1.539863325740319, + "grad_norm": 0.3430947959423065, + "learning_rate": 5e-06, + "loss": 0.0667, + "step": 676 + }, + { + "epoch": 1.5421412300683373, + "grad_norm": 0.3969891667366028, + "learning_rate": 5e-06, + "loss": 0.0625, + "step": 677 + }, + { + "epoch": 1.5444191343963554, + "grad_norm": 0.34465450048446655, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 678 + }, + { + "epoch": 1.5466970387243735, + "grad_norm": 0.4321921169757843, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 679 + }, + { + "epoch": 1.5489749430523918, + "grad_norm": 0.4178912937641144, + "learning_rate": 5e-06, + "loss": 0.0673, + "step": 680 + }, + { + "epoch": 1.5512528473804101, + "grad_norm": 0.4607953727245331, + "learning_rate": 5e-06, + "loss": 0.0626, + "step": 681 + }, + { + "epoch": 1.5535307517084282, + "grad_norm": 0.36263149976730347, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 682 + }, + { + "epoch": 1.5558086560364464, + "grad_norm": 0.4377477765083313, + "learning_rate": 5e-06, + "loss": 0.0666, + "step": 683 + }, + { + "epoch": 1.5580865603644647, + "grad_norm": 0.40092626214027405, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 684 + }, + { + "epoch": 1.560364464692483, + "grad_norm": 0.4000115990638733, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 685 + }, + { + "epoch": 1.5626423690205011, + "grad_norm": 0.3874872326850891, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 686 + }, + { + "epoch": 1.5649202733485192, + "grad_norm": 0.40038931369781494, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 687 + }, + { + "epoch": 1.5671981776765376, + "grad_norm": 0.39067986607551575, + "learning_rate": 5e-06, + "loss": 0.0666, + "step": 688 + }, + { + "epoch": 1.569476082004556, + "grad_norm": 0.4196305572986603, + "learning_rate": 5e-06, + "loss": 0.0688, + "step": 689 + }, + { + "epoch": 1.571753986332574, + "grad_norm": 0.36991575360298157, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 690 + }, + { + "epoch": 1.5740318906605921, + "grad_norm": 0.3553236424922943, + "learning_rate": 5e-06, + "loss": 0.0634, + "step": 691 + }, + { + "epoch": 1.5763097949886105, + "grad_norm": 0.354350209236145, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 692 + }, + { + "epoch": 1.5785876993166288, + "grad_norm": 0.420137494802475, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 693 + }, + { + "epoch": 1.580865603644647, + "grad_norm": 0.32134318351745605, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 694 + }, + { + "epoch": 1.583143507972665, + "grad_norm": 0.367403507232666, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 695 + }, + { + "epoch": 1.5854214123006833, + "grad_norm": 0.3804835081100464, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 696 + }, + { + "epoch": 1.5876993166287017, + "grad_norm": 0.38060566782951355, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 697 + }, + { + "epoch": 1.5899772209567198, + "grad_norm": 0.40258821845054626, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 698 + }, + { + "epoch": 1.592255125284738, + "grad_norm": 0.3835340738296509, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 699 + }, + { + "epoch": 1.5945330296127562, + "grad_norm": 0.3699604272842407, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 700 + }, + { + "epoch": 1.5968109339407746, + "grad_norm": 0.4031272828578949, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 701 + }, + { + "epoch": 1.5990888382687927, + "grad_norm": 0.3443854749202728, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 702 + }, + { + "epoch": 1.6013667425968108, + "grad_norm": 0.41966134309768677, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 703 + }, + { + "epoch": 1.603644646924829, + "grad_norm": 0.3769599199295044, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 704 + }, + { + "epoch": 1.6059225512528474, + "grad_norm": 0.40910017490386963, + "learning_rate": 5e-06, + "loss": 0.0627, + "step": 705 + }, + { + "epoch": 1.6082004555808656, + "grad_norm": 0.370617151260376, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 706 + }, + { + "epoch": 1.6104783599088837, + "grad_norm": 0.3756926655769348, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 707 + }, + { + "epoch": 1.6127562642369022, + "grad_norm": 0.3458310067653656, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 708 + }, + { + "epoch": 1.6150341685649203, + "grad_norm": 0.4030172824859619, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 709 + }, + { + "epoch": 1.6173120728929384, + "grad_norm": 0.32781001925468445, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 710 + }, + { + "epoch": 1.6195899772209568, + "grad_norm": 0.3912848234176636, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 711 + }, + { + "epoch": 1.621867881548975, + "grad_norm": 0.326858788728714, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 712 + }, + { + "epoch": 1.6241457858769932, + "grad_norm": 0.40474191308021545, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 713 + }, + { + "epoch": 1.6264236902050113, + "grad_norm": 0.3906795084476471, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 714 + }, + { + "epoch": 1.6287015945330297, + "grad_norm": 0.3630123734474182, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 715 + }, + { + "epoch": 1.630979498861048, + "grad_norm": 0.3990236222743988, + "learning_rate": 5e-06, + "loss": 0.0677, + "step": 716 + }, + { + "epoch": 1.633257403189066, + "grad_norm": 0.35970017313957214, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 717 + }, + { + "epoch": 1.6355353075170842, + "grad_norm": 0.4213522672653198, + "learning_rate": 5e-06, + "loss": 0.063, + "step": 718 + }, + { + "epoch": 1.6378132118451025, + "grad_norm": 0.3456747829914093, + "learning_rate": 5e-06, + "loss": 0.0628, + "step": 719 + }, + { + "epoch": 1.6400911161731209, + "grad_norm": 0.4727102816104889, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 720 + }, + { + "epoch": 1.642369020501139, + "grad_norm": 0.3923806846141815, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 721 + }, + { + "epoch": 1.644646924829157, + "grad_norm": 0.3525092303752899, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 722 + }, + { + "epoch": 1.6469248291571754, + "grad_norm": 0.3721219301223755, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 723 + }, + { + "epoch": 1.6492027334851938, + "grad_norm": 0.3733888268470764, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 724 + }, + { + "epoch": 1.6514806378132119, + "grad_norm": 0.39531344175338745, + "learning_rate": 5e-06, + "loss": 0.0675, + "step": 725 + }, + { + "epoch": 1.65375854214123, + "grad_norm": 0.4157196879386902, + "learning_rate": 5e-06, + "loss": 0.0654, + "step": 726 + }, + { + "epoch": 1.6560364464692483, + "grad_norm": 0.3935493230819702, + "learning_rate": 5e-06, + "loss": 0.0669, + "step": 727 + }, + { + "epoch": 1.6583143507972666, + "grad_norm": 0.43975120782852173, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 728 + }, + { + "epoch": 1.6605922551252847, + "grad_norm": 0.40202975273132324, + "learning_rate": 5e-06, + "loss": 0.0614, + "step": 729 + }, + { + "epoch": 1.6628701594533029, + "grad_norm": 0.3833255171775818, + "learning_rate": 5e-06, + "loss": 0.0629, + "step": 730 + }, + { + "epoch": 1.6651480637813212, + "grad_norm": 0.4237973690032959, + "learning_rate": 5e-06, + "loss": 0.0694, + "step": 731 + }, + { + "epoch": 1.6674259681093395, + "grad_norm": 0.41160348057746887, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 732 + }, + { + "epoch": 1.6697038724373576, + "grad_norm": 0.4225656986236572, + "learning_rate": 5e-06, + "loss": 0.0654, + "step": 733 + }, + { + "epoch": 1.6719817767653757, + "grad_norm": 0.40820473432540894, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 734 + }, + { + "epoch": 1.674259681093394, + "grad_norm": 0.4059019982814789, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 735 + }, + { + "epoch": 1.6765375854214124, + "grad_norm": 0.3632740378379822, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 736 + }, + { + "epoch": 1.6788154897494305, + "grad_norm": 0.3080376982688904, + "learning_rate": 5e-06, + "loss": 0.0627, + "step": 737 + }, + { + "epoch": 1.6810933940774486, + "grad_norm": 0.36947083473205566, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 738 + }, + { + "epoch": 1.683371298405467, + "grad_norm": 0.3585936725139618, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 739 + }, + { + "epoch": 1.6856492027334853, + "grad_norm": 0.4264557957649231, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 740 + }, + { + "epoch": 1.6879271070615034, + "grad_norm": 0.3662192225456238, + "learning_rate": 5e-06, + "loss": 0.0668, + "step": 741 + }, + { + "epoch": 1.6902050113895215, + "grad_norm": 0.34586918354034424, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 742 + }, + { + "epoch": 1.6924829157175398, + "grad_norm": 0.3758877217769623, + "learning_rate": 5e-06, + "loss": 0.0654, + "step": 743 + }, + { + "epoch": 1.6947608200455582, + "grad_norm": 0.37783950567245483, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 744 + }, + { + "epoch": 1.6970387243735763, + "grad_norm": 0.48236802220344543, + "learning_rate": 5e-06, + "loss": 0.0629, + "step": 745 + }, + { + "epoch": 1.6993166287015944, + "grad_norm": 0.3640000522136688, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 746 + }, + { + "epoch": 1.7015945330296127, + "grad_norm": 0.3304060995578766, + "learning_rate": 5e-06, + "loss": 0.0628, + "step": 747 + }, + { + "epoch": 1.703872437357631, + "grad_norm": 0.42156603932380676, + "learning_rate": 5e-06, + "loss": 0.069, + "step": 748 + }, + { + "epoch": 1.7061503416856492, + "grad_norm": 0.4093463122844696, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 749 + }, + { + "epoch": 1.7084282460136673, + "grad_norm": 0.3772518038749695, + "learning_rate": 5e-06, + "loss": 0.0669, + "step": 750 + }, + { + "epoch": 1.7107061503416856, + "grad_norm": 0.3777291476726532, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 751 + }, + { + "epoch": 1.712984054669704, + "grad_norm": 0.32958298921585083, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 752 + }, + { + "epoch": 1.715261958997722, + "grad_norm": 0.3931344151496887, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 753 + }, + { + "epoch": 1.7175398633257402, + "grad_norm": 0.38946473598480225, + "learning_rate": 5e-06, + "loss": 0.0683, + "step": 754 + }, + { + "epoch": 1.7198177676537585, + "grad_norm": 0.38754236698150635, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 755 + }, + { + "epoch": 1.7220956719817768, + "grad_norm": 0.358132541179657, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 756 + }, + { + "epoch": 1.724373576309795, + "grad_norm": 0.3571512997150421, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 757 + }, + { + "epoch": 1.7266514806378133, + "grad_norm": 0.41262051463127136, + "learning_rate": 5e-06, + "loss": 0.0673, + "step": 758 + }, + { + "epoch": 1.7289293849658316, + "grad_norm": 0.3680334687232971, + "learning_rate": 5e-06, + "loss": 0.0686, + "step": 759 + }, + { + "epoch": 1.7312072892938497, + "grad_norm": 0.38827383518218994, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 760 + }, + { + "epoch": 1.7334851936218678, + "grad_norm": 0.36443886160850525, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 761 + }, + { + "epoch": 1.7357630979498861, + "grad_norm": 0.44589725136756897, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 762 + }, + { + "epoch": 1.7380410022779045, + "grad_norm": 0.36552733182907104, + "learning_rate": 5e-06, + "loss": 0.0668, + "step": 763 + }, + { + "epoch": 1.7403189066059226, + "grad_norm": 0.3776485025882721, + "learning_rate": 5e-06, + "loss": 0.0669, + "step": 764 + }, + { + "epoch": 1.7425968109339407, + "grad_norm": 0.4614730477333069, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 765 + }, + { + "epoch": 1.744874715261959, + "grad_norm": 0.3474021553993225, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 766 + }, + { + "epoch": 1.7471526195899774, + "grad_norm": 0.4213907718658447, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 767 + }, + { + "epoch": 1.7494305239179955, + "grad_norm": 0.32314637303352356, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 768 + }, + { + "epoch": 1.7517084282460136, + "grad_norm": 0.3781552016735077, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 769 + }, + { + "epoch": 1.753986332574032, + "grad_norm": 0.40325456857681274, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 770 + }, + { + "epoch": 1.7562642369020502, + "grad_norm": 0.394195556640625, + "learning_rate": 5e-06, + "loss": 0.0654, + "step": 771 + }, + { + "epoch": 1.7585421412300684, + "grad_norm": 0.35617461800575256, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 772 + }, + { + "epoch": 1.7608200455580865, + "grad_norm": 0.31758973002433777, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 773 + }, + { + "epoch": 1.7630979498861048, + "grad_norm": 0.35947299003601074, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 774 + }, + { + "epoch": 1.7653758542141231, + "grad_norm": 0.3657311797142029, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 775 + }, + { + "epoch": 1.7676537585421412, + "grad_norm": 0.471209853887558, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 776 + }, + { + "epoch": 1.7699316628701594, + "grad_norm": 0.40624454617500305, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 777 + }, + { + "epoch": 1.7722095671981777, + "grad_norm": 0.37600862979888916, + "learning_rate": 5e-06, + "loss": 0.0628, + "step": 778 + }, + { + "epoch": 1.774487471526196, + "grad_norm": 0.37594738602638245, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 779 + }, + { + "epoch": 1.7767653758542141, + "grad_norm": 0.3494108319282532, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 780 + }, + { + "epoch": 1.7790432801822322, + "grad_norm": 0.3323187232017517, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 781 + }, + { + "epoch": 1.7813211845102506, + "grad_norm": 0.3645932376384735, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 782 + }, + { + "epoch": 1.783599088838269, + "grad_norm": 0.4065336287021637, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 783 + }, + { + "epoch": 1.785876993166287, + "grad_norm": 0.38880616426467896, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 784 + }, + { + "epoch": 1.7881548974943051, + "grad_norm": 0.375024676322937, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 785 + }, + { + "epoch": 1.7904328018223234, + "grad_norm": 0.37921419739723206, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 786 + }, + { + "epoch": 1.7927107061503418, + "grad_norm": 0.3726096749305725, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 787 + }, + { + "epoch": 1.79498861047836, + "grad_norm": 0.3758138120174408, + "learning_rate": 5e-06, + "loss": 0.0682, + "step": 788 + }, + { + "epoch": 1.797266514806378, + "grad_norm": 0.42794129252433777, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 789 + }, + { + "epoch": 1.7995444191343963, + "grad_norm": 0.38579314947128296, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 790 + }, + { + "epoch": 1.8018223234624147, + "grad_norm": 0.3638302683830261, + "learning_rate": 5e-06, + "loss": 0.0631, + "step": 791 + }, + { + "epoch": 1.8041002277904328, + "grad_norm": 0.4297358989715576, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 792 + }, + { + "epoch": 1.8063781321184509, + "grad_norm": 0.4153372645378113, + "learning_rate": 5e-06, + "loss": 0.0673, + "step": 793 + }, + { + "epoch": 1.8086560364464692, + "grad_norm": 0.3808562755584717, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 794 + }, + { + "epoch": 1.8109339407744875, + "grad_norm": 0.4327223002910614, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 795 + }, + { + "epoch": 1.8132118451025057, + "grad_norm": 0.44192948937416077, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 796 + }, + { + "epoch": 1.8154897494305238, + "grad_norm": 0.34678900241851807, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 797 + }, + { + "epoch": 1.817767653758542, + "grad_norm": 0.34950852394104004, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 798 + }, + { + "epoch": 1.8200455580865604, + "grad_norm": 0.36530327796936035, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 799 + }, + { + "epoch": 1.8223234624145785, + "grad_norm": 0.3592337667942047, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 800 + }, + { + "epoch": 1.8246013667425967, + "grad_norm": 0.3423630893230438, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 801 + }, + { + "epoch": 1.826879271070615, + "grad_norm": 0.35642048716545105, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 802 + }, + { + "epoch": 1.8291571753986333, + "grad_norm": 0.36607348918914795, + "learning_rate": 5e-06, + "loss": 0.0624, + "step": 803 + }, + { + "epoch": 1.8314350797266514, + "grad_norm": 0.43585270643234253, + "learning_rate": 5e-06, + "loss": 0.0671, + "step": 804 + }, + { + "epoch": 1.8337129840546698, + "grad_norm": 0.3595200777053833, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 805 + }, + { + "epoch": 1.835990888382688, + "grad_norm": 0.39456018805503845, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 806 + }, + { + "epoch": 1.8382687927107062, + "grad_norm": 0.37392258644104004, + "learning_rate": 5e-06, + "loss": 0.0667, + "step": 807 + }, + { + "epoch": 1.8405466970387243, + "grad_norm": 0.4371819496154785, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 808 + }, + { + "epoch": 1.8428246013667426, + "grad_norm": 0.4336298108100891, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 809 + }, + { + "epoch": 1.845102505694761, + "grad_norm": 0.3856502175331116, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 810 + }, + { + "epoch": 1.847380410022779, + "grad_norm": 0.4522041082382202, + "learning_rate": 5e-06, + "loss": 0.0692, + "step": 811 + }, + { + "epoch": 1.8496583143507972, + "grad_norm": 0.3512691259384155, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 812 + }, + { + "epoch": 1.8519362186788155, + "grad_norm": 0.43539708852767944, + "learning_rate": 5e-06, + "loss": 0.0667, + "step": 813 + }, + { + "epoch": 1.8542141230068339, + "grad_norm": 0.36477357149124146, + "learning_rate": 5e-06, + "loss": 0.0681, + "step": 814 + }, + { + "epoch": 1.856492027334852, + "grad_norm": 0.3925856649875641, + "learning_rate": 5e-06, + "loss": 0.0672, + "step": 815 + }, + { + "epoch": 1.85876993166287, + "grad_norm": 0.3788635730743408, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 816 + }, + { + "epoch": 1.8610478359908884, + "grad_norm": 0.4003060460090637, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 817 + }, + { + "epoch": 1.8633257403189067, + "grad_norm": 0.35700076818466187, + "learning_rate": 5e-06, + "loss": 0.0628, + "step": 818 + }, + { + "epoch": 1.8656036446469249, + "grad_norm": 0.3399184048175812, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 819 + }, + { + "epoch": 1.867881548974943, + "grad_norm": 0.3661590814590454, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 820 + }, + { + "epoch": 1.8701594533029613, + "grad_norm": 0.33902713656425476, + "learning_rate": 5e-06, + "loss": 0.0638, + "step": 821 + }, + { + "epoch": 1.8724373576309796, + "grad_norm": 0.447002112865448, + "learning_rate": 5e-06, + "loss": 0.0672, + "step": 822 + }, + { + "epoch": 1.8747152619589977, + "grad_norm": 0.38841983675956726, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 823 + }, + { + "epoch": 1.8769931662870158, + "grad_norm": 0.46865421533584595, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 824 + }, + { + "epoch": 1.8792710706150342, + "grad_norm": 0.3745507597923279, + "learning_rate": 5e-06, + "loss": 0.0667, + "step": 825 + }, + { + "epoch": 1.8815489749430525, + "grad_norm": 0.34042197465896606, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 826 + }, + { + "epoch": 1.8838268792710706, + "grad_norm": 0.36797165870666504, + "learning_rate": 5e-06, + "loss": 0.0626, + "step": 827 + }, + { + "epoch": 1.8861047835990887, + "grad_norm": 0.39723649621009827, + "learning_rate": 5e-06, + "loss": 0.0654, + "step": 828 + }, + { + "epoch": 1.888382687927107, + "grad_norm": 0.3646100163459778, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 829 + }, + { + "epoch": 1.8906605922551254, + "grad_norm": 0.35160142183303833, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 830 + }, + { + "epoch": 1.8929384965831435, + "grad_norm": 0.3282979130744934, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 831 + }, + { + "epoch": 1.8952164009111616, + "grad_norm": 0.33924567699432373, + "learning_rate": 5e-06, + "loss": 0.0633, + "step": 832 + }, + { + "epoch": 1.89749430523918, + "grad_norm": 0.34475693106651306, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 833 + }, + { + "epoch": 1.8997722095671983, + "grad_norm": 0.38452866673469543, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 834 + }, + { + "epoch": 1.9020501138952164, + "grad_norm": 0.36789998412132263, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 835 + }, + { + "epoch": 1.9043280182232345, + "grad_norm": 0.34242087602615356, + "learning_rate": 5e-06, + "loss": 0.0621, + "step": 836 + }, + { + "epoch": 1.9066059225512528, + "grad_norm": 0.33219853043556213, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 837 + }, + { + "epoch": 1.9088838268792712, + "grad_norm": 0.3524526059627533, + "learning_rate": 5e-06, + "loss": 0.0625, + "step": 838 + }, + { + "epoch": 1.9111617312072893, + "grad_norm": 0.3499576449394226, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 839 + }, + { + "epoch": 1.9134396355353074, + "grad_norm": 0.41930925846099854, + "learning_rate": 5e-06, + "loss": 0.0675, + "step": 840 + }, + { + "epoch": 1.9157175398633257, + "grad_norm": 0.36949628591537476, + "learning_rate": 5e-06, + "loss": 0.0673, + "step": 841 + }, + { + "epoch": 1.917995444191344, + "grad_norm": 0.354651540517807, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 842 + }, + { + "epoch": 1.9202733485193622, + "grad_norm": 0.37935441732406616, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 843 + }, + { + "epoch": 1.9225512528473803, + "grad_norm": 0.43017083406448364, + "learning_rate": 5e-06, + "loss": 0.0678, + "step": 844 + }, + { + "epoch": 1.9248291571753986, + "grad_norm": 0.37563037872314453, + "learning_rate": 5e-06, + "loss": 0.0669, + "step": 845 + }, + { + "epoch": 1.927107061503417, + "grad_norm": 0.3513509929180145, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 846 + }, + { + "epoch": 1.929384965831435, + "grad_norm": 0.3746267855167389, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 847 + }, + { + "epoch": 1.9316628701594531, + "grad_norm": 0.4095878303050995, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 848 + }, + { + "epoch": 1.9339407744874715, + "grad_norm": 0.4189017713069916, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 849 + }, + { + "epoch": 1.9362186788154898, + "grad_norm": 0.37285467982292175, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 850 + }, + { + "epoch": 1.938496583143508, + "grad_norm": 0.4462735950946808, + "learning_rate": 5e-06, + "loss": 0.0667, + "step": 851 + }, + { + "epoch": 1.940774487471526, + "grad_norm": 0.41118356585502625, + "learning_rate": 5e-06, + "loss": 0.0634, + "step": 852 + }, + { + "epoch": 1.9430523917995444, + "grad_norm": 0.3878600597381592, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 853 + }, + { + "epoch": 1.9453302961275627, + "grad_norm": 0.40632522106170654, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 854 + }, + { + "epoch": 1.9476082004555808, + "grad_norm": 0.35835206508636475, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 855 + }, + { + "epoch": 1.9498861047835991, + "grad_norm": 0.3729773163795471, + "learning_rate": 5e-06, + "loss": 0.0669, + "step": 856 + }, + { + "epoch": 1.9521640091116175, + "grad_norm": 0.32670673727989197, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 857 + }, + { + "epoch": 1.9544419134396356, + "grad_norm": 0.32948240637779236, + "learning_rate": 5e-06, + "loss": 0.0629, + "step": 858 + }, + { + "epoch": 1.9567198177676537, + "grad_norm": 0.30624064803123474, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 859 + }, + { + "epoch": 1.958997722095672, + "grad_norm": 0.32112807035446167, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 860 + }, + { + "epoch": 1.9612756264236904, + "grad_norm": 0.3518359959125519, + "learning_rate": 5e-06, + "loss": 0.0623, + "step": 861 + }, + { + "epoch": 1.9635535307517085, + "grad_norm": 0.3307065963745117, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 862 + }, + { + "epoch": 1.9658314350797266, + "grad_norm": 0.42508912086486816, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 863 + }, + { + "epoch": 1.968109339407745, + "grad_norm": 0.32008594274520874, + "learning_rate": 5e-06, + "loss": 0.0626, + "step": 864 + }, + { + "epoch": 1.9703872437357632, + "grad_norm": 0.36606013774871826, + "learning_rate": 5e-06, + "loss": 0.0638, + "step": 865 + }, + { + "epoch": 1.9726651480637813, + "grad_norm": 0.347261905670166, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 866 + }, + { + "epoch": 1.9749430523917995, + "grad_norm": 0.32424452900886536, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 867 + }, + { + "epoch": 1.9772209567198178, + "grad_norm": 0.3837982714176178, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 868 + }, + { + "epoch": 1.9794988610478361, + "grad_norm": 0.38964781165122986, + "learning_rate": 5e-06, + "loss": 0.061, + "step": 869 + }, + { + "epoch": 1.9817767653758542, + "grad_norm": 0.3276178538799286, + "learning_rate": 5e-06, + "loss": 0.0673, + "step": 870 + }, + { + "epoch": 1.9840546697038723, + "grad_norm": 0.35975369811058044, + "learning_rate": 5e-06, + "loss": 0.0628, + "step": 871 + }, + { + "epoch": 1.9863325740318907, + "grad_norm": 0.3330411911010742, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 872 + }, + { + "epoch": 1.988610478359909, + "grad_norm": 0.33740854263305664, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 873 + }, + { + "epoch": 1.9908883826879271, + "grad_norm": 0.3705332577228546, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 874 + }, + { + "epoch": 1.9931662870159452, + "grad_norm": 0.3322344720363617, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 875 + }, + { + "epoch": 1.9954441913439636, + "grad_norm": 0.37892571091651917, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 876 + }, + { + "epoch": 1.997722095671982, + "grad_norm": 0.4094819724559784, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 877 + }, + { + "epoch": 2.0, + "grad_norm": 0.3718055188655853, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 878 + }, + { + "epoch": 2.002277904328018, + "grad_norm": 0.4386118948459625, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 879 + }, + { + "epoch": 2.0045558086560367, + "grad_norm": 0.3467435836791992, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 880 + }, + { + "epoch": 2.0068337129840548, + "grad_norm": 0.3573801517486572, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 881 + }, + { + "epoch": 2.009111617312073, + "grad_norm": 0.36609163880348206, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 882 + }, + { + "epoch": 2.011389521640091, + "grad_norm": 0.36750468611717224, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 883 + }, + { + "epoch": 2.0136674259681095, + "grad_norm": 0.3811657726764679, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 884 + }, + { + "epoch": 2.0159453302961277, + "grad_norm": 0.38328248262405396, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 885 + }, + { + "epoch": 2.0182232346241458, + "grad_norm": 0.35302114486694336, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 886 + }, + { + "epoch": 2.020501138952164, + "grad_norm": 0.35229402780532837, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 887 + }, + { + "epoch": 2.0227790432801824, + "grad_norm": 0.4118206202983856, + "learning_rate": 5e-06, + "loss": 0.0671, + "step": 888 + }, + { + "epoch": 2.0250569476082005, + "grad_norm": 0.35905227065086365, + "learning_rate": 5e-06, + "loss": 0.0638, + "step": 889 + }, + { + "epoch": 2.0273348519362187, + "grad_norm": 0.3849807381629944, + "learning_rate": 5e-06, + "loss": 0.0626, + "step": 890 + }, + { + "epoch": 2.0296127562642368, + "grad_norm": 0.38305532932281494, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 891 + }, + { + "epoch": 2.0318906605922553, + "grad_norm": 0.3660757839679718, + "learning_rate": 5e-06, + "loss": 0.0651, + "step": 892 + }, + { + "epoch": 2.0341685649202734, + "grad_norm": 0.37978801131248474, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 893 + }, + { + "epoch": 2.0364464692482915, + "grad_norm": 0.3686749041080475, + "learning_rate": 5e-06, + "loss": 0.0664, + "step": 894 + }, + { + "epoch": 2.0387243735763096, + "grad_norm": 0.3721759617328644, + "learning_rate": 5e-06, + "loss": 0.0668, + "step": 895 + }, + { + "epoch": 2.041002277904328, + "grad_norm": 0.4217437207698822, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 896 + }, + { + "epoch": 2.0432801822323463, + "grad_norm": 0.37084710597991943, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 897 + }, + { + "epoch": 2.0455580865603644, + "grad_norm": 0.3468475341796875, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 898 + }, + { + "epoch": 2.0478359908883825, + "grad_norm": 0.3841940462589264, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 899 + }, + { + "epoch": 2.050113895216401, + "grad_norm": 0.4270191192626953, + "learning_rate": 5e-06, + "loss": 0.0679, + "step": 900 + }, + { + "epoch": 2.052391799544419, + "grad_norm": 0.3628332316875458, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 901 + }, + { + "epoch": 2.0546697038724373, + "grad_norm": 0.4229717254638672, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 902 + }, + { + "epoch": 2.0569476082004554, + "grad_norm": 0.37091630697250366, + "learning_rate": 5e-06, + "loss": 0.0631, + "step": 903 + }, + { + "epoch": 2.059225512528474, + "grad_norm": 0.3613353669643402, + "learning_rate": 5e-06, + "loss": 0.0682, + "step": 904 + }, + { + "epoch": 2.061503416856492, + "grad_norm": 0.4177112281322479, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 905 + }, + { + "epoch": 2.06378132118451, + "grad_norm": 0.3757723271846771, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 906 + }, + { + "epoch": 2.0660592255125283, + "grad_norm": 0.4058937430381775, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 907 + }, + { + "epoch": 2.068337129840547, + "grad_norm": 0.3449467420578003, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 908 + }, + { + "epoch": 2.070615034168565, + "grad_norm": 0.39998921751976013, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 909 + }, + { + "epoch": 2.072892938496583, + "grad_norm": 0.3876377046108246, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 910 + }, + { + "epoch": 2.075170842824601, + "grad_norm": 0.38587743043899536, + "learning_rate": 5e-06, + "loss": 0.0614, + "step": 911 + }, + { + "epoch": 2.0774487471526197, + "grad_norm": 0.38243022561073303, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 912 + }, + { + "epoch": 2.079726651480638, + "grad_norm": 0.37763819098472595, + "learning_rate": 5e-06, + "loss": 0.0631, + "step": 913 + }, + { + "epoch": 2.082004555808656, + "grad_norm": 0.39426735043525696, + "learning_rate": 5e-06, + "loss": 0.0634, + "step": 914 + }, + { + "epoch": 2.084282460136674, + "grad_norm": 0.4097016453742981, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 915 + }, + { + "epoch": 2.0865603644646926, + "grad_norm": 0.31911271810531616, + "learning_rate": 5e-06, + "loss": 0.0625, + "step": 916 + }, + { + "epoch": 2.0888382687927107, + "grad_norm": 0.37096673250198364, + "learning_rate": 5e-06, + "loss": 0.0631, + "step": 917 + }, + { + "epoch": 2.091116173120729, + "grad_norm": 0.37522169947624207, + "learning_rate": 5e-06, + "loss": 0.0638, + "step": 918 + }, + { + "epoch": 2.093394077448747, + "grad_norm": 0.35180211067199707, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 919 + }, + { + "epoch": 2.0956719817767655, + "grad_norm": 0.3494492173194885, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 920 + }, + { + "epoch": 2.0979498861047836, + "grad_norm": 0.38248974084854126, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 921 + }, + { + "epoch": 2.1002277904328017, + "grad_norm": 0.3548568785190582, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 922 + }, + { + "epoch": 2.10250569476082, + "grad_norm": 0.34141600131988525, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 923 + }, + { + "epoch": 2.1047835990888384, + "grad_norm": 0.43948766589164734, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 924 + }, + { + "epoch": 2.1070615034168565, + "grad_norm": 0.370527446269989, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 925 + }, + { + "epoch": 2.1093394077448746, + "grad_norm": 0.44553908705711365, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 926 + }, + { + "epoch": 2.1116173120728927, + "grad_norm": 0.4327397346496582, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 927 + }, + { + "epoch": 2.1138952164009113, + "grad_norm": 0.42272982001304626, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 928 + }, + { + "epoch": 2.1161731207289294, + "grad_norm": 0.4232465326786041, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 929 + }, + { + "epoch": 2.1184510250569475, + "grad_norm": 0.40464767813682556, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 930 + }, + { + "epoch": 2.120728929384966, + "grad_norm": 0.39294788241386414, + "learning_rate": 5e-06, + "loss": 0.062, + "step": 931 + }, + { + "epoch": 2.123006833712984, + "grad_norm": 0.3918789029121399, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 932 + }, + { + "epoch": 2.1252847380410023, + "grad_norm": 0.47361522912979126, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 933 + }, + { + "epoch": 2.1275626423690204, + "grad_norm": 0.3954993784427643, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 934 + }, + { + "epoch": 2.129840546697039, + "grad_norm": 0.4152219295501709, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 935 + }, + { + "epoch": 2.132118451025057, + "grad_norm": 0.42703163623809814, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 936 + }, + { + "epoch": 2.134396355353075, + "grad_norm": 0.3991785943508148, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 937 + }, + { + "epoch": 2.1366742596810933, + "grad_norm": 0.4113765060901642, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 938 + }, + { + "epoch": 2.138952164009112, + "grad_norm": 0.348736047744751, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 939 + }, + { + "epoch": 2.14123006833713, + "grad_norm": 0.34686192870140076, + "learning_rate": 5e-06, + "loss": 0.0638, + "step": 940 + }, + { + "epoch": 2.143507972665148, + "grad_norm": 0.4006514251232147, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 941 + }, + { + "epoch": 2.145785876993166, + "grad_norm": 0.360848605632782, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 942 + }, + { + "epoch": 2.1480637813211847, + "grad_norm": 0.39627739787101746, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 943 + }, + { + "epoch": 2.150341685649203, + "grad_norm": 0.3971979320049286, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 944 + }, + { + "epoch": 2.152619589977221, + "grad_norm": 0.3891604542732239, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 945 + }, + { + "epoch": 2.154897494305239, + "grad_norm": 0.42411237955093384, + "learning_rate": 5e-06, + "loss": 0.0609, + "step": 946 + }, + { + "epoch": 2.1571753986332576, + "grad_norm": 0.32573559880256653, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 947 + }, + { + "epoch": 2.1594533029612757, + "grad_norm": 0.46468672156333923, + "learning_rate": 5e-06, + "loss": 0.0616, + "step": 948 + }, + { + "epoch": 2.161731207289294, + "grad_norm": 0.3554407060146332, + "learning_rate": 5e-06, + "loss": 0.063, + "step": 949 + }, + { + "epoch": 2.164009111617312, + "grad_norm": 0.4320879280567169, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 950 + }, + { + "epoch": 2.1662870159453305, + "grad_norm": 0.347248911857605, + "learning_rate": 5e-06, + "loss": 0.0614, + "step": 951 + }, + { + "epoch": 2.1685649202733486, + "grad_norm": 0.36107945442199707, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 952 + }, + { + "epoch": 2.1708428246013667, + "grad_norm": 0.33618679642677307, + "learning_rate": 5e-06, + "loss": 0.0629, + "step": 953 + }, + { + "epoch": 2.173120728929385, + "grad_norm": 0.359061598777771, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 954 + }, + { + "epoch": 2.1753986332574033, + "grad_norm": 0.3282698094844818, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 955 + }, + { + "epoch": 2.1776765375854215, + "grad_norm": 0.3779278099536896, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 956 + }, + { + "epoch": 2.1799544419134396, + "grad_norm": 0.35100382566452026, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 957 + }, + { + "epoch": 2.1822323462414577, + "grad_norm": 0.41122227907180786, + "learning_rate": 5e-06, + "loss": 0.0612, + "step": 958 + }, + { + "epoch": 2.1845102505694762, + "grad_norm": 0.31362035870552063, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 959 + }, + { + "epoch": 2.1867881548974943, + "grad_norm": 0.32592323422431946, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 960 + }, + { + "epoch": 2.1890660592255125, + "grad_norm": 0.3565470278263092, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 961 + }, + { + "epoch": 2.1913439635535306, + "grad_norm": 0.33629193902015686, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 962 + }, + { + "epoch": 2.193621867881549, + "grad_norm": 0.32078197598457336, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 963 + }, + { + "epoch": 2.1958997722095672, + "grad_norm": 0.3121110200881958, + "learning_rate": 5e-06, + "loss": 0.0624, + "step": 964 + }, + { + "epoch": 2.1981776765375853, + "grad_norm": 0.35310298204421997, + "learning_rate": 5e-06, + "loss": 0.061, + "step": 965 + }, + { + "epoch": 2.2004555808656034, + "grad_norm": 0.3640041649341583, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 966 + }, + { + "epoch": 2.202733485193622, + "grad_norm": 0.3210286498069763, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 967 + }, + { + "epoch": 2.20501138952164, + "grad_norm": 0.4181840419769287, + "learning_rate": 5e-06, + "loss": 0.0654, + "step": 968 + }, + { + "epoch": 2.207289293849658, + "grad_norm": 0.3561514616012573, + "learning_rate": 5e-06, + "loss": 0.0669, + "step": 969 + }, + { + "epoch": 2.2095671981776768, + "grad_norm": 0.3986862599849701, + "learning_rate": 5e-06, + "loss": 0.0654, + "step": 970 + }, + { + "epoch": 2.211845102505695, + "grad_norm": 0.349417507648468, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 971 + }, + { + "epoch": 2.214123006833713, + "grad_norm": 0.37399938702583313, + "learning_rate": 5e-06, + "loss": 0.0633, + "step": 972 + }, + { + "epoch": 2.216400911161731, + "grad_norm": 0.3361985385417938, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 973 + }, + { + "epoch": 2.218678815489749, + "grad_norm": 0.359173059463501, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 974 + }, + { + "epoch": 2.2209567198177678, + "grad_norm": 0.3447031080722809, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 975 + }, + { + "epoch": 2.223234624145786, + "grad_norm": 0.3254874050617218, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 976 + }, + { + "epoch": 2.225512528473804, + "grad_norm": 0.3558181822299957, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 977 + }, + { + "epoch": 2.2277904328018225, + "grad_norm": 0.37203213572502136, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 978 + }, + { + "epoch": 2.2300683371298406, + "grad_norm": 0.4200899600982666, + "learning_rate": 5e-06, + "loss": 0.0668, + "step": 979 + }, + { + "epoch": 2.2323462414578588, + "grad_norm": 0.3482038080692291, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 980 + }, + { + "epoch": 2.234624145785877, + "grad_norm": 0.38406234979629517, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 981 + }, + { + "epoch": 2.2369020501138954, + "grad_norm": 0.3911779820919037, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 982 + }, + { + "epoch": 2.2391799544419135, + "grad_norm": 0.3870941698551178, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 983 + }, + { + "epoch": 2.2414578587699316, + "grad_norm": 0.426997572183609, + "learning_rate": 5e-06, + "loss": 0.0638, + "step": 984 + }, + { + "epoch": 2.2437357630979498, + "grad_norm": 0.3327006697654724, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 985 + }, + { + "epoch": 2.2460136674259683, + "grad_norm": 0.40405985713005066, + "learning_rate": 5e-06, + "loss": 0.0629, + "step": 986 + }, + { + "epoch": 2.2482915717539864, + "grad_norm": 0.3728192448616028, + "learning_rate": 5e-06, + "loss": 0.0631, + "step": 987 + }, + { + "epoch": 2.2505694760820045, + "grad_norm": 0.447054922580719, + "learning_rate": 5e-06, + "loss": 0.0679, + "step": 988 + }, + { + "epoch": 2.2528473804100226, + "grad_norm": 0.4169905483722687, + "learning_rate": 5e-06, + "loss": 0.0675, + "step": 989 + }, + { + "epoch": 2.255125284738041, + "grad_norm": 0.4643232524394989, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 990 + }, + { + "epoch": 2.2574031890660593, + "grad_norm": 0.370007187128067, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 991 + }, + { + "epoch": 2.2596810933940774, + "grad_norm": 0.5323000550270081, + "learning_rate": 5e-06, + "loss": 0.068, + "step": 992 + }, + { + "epoch": 2.2619589977220955, + "grad_norm": 0.41063234210014343, + "learning_rate": 5e-06, + "loss": 0.0631, + "step": 993 + }, + { + "epoch": 2.264236902050114, + "grad_norm": 0.4756401777267456, + "learning_rate": 5e-06, + "loss": 0.0615, + "step": 994 + }, + { + "epoch": 2.266514806378132, + "grad_norm": 0.47712597250938416, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 995 + }, + { + "epoch": 2.2687927107061503, + "grad_norm": 0.43546780943870544, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 996 + }, + { + "epoch": 2.2710706150341684, + "grad_norm": 0.4730824828147888, + "learning_rate": 5e-06, + "loss": 0.0654, + "step": 997 + }, + { + "epoch": 2.273348519362187, + "grad_norm": 0.3857640027999878, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 998 + }, + { + "epoch": 2.275626423690205, + "grad_norm": 0.48187997937202454, + "learning_rate": 5e-06, + "loss": 0.0622, + "step": 999 + }, + { + "epoch": 2.277904328018223, + "grad_norm": 0.41435930132865906, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 1000 + }, + { + "epoch": 2.2801822323462413, + "grad_norm": 0.4311675727367401, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 1001 + }, + { + "epoch": 2.28246013667426, + "grad_norm": 0.3588732182979584, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 1002 + }, + { + "epoch": 2.284738041002278, + "grad_norm": 0.39682042598724365, + "learning_rate": 5e-06, + "loss": 0.0624, + "step": 1003 + }, + { + "epoch": 2.287015945330296, + "grad_norm": 0.32970863580703735, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 1004 + }, + { + "epoch": 2.289293849658314, + "grad_norm": 0.41080474853515625, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 1005 + }, + { + "epoch": 2.2915717539863327, + "grad_norm": 0.3642153739929199, + "learning_rate": 5e-06, + "loss": 0.0628, + "step": 1006 + }, + { + "epoch": 2.293849658314351, + "grad_norm": 0.3400186002254486, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 1007 + }, + { + "epoch": 2.296127562642369, + "grad_norm": 0.36358851194381714, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 1008 + }, + { + "epoch": 2.298405466970387, + "grad_norm": 0.3627476990222931, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 1009 + }, + { + "epoch": 2.3006833712984056, + "grad_norm": 0.33303695917129517, + "learning_rate": 5e-06, + "loss": 0.0623, + "step": 1010 + }, + { + "epoch": 2.3029612756264237, + "grad_norm": 0.3159809112548828, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 1011 + }, + { + "epoch": 2.305239179954442, + "grad_norm": 0.3029627203941345, + "learning_rate": 5e-06, + "loss": 0.0615, + "step": 1012 + }, + { + "epoch": 2.30751708428246, + "grad_norm": 0.33818596601486206, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 1013 + }, + { + "epoch": 2.3097949886104785, + "grad_norm": 0.3519606292247772, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 1014 + }, + { + "epoch": 2.3120728929384966, + "grad_norm": 0.3234321177005768, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 1015 + }, + { + "epoch": 2.3143507972665147, + "grad_norm": 0.3304242193698883, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 1016 + }, + { + "epoch": 2.3166287015945333, + "grad_norm": 0.3527721166610718, + "learning_rate": 5e-06, + "loss": 0.0617, + "step": 1017 + }, + { + "epoch": 2.3189066059225514, + "grad_norm": 0.32102298736572266, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 1018 + }, + { + "epoch": 2.3211845102505695, + "grad_norm": 0.3999481797218323, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 1019 + }, + { + "epoch": 2.3234624145785876, + "grad_norm": 0.3413870632648468, + "learning_rate": 5e-06, + "loss": 0.063, + "step": 1020 + }, + { + "epoch": 2.3257403189066057, + "grad_norm": 0.3446578085422516, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 1021 + }, + { + "epoch": 2.3280182232346243, + "grad_norm": 0.3409595489501953, + "learning_rate": 5e-06, + "loss": 0.0627, + "step": 1022 + }, + { + "epoch": 2.3302961275626424, + "grad_norm": 0.38335004448890686, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 1023 + }, + { + "epoch": 2.3325740318906605, + "grad_norm": 0.34537696838378906, + "learning_rate": 5e-06, + "loss": 0.0672, + "step": 1024 + }, + { + "epoch": 2.334851936218679, + "grad_norm": 0.35324957966804504, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 1025 + }, + { + "epoch": 2.337129840546697, + "grad_norm": 0.3460294306278229, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 1026 + }, + { + "epoch": 2.3394077448747153, + "grad_norm": 0.3597829341888428, + "learning_rate": 5e-06, + "loss": 0.0613, + "step": 1027 + }, + { + "epoch": 2.3416856492027334, + "grad_norm": 0.34790682792663574, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 1028 + }, + { + "epoch": 2.3439635535307515, + "grad_norm": 0.372488796710968, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 1029 + }, + { + "epoch": 2.34624145785877, + "grad_norm": 0.3652331233024597, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 1030 + }, + { + "epoch": 2.348519362186788, + "grad_norm": 0.35086071491241455, + "learning_rate": 5e-06, + "loss": 0.0629, + "step": 1031 + }, + { + "epoch": 2.3507972665148062, + "grad_norm": 0.3590526282787323, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 1032 + }, + { + "epoch": 2.353075170842825, + "grad_norm": 0.3077162504196167, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 1033 + }, + { + "epoch": 2.355353075170843, + "grad_norm": 0.36362966895103455, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 1034 + }, + { + "epoch": 2.357630979498861, + "grad_norm": 0.35024797916412354, + "learning_rate": 5e-06, + "loss": 0.0625, + "step": 1035 + }, + { + "epoch": 2.359908883826879, + "grad_norm": 0.35729578137397766, + "learning_rate": 5e-06, + "loss": 0.0651, + "step": 1036 + }, + { + "epoch": 2.3621867881548977, + "grad_norm": 0.3868909776210785, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 1037 + }, + { + "epoch": 2.364464692482916, + "grad_norm": 0.3617287874221802, + "learning_rate": 5e-06, + "loss": 0.0685, + "step": 1038 + }, + { + "epoch": 2.366742596810934, + "grad_norm": 0.40562912821769714, + "learning_rate": 5e-06, + "loss": 0.0624, + "step": 1039 + }, + { + "epoch": 2.369020501138952, + "grad_norm": 0.3510848581790924, + "learning_rate": 5e-06, + "loss": 0.0669, + "step": 1040 + }, + { + "epoch": 2.3712984054669706, + "grad_norm": 0.42867493629455566, + "learning_rate": 5e-06, + "loss": 0.0604, + "step": 1041 + }, + { + "epoch": 2.3735763097949887, + "grad_norm": 0.35482341051101685, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 1042 + }, + { + "epoch": 2.375854214123007, + "grad_norm": 0.42982247471809387, + "learning_rate": 5e-06, + "loss": 0.0633, + "step": 1043 + }, + { + "epoch": 2.378132118451025, + "grad_norm": 0.41004472970962524, + "learning_rate": 5e-06, + "loss": 0.0628, + "step": 1044 + }, + { + "epoch": 2.3804100227790435, + "grad_norm": 0.37374380230903625, + "learning_rate": 5e-06, + "loss": 0.0638, + "step": 1045 + }, + { + "epoch": 2.3826879271070616, + "grad_norm": 0.40902790427207947, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 1046 + }, + { + "epoch": 2.3849658314350797, + "grad_norm": 0.3632076680660248, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 1047 + }, + { + "epoch": 2.387243735763098, + "grad_norm": 0.34351205825805664, + "learning_rate": 5e-06, + "loss": 0.0634, + "step": 1048 + }, + { + "epoch": 2.3895216400911163, + "grad_norm": 0.38355907797813416, + "learning_rate": 5e-06, + "loss": 0.0611, + "step": 1049 + }, + { + "epoch": 2.3917995444191344, + "grad_norm": 0.33864936232566833, + "learning_rate": 5e-06, + "loss": 0.063, + "step": 1050 + }, + { + "epoch": 2.3940774487471526, + "grad_norm": 0.42707204818725586, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 1051 + }, + { + "epoch": 2.3963553530751707, + "grad_norm": 0.35341501235961914, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 1052 + }, + { + "epoch": 2.3986332574031892, + "grad_norm": 0.4059014618396759, + "learning_rate": 5e-06, + "loss": 0.0627, + "step": 1053 + }, + { + "epoch": 2.4009111617312073, + "grad_norm": 0.3321845233440399, + "learning_rate": 5e-06, + "loss": 0.0627, + "step": 1054 + }, + { + "epoch": 2.4031890660592254, + "grad_norm": 0.4802558422088623, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 1055 + }, + { + "epoch": 2.4054669703872436, + "grad_norm": 0.37992075085639954, + "learning_rate": 5e-06, + "loss": 0.0651, + "step": 1056 + }, + { + "epoch": 2.407744874715262, + "grad_norm": 0.4289088249206543, + "learning_rate": 5e-06, + "loss": 0.0622, + "step": 1057 + }, + { + "epoch": 2.41002277904328, + "grad_norm": 0.3717344403266907, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 1058 + }, + { + "epoch": 2.4123006833712983, + "grad_norm": 0.3617473840713501, + "learning_rate": 5e-06, + "loss": 0.0626, + "step": 1059 + }, + { + "epoch": 2.4145785876993164, + "grad_norm": 0.3671506643295288, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 1060 + }, + { + "epoch": 2.416856492027335, + "grad_norm": 0.3733121156692505, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 1061 + }, + { + "epoch": 2.419134396355353, + "grad_norm": 0.3563171625137329, + "learning_rate": 5e-06, + "loss": 0.0667, + "step": 1062 + }, + { + "epoch": 2.421412300683371, + "grad_norm": 0.41789793968200684, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 1063 + }, + { + "epoch": 2.4236902050113898, + "grad_norm": 0.3826461732387543, + "learning_rate": 5e-06, + "loss": 0.0651, + "step": 1064 + }, + { + "epoch": 2.425968109339408, + "grad_norm": 0.38771623373031616, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 1065 + }, + { + "epoch": 2.428246013667426, + "grad_norm": 0.34157484769821167, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 1066 + }, + { + "epoch": 2.430523917995444, + "grad_norm": 0.3616361916065216, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 1067 + }, + { + "epoch": 2.432801822323462, + "grad_norm": 0.3935610353946686, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 1068 + }, + { + "epoch": 2.4350797266514808, + "grad_norm": 0.38885191082954407, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 1069 + }, + { + "epoch": 2.437357630979499, + "grad_norm": 0.36633971333503723, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 1070 + }, + { + "epoch": 2.439635535307517, + "grad_norm": 0.39490458369255066, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 1071 + }, + { + "epoch": 2.4419134396355355, + "grad_norm": 0.3240135908126831, + "learning_rate": 5e-06, + "loss": 0.0622, + "step": 1072 + }, + { + "epoch": 2.4441913439635536, + "grad_norm": 0.3697579503059387, + "learning_rate": 5e-06, + "loss": 0.0604, + "step": 1073 + }, + { + "epoch": 2.4464692482915718, + "grad_norm": 0.4132577180862427, + "learning_rate": 5e-06, + "loss": 0.0669, + "step": 1074 + }, + { + "epoch": 2.44874715261959, + "grad_norm": 0.39057809114456177, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 1075 + }, + { + "epoch": 2.451025056947608, + "grad_norm": 0.3658410310745239, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 1076 + }, + { + "epoch": 2.4533029612756265, + "grad_norm": 0.4368787109851837, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 1077 + }, + { + "epoch": 2.4555808656036446, + "grad_norm": 0.3769480586051941, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 1078 + }, + { + "epoch": 2.4578587699316627, + "grad_norm": 0.3672235906124115, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 1079 + }, + { + "epoch": 2.4601366742596813, + "grad_norm": 0.39357760548591614, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 1080 + }, + { + "epoch": 2.4624145785876994, + "grad_norm": 0.37806954979896545, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 1081 + }, + { + "epoch": 2.4646924829157175, + "grad_norm": 0.35902294516563416, + "learning_rate": 5e-06, + "loss": 0.0668, + "step": 1082 + }, + { + "epoch": 2.4669703872437356, + "grad_norm": 0.3983694314956665, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 1083 + }, + { + "epoch": 2.4692482915717537, + "grad_norm": 0.3836486041545868, + "learning_rate": 5e-06, + "loss": 0.0638, + "step": 1084 + }, + { + "epoch": 2.4715261958997723, + "grad_norm": 0.3957916796207428, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 1085 + }, + { + "epoch": 2.4738041002277904, + "grad_norm": 0.39857879281044006, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 1086 + }, + { + "epoch": 2.4760820045558085, + "grad_norm": 0.4524567425251007, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 1087 + }, + { + "epoch": 2.478359908883827, + "grad_norm": 0.35638338327407837, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 1088 + }, + { + "epoch": 2.480637813211845, + "grad_norm": 0.4003960192203522, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 1089 + }, + { + "epoch": 2.4829157175398633, + "grad_norm": 0.3599901497364044, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 1090 + }, + { + "epoch": 2.4851936218678814, + "grad_norm": 0.3281540870666504, + "learning_rate": 5e-06, + "loss": 0.0638, + "step": 1091 + }, + { + "epoch": 2.4874715261959, + "grad_norm": 0.3587161600589752, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 1092 + }, + { + "epoch": 2.489749430523918, + "grad_norm": 0.32150331139564514, + "learning_rate": 5e-06, + "loss": 0.0621, + "step": 1093 + }, + { + "epoch": 2.492027334851936, + "grad_norm": 0.38395655155181885, + "learning_rate": 5e-06, + "loss": 0.0671, + "step": 1094 + }, + { + "epoch": 2.4943052391799543, + "grad_norm": 0.3485938608646393, + "learning_rate": 5e-06, + "loss": 0.0664, + "step": 1095 + }, + { + "epoch": 2.496583143507973, + "grad_norm": 0.34448081254959106, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 1096 + }, + { + "epoch": 2.498861047835991, + "grad_norm": 0.3400004804134369, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 1097 + }, + { + "epoch": 2.501138952164009, + "grad_norm": 0.3427201211452484, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 1098 + }, + { + "epoch": 2.503416856492027, + "grad_norm": 0.3488285541534424, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 1099 + }, + { + "epoch": 2.5056947608200457, + "grad_norm": 0.3434089422225952, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 1100 + }, + { + "epoch": 2.507972665148064, + "grad_norm": 0.33915603160858154, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 1101 + }, + { + "epoch": 2.510250569476082, + "grad_norm": 0.33430543541908264, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 1102 + }, + { + "epoch": 2.5125284738041005, + "grad_norm": 0.4376661777496338, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 1103 + }, + { + "epoch": 2.5148063781321186, + "grad_norm": 0.3509063422679901, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 1104 + }, + { + "epoch": 2.5170842824601367, + "grad_norm": 0.4079601466655731, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 1105 + }, + { + "epoch": 2.519362186788155, + "grad_norm": 0.36879250407218933, + "learning_rate": 5e-06, + "loss": 0.0666, + "step": 1106 + }, + { + "epoch": 2.521640091116173, + "grad_norm": 0.4270707964897156, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 1107 + }, + { + "epoch": 2.5239179954441915, + "grad_norm": 0.3785046637058258, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 1108 + }, + { + "epoch": 2.5261958997722096, + "grad_norm": 0.37609314918518066, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 1109 + }, + { + "epoch": 2.5284738041002277, + "grad_norm": 0.40449628233909607, + "learning_rate": 5e-06, + "loss": 0.0651, + "step": 1110 + }, + { + "epoch": 2.5307517084282463, + "grad_norm": 0.33656638860702515, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 1111 + }, + { + "epoch": 2.5330296127562644, + "grad_norm": 0.34896692633628845, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 1112 + }, + { + "epoch": 2.5353075170842825, + "grad_norm": 0.3910132050514221, + "learning_rate": 5e-06, + "loss": 0.0676, + "step": 1113 + }, + { + "epoch": 2.5375854214123006, + "grad_norm": 0.4085399806499481, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 1114 + }, + { + "epoch": 2.5398633257403187, + "grad_norm": 0.40095609426498413, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 1115 + }, + { + "epoch": 2.5421412300683373, + "grad_norm": 0.388229101896286, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 1116 + }, + { + "epoch": 2.5444191343963554, + "grad_norm": 0.33334848284721375, + "learning_rate": 5e-06, + "loss": 0.0619, + "step": 1117 + }, + { + "epoch": 2.5466970387243735, + "grad_norm": 0.3782349228858948, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 1118 + }, + { + "epoch": 2.548974943052392, + "grad_norm": 0.34723809361457825, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 1119 + }, + { + "epoch": 2.55125284738041, + "grad_norm": 0.3522922992706299, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 1120 + }, + { + "epoch": 2.5535307517084282, + "grad_norm": 0.3559888005256653, + "learning_rate": 5e-06, + "loss": 0.0674, + "step": 1121 + }, + { + "epoch": 2.5558086560364464, + "grad_norm": 0.3526102602481842, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 1122 + }, + { + "epoch": 2.5580865603644645, + "grad_norm": 0.3434252142906189, + "learning_rate": 5e-06, + "loss": 0.0662, + "step": 1123 + }, + { + "epoch": 2.560364464692483, + "grad_norm": 0.3523869216442108, + "learning_rate": 5e-06, + "loss": 0.0627, + "step": 1124 + }, + { + "epoch": 2.562642369020501, + "grad_norm": 0.3304862380027771, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 1125 + }, + { + "epoch": 2.5649202733485192, + "grad_norm": 0.3856379985809326, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 1126 + }, + { + "epoch": 2.567198177676538, + "grad_norm": 0.35280197858810425, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 1127 + }, + { + "epoch": 2.569476082004556, + "grad_norm": 0.3772255480289459, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 1128 + }, + { + "epoch": 2.571753986332574, + "grad_norm": 0.3618199825286865, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 1129 + }, + { + "epoch": 2.574031890660592, + "grad_norm": 0.40705057978630066, + "learning_rate": 5e-06, + "loss": 0.0666, + "step": 1130 + }, + { + "epoch": 2.5763097949886102, + "grad_norm": 0.3558051288127899, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 1131 + }, + { + "epoch": 2.578587699316629, + "grad_norm": 0.34395596385002136, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 1132 + }, + { + "epoch": 2.580865603644647, + "grad_norm": 0.36210745573043823, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 1133 + }, + { + "epoch": 2.583143507972665, + "grad_norm": 0.3228192627429962, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 1134 + }, + { + "epoch": 2.5854214123006836, + "grad_norm": 0.42203786969184875, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 1135 + }, + { + "epoch": 2.5876993166287017, + "grad_norm": 0.3415069878101349, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 1136 + }, + { + "epoch": 2.58997722095672, + "grad_norm": 0.42018458247184753, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 1137 + }, + { + "epoch": 2.592255125284738, + "grad_norm": 0.3087046444416046, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 1138 + }, + { + "epoch": 2.594533029612756, + "grad_norm": 0.44265905022621155, + "learning_rate": 5e-06, + "loss": 0.0624, + "step": 1139 + }, + { + "epoch": 2.5968109339407746, + "grad_norm": 0.40397506952285767, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 1140 + }, + { + "epoch": 2.5990888382687927, + "grad_norm": 0.37356775999069214, + "learning_rate": 5e-06, + "loss": 0.0624, + "step": 1141 + }, + { + "epoch": 2.6013667425968108, + "grad_norm": 0.36775606870651245, + "learning_rate": 5e-06, + "loss": 0.0631, + "step": 1142 + }, + { + "epoch": 2.6036446469248293, + "grad_norm": 0.369292289018631, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 1143 + }, + { + "epoch": 2.6059225512528474, + "grad_norm": 0.36045145988464355, + "learning_rate": 5e-06, + "loss": 0.0669, + "step": 1144 + }, + { + "epoch": 2.6082004555808656, + "grad_norm": 0.328274667263031, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 1145 + }, + { + "epoch": 2.6104783599088837, + "grad_norm": 0.38536250591278076, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 1146 + }, + { + "epoch": 2.612756264236902, + "grad_norm": 0.36271294951438904, + "learning_rate": 5e-06, + "loss": 0.063, + "step": 1147 + }, + { + "epoch": 2.6150341685649203, + "grad_norm": 0.3569026589393616, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 1148 + }, + { + "epoch": 2.6173120728929384, + "grad_norm": 0.36661359667778015, + "learning_rate": 5e-06, + "loss": 0.0654, + "step": 1149 + }, + { + "epoch": 2.619589977220957, + "grad_norm": 0.357048362493515, + "learning_rate": 5e-06, + "loss": 0.0617, + "step": 1150 + }, + { + "epoch": 2.621867881548975, + "grad_norm": 0.3479042649269104, + "learning_rate": 5e-06, + "loss": 0.063, + "step": 1151 + }, + { + "epoch": 2.624145785876993, + "grad_norm": 0.37682074308395386, + "learning_rate": 5e-06, + "loss": 0.0672, + "step": 1152 + }, + { + "epoch": 2.6264236902050113, + "grad_norm": 0.3442467749118805, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 1153 + }, + { + "epoch": 2.6287015945330294, + "grad_norm": 0.350748747587204, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 1154 + }, + { + "epoch": 2.630979498861048, + "grad_norm": 0.3581165671348572, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 1155 + }, + { + "epoch": 2.633257403189066, + "grad_norm": 0.35141557455062866, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 1156 + }, + { + "epoch": 2.635535307517084, + "grad_norm": 0.3422101140022278, + "learning_rate": 5e-06, + "loss": 0.0621, + "step": 1157 + }, + { + "epoch": 2.6378132118451028, + "grad_norm": 0.32578712701797485, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 1158 + }, + { + "epoch": 2.640091116173121, + "grad_norm": 0.3447733223438263, + "learning_rate": 5e-06, + "loss": 0.0647, + "step": 1159 + }, + { + "epoch": 2.642369020501139, + "grad_norm": 0.31952229142189026, + "learning_rate": 5e-06, + "loss": 0.0608, + "step": 1160 + }, + { + "epoch": 2.644646924829157, + "grad_norm": 0.3545774519443512, + "learning_rate": 5e-06, + "loss": 0.0625, + "step": 1161 + }, + { + "epoch": 2.646924829157175, + "grad_norm": 0.306447833776474, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 1162 + }, + { + "epoch": 2.6492027334851938, + "grad_norm": 0.3436081111431122, + "learning_rate": 5e-06, + "loss": 0.0631, + "step": 1163 + }, + { + "epoch": 2.651480637813212, + "grad_norm": 0.3364735543727875, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 1164 + }, + { + "epoch": 2.65375854214123, + "grad_norm": 0.36356112360954285, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 1165 + }, + { + "epoch": 2.6560364464692485, + "grad_norm": 0.318620890378952, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 1166 + }, + { + "epoch": 2.6583143507972666, + "grad_norm": 0.3614048957824707, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 1167 + }, + { + "epoch": 2.6605922551252847, + "grad_norm": 0.3666227459907532, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 1168 + }, + { + "epoch": 2.662870159453303, + "grad_norm": 0.35970601439476013, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 1169 + }, + { + "epoch": 2.665148063781321, + "grad_norm": 0.3734301030635834, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 1170 + }, + { + "epoch": 2.6674259681093395, + "grad_norm": 0.361607164144516, + "learning_rate": 5e-06, + "loss": 0.066, + "step": 1171 + }, + { + "epoch": 2.6697038724373576, + "grad_norm": 0.36409586668014526, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 1172 + }, + { + "epoch": 2.6719817767653757, + "grad_norm": 0.3264155983924866, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 1173 + }, + { + "epoch": 2.6742596810933943, + "grad_norm": 0.3416155278682709, + "learning_rate": 5e-06, + "loss": 0.0622, + "step": 1174 + }, + { + "epoch": 2.6765375854214124, + "grad_norm": 0.33026736974716187, + "learning_rate": 5e-06, + "loss": 0.0627, + "step": 1175 + }, + { + "epoch": 2.6788154897494305, + "grad_norm": 0.30708789825439453, + "learning_rate": 5e-06, + "loss": 0.0629, + "step": 1176 + }, + { + "epoch": 2.6810933940774486, + "grad_norm": 0.34217360615730286, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 1177 + }, + { + "epoch": 2.6833712984054667, + "grad_norm": 0.3175274729728699, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 1178 + }, + { + "epoch": 2.6856492027334853, + "grad_norm": 0.3422200083732605, + "learning_rate": 5e-06, + "loss": 0.0627, + "step": 1179 + }, + { + "epoch": 2.6879271070615034, + "grad_norm": 0.3193976879119873, + "learning_rate": 5e-06, + "loss": 0.0624, + "step": 1180 + }, + { + "epoch": 2.6902050113895215, + "grad_norm": 0.3266102075576782, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 1181 + }, + { + "epoch": 2.69248291571754, + "grad_norm": 0.37876564264297485, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 1182 + }, + { + "epoch": 2.694760820045558, + "grad_norm": 0.35405901074409485, + "learning_rate": 5e-06, + "loss": 0.0656, + "step": 1183 + }, + { + "epoch": 2.6970387243735763, + "grad_norm": 0.35572513937950134, + "learning_rate": 5e-06, + "loss": 0.0663, + "step": 1184 + }, + { + "epoch": 2.6993166287015944, + "grad_norm": 0.3613019287586212, + "learning_rate": 5e-06, + "loss": 0.0609, + "step": 1185 + }, + { + "epoch": 2.7015945330296125, + "grad_norm": 0.37498390674591064, + "learning_rate": 5e-06, + "loss": 0.0627, + "step": 1186 + }, + { + "epoch": 2.703872437357631, + "grad_norm": 0.31806430220603943, + "learning_rate": 5e-06, + "loss": 0.0666, + "step": 1187 + }, + { + "epoch": 2.706150341685649, + "grad_norm": 0.4339022934436798, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 1188 + }, + { + "epoch": 2.7084282460136673, + "grad_norm": 0.32378706336021423, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 1189 + }, + { + "epoch": 2.710706150341686, + "grad_norm": 0.34744852781295776, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 1190 + }, + { + "epoch": 2.712984054669704, + "grad_norm": 0.32550448179244995, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 1191 + }, + { + "epoch": 2.715261958997722, + "grad_norm": 0.33748725056648254, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 1192 + }, + { + "epoch": 2.71753986332574, + "grad_norm": 0.396217405796051, + "learning_rate": 5e-06, + "loss": 0.0624, + "step": 1193 + }, + { + "epoch": 2.7198177676537583, + "grad_norm": 0.33462515473365784, + "learning_rate": 5e-06, + "loss": 0.0622, + "step": 1194 + }, + { + "epoch": 2.722095671981777, + "grad_norm": 0.3674948811531067, + "learning_rate": 5e-06, + "loss": 0.0621, + "step": 1195 + }, + { + "epoch": 2.724373576309795, + "grad_norm": 0.37018531560897827, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 1196 + }, + { + "epoch": 2.7266514806378135, + "grad_norm": 0.33755621314048767, + "learning_rate": 5e-06, + "loss": 0.067, + "step": 1197 + }, + { + "epoch": 2.7289293849658316, + "grad_norm": 0.41091012954711914, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 1198 + }, + { + "epoch": 2.7312072892938497, + "grad_norm": 0.41686341166496277, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 1199 + }, + { + "epoch": 2.733485193621868, + "grad_norm": 0.4030642807483673, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 1200 + }, + { + "epoch": 2.735763097949886, + "grad_norm": 0.3342956602573395, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 1201 + }, + { + "epoch": 2.7380410022779045, + "grad_norm": 0.3751448094844818, + "learning_rate": 5e-06, + "loss": 0.0664, + "step": 1202 + }, + { + "epoch": 2.7403189066059226, + "grad_norm": 0.34822869300842285, + "learning_rate": 5e-06, + "loss": 0.0645, + "step": 1203 + }, + { + "epoch": 2.7425968109339407, + "grad_norm": 0.342708945274353, + "learning_rate": 5e-06, + "loss": 0.0634, + "step": 1204 + }, + { + "epoch": 2.7448747152619593, + "grad_norm": 0.34886717796325684, + "learning_rate": 5e-06, + "loss": 0.0677, + "step": 1205 + }, + { + "epoch": 2.7471526195899774, + "grad_norm": 0.36077600717544556, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 1206 + }, + { + "epoch": 2.7494305239179955, + "grad_norm": 0.38435259461402893, + "learning_rate": 5e-06, + "loss": 0.0628, + "step": 1207 + }, + { + "epoch": 2.7517084282460136, + "grad_norm": 0.3360936641693115, + "learning_rate": 5e-06, + "loss": 0.0621, + "step": 1208 + }, + { + "epoch": 2.7539863325740317, + "grad_norm": 0.32045936584472656, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 1209 + }, + { + "epoch": 2.7562642369020502, + "grad_norm": 0.3480812907218933, + "learning_rate": 5e-06, + "loss": 0.0623, + "step": 1210 + }, + { + "epoch": 2.7585421412300684, + "grad_norm": 0.38358426094055176, + "learning_rate": 5e-06, + "loss": 0.0634, + "step": 1211 + }, + { + "epoch": 2.7608200455580865, + "grad_norm": 0.33160167932510376, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 1212 + }, + { + "epoch": 2.763097949886105, + "grad_norm": 0.3674454689025879, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 1213 + }, + { + "epoch": 2.765375854214123, + "grad_norm": 0.3301061689853668, + "learning_rate": 5e-06, + "loss": 0.0593, + "step": 1214 + }, + { + "epoch": 2.7676537585421412, + "grad_norm": 0.3309481739997864, + "learning_rate": 5e-06, + "loss": 0.0628, + "step": 1215 + }, + { + "epoch": 2.7699316628701594, + "grad_norm": 0.347108393907547, + "learning_rate": 5e-06, + "loss": 0.0621, + "step": 1216 + }, + { + "epoch": 2.7722095671981775, + "grad_norm": 0.3608108162879944, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 1217 + }, + { + "epoch": 2.774487471526196, + "grad_norm": 0.3953985273838043, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 1218 + }, + { + "epoch": 2.776765375854214, + "grad_norm": 0.3932558298110962, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 1219 + }, + { + "epoch": 2.7790432801822322, + "grad_norm": 0.3517478108406067, + "learning_rate": 5e-06, + "loss": 0.0661, + "step": 1220 + }, + { + "epoch": 2.781321184510251, + "grad_norm": 0.3522094488143921, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 1221 + }, + { + "epoch": 2.783599088838269, + "grad_norm": 0.3189462125301361, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 1222 + }, + { + "epoch": 2.785876993166287, + "grad_norm": 0.3788624703884125, + "learning_rate": 5e-06, + "loss": 0.0629, + "step": 1223 + }, + { + "epoch": 2.788154897494305, + "grad_norm": 0.3391224145889282, + "learning_rate": 5e-06, + "loss": 0.0628, + "step": 1224 + }, + { + "epoch": 2.7904328018223232, + "grad_norm": 0.36144745349884033, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 1225 + }, + { + "epoch": 2.792710706150342, + "grad_norm": 0.3475210964679718, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 1226 + }, + { + "epoch": 2.79498861047836, + "grad_norm": 0.3617037534713745, + "learning_rate": 5e-06, + "loss": 0.0651, + "step": 1227 + }, + { + "epoch": 2.797266514806378, + "grad_norm": 0.39694684743881226, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 1228 + }, + { + "epoch": 2.7995444191343966, + "grad_norm": 0.3512531816959381, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 1229 + }, + { + "epoch": 2.8018223234624147, + "grad_norm": 0.40816575288772583, + "learning_rate": 5e-06, + "loss": 0.065, + "step": 1230 + }, + { + "epoch": 2.8041002277904328, + "grad_norm": 0.3422633111476898, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 1231 + }, + { + "epoch": 2.806378132118451, + "grad_norm": 0.36634865403175354, + "learning_rate": 5e-06, + "loss": 0.0628, + "step": 1232 + }, + { + "epoch": 2.808656036446469, + "grad_norm": 0.3451927602291107, + "learning_rate": 5e-06, + "loss": 0.0609, + "step": 1233 + }, + { + "epoch": 2.8109339407744875, + "grad_norm": 0.36408817768096924, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 1234 + }, + { + "epoch": 2.8132118451025057, + "grad_norm": 0.35038718581199646, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 1235 + }, + { + "epoch": 2.8154897494305238, + "grad_norm": 0.33411678671836853, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 1236 + }, + { + "epoch": 2.8177676537585423, + "grad_norm": 0.3636171221733093, + "learning_rate": 5e-06, + "loss": 0.0652, + "step": 1237 + }, + { + "epoch": 2.8200455580865604, + "grad_norm": 0.3390803933143616, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 1238 + }, + { + "epoch": 2.8223234624145785, + "grad_norm": 0.3592444062232971, + "learning_rate": 5e-06, + "loss": 0.0641, + "step": 1239 + }, + { + "epoch": 2.8246013667425967, + "grad_norm": 0.3812445104122162, + "learning_rate": 5e-06, + "loss": 0.0651, + "step": 1240 + }, + { + "epoch": 2.8268792710706148, + "grad_norm": 0.39333441853523254, + "learning_rate": 5e-06, + "loss": 0.0658, + "step": 1241 + }, + { + "epoch": 2.8291571753986333, + "grad_norm": 0.31484416127204895, + "learning_rate": 5e-06, + "loss": 0.0625, + "step": 1242 + }, + { + "epoch": 2.8314350797266514, + "grad_norm": 0.37235206365585327, + "learning_rate": 5e-06, + "loss": 0.063, + "step": 1243 + }, + { + "epoch": 2.83371298405467, + "grad_norm": 0.31487953662872314, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 1244 + }, + { + "epoch": 2.835990888382688, + "grad_norm": 0.3140309154987335, + "learning_rate": 5e-06, + "loss": 0.0607, + "step": 1245 + }, + { + "epoch": 2.838268792710706, + "grad_norm": 0.3087141215801239, + "learning_rate": 5e-06, + "loss": 0.0625, + "step": 1246 + }, + { + "epoch": 2.8405466970387243, + "grad_norm": 0.3366031050682068, + "learning_rate": 5e-06, + "loss": 0.0617, + "step": 1247 + }, + { + "epoch": 2.8428246013667424, + "grad_norm": 0.33195042610168457, + "learning_rate": 5e-06, + "loss": 0.0617, + "step": 1248 + }, + { + "epoch": 2.845102505694761, + "grad_norm": 0.34049350023269653, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 1249 + }, + { + "epoch": 2.847380410022779, + "grad_norm": 0.308563232421875, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 1250 + }, + { + "epoch": 2.849658314350797, + "grad_norm": 0.363180547952652, + "learning_rate": 5e-06, + "loss": 0.0617, + "step": 1251 + }, + { + "epoch": 2.8519362186788157, + "grad_norm": 0.3409302830696106, + "learning_rate": 5e-06, + "loss": 0.0615, + "step": 1252 + }, + { + "epoch": 2.854214123006834, + "grad_norm": 0.33143019676208496, + "learning_rate": 5e-06, + "loss": 0.0655, + "step": 1253 + }, + { + "epoch": 2.856492027334852, + "grad_norm": 0.37796229124069214, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 1254 + }, + { + "epoch": 2.85876993166287, + "grad_norm": 0.32886597514152527, + "learning_rate": 5e-06, + "loss": 0.0603, + "step": 1255 + }, + { + "epoch": 2.861047835990888, + "grad_norm": 0.4048435688018799, + "learning_rate": 5e-06, + "loss": 0.0664, + "step": 1256 + }, + { + "epoch": 2.8633257403189067, + "grad_norm": 0.36198365688323975, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 1257 + }, + { + "epoch": 2.865603644646925, + "grad_norm": 0.5110198855400085, + "learning_rate": 5e-06, + "loss": 0.0665, + "step": 1258 + }, + { + "epoch": 2.867881548974943, + "grad_norm": 0.37884554266929626, + "learning_rate": 5e-06, + "loss": 0.0614, + "step": 1259 + }, + { + "epoch": 2.8701594533029615, + "grad_norm": 0.4248731732368469, + "learning_rate": 5e-06, + "loss": 0.0638, + "step": 1260 + }, + { + "epoch": 2.8724373576309796, + "grad_norm": 0.3942655324935913, + "learning_rate": 5e-06, + "loss": 0.0629, + "step": 1261 + }, + { + "epoch": 2.8747152619589977, + "grad_norm": 0.40612831711769104, + "learning_rate": 5e-06, + "loss": 0.0615, + "step": 1262 + }, + { + "epoch": 2.876993166287016, + "grad_norm": 0.41458389163017273, + "learning_rate": 5e-06, + "loss": 0.0627, + "step": 1263 + }, + { + "epoch": 2.879271070615034, + "grad_norm": 0.40696799755096436, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 1264 + }, + { + "epoch": 2.8815489749430525, + "grad_norm": 0.3632875978946686, + "learning_rate": 5e-06, + "loss": 0.0635, + "step": 1265 + }, + { + "epoch": 2.8838268792710706, + "grad_norm": 0.3959384858608246, + "learning_rate": 5e-06, + "loss": 0.0605, + "step": 1266 + }, + { + "epoch": 2.8861047835990887, + "grad_norm": 0.36218011379241943, + "learning_rate": 5e-06, + "loss": 0.0628, + "step": 1267 + }, + { + "epoch": 2.8883826879271073, + "grad_norm": 0.37656182050704956, + "learning_rate": 5e-06, + "loss": 0.0607, + "step": 1268 + }, + { + "epoch": 2.8906605922551254, + "grad_norm": 0.36499765515327454, + "learning_rate": 5e-06, + "loss": 0.0638, + "step": 1269 + }, + { + "epoch": 2.8929384965831435, + "grad_norm": 0.3435059189796448, + "learning_rate": 5e-06, + "loss": 0.0634, + "step": 1270 + }, + { + "epoch": 2.8952164009111616, + "grad_norm": 0.36923810839653015, + "learning_rate": 5e-06, + "loss": 0.0607, + "step": 1271 + }, + { + "epoch": 2.8974943052391797, + "grad_norm": 0.37164321541786194, + "learning_rate": 5e-06, + "loss": 0.0626, + "step": 1272 + }, + { + "epoch": 2.8997722095671983, + "grad_norm": 0.42563021183013916, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 1273 + }, + { + "epoch": 2.9020501138952164, + "grad_norm": 0.3721482753753662, + "learning_rate": 5e-06, + "loss": 0.0632, + "step": 1274 + }, + { + "epoch": 2.9043280182232345, + "grad_norm": 0.36868250370025635, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 1275 + }, + { + "epoch": 2.906605922551253, + "grad_norm": 0.3597923517227173, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 1276 + }, + { + "epoch": 2.908883826879271, + "grad_norm": 0.3550402522087097, + "learning_rate": 5e-06, + "loss": 0.0648, + "step": 1277 + }, + { + "epoch": 2.9111617312072893, + "grad_norm": 0.3539983928203583, + "learning_rate": 5e-06, + "loss": 0.0627, + "step": 1278 + }, + { + "epoch": 2.9134396355353074, + "grad_norm": 0.32854676246643066, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 1279 + }, + { + "epoch": 2.9157175398633255, + "grad_norm": 0.33244210481643677, + "learning_rate": 5e-06, + "loss": 0.0622, + "step": 1280 + }, + { + "epoch": 2.917995444191344, + "grad_norm": 0.3468360900878906, + "learning_rate": 5e-06, + "loss": 0.0634, + "step": 1281 + }, + { + "epoch": 2.920273348519362, + "grad_norm": 0.35334527492523193, + "learning_rate": 5e-06, + "loss": 0.0627, + "step": 1282 + }, + { + "epoch": 2.9225512528473803, + "grad_norm": 0.35016804933547974, + "learning_rate": 5e-06, + "loss": 0.0636, + "step": 1283 + }, + { + "epoch": 2.924829157175399, + "grad_norm": 0.4512026906013489, + "learning_rate": 5e-06, + "loss": 0.064, + "step": 1284 + }, + { + "epoch": 2.927107061503417, + "grad_norm": 0.359772264957428, + "learning_rate": 5e-06, + "loss": 0.0643, + "step": 1285 + }, + { + "epoch": 2.929384965831435, + "grad_norm": 0.3655630350112915, + "learning_rate": 5e-06, + "loss": 0.0629, + "step": 1286 + }, + { + "epoch": 2.931662870159453, + "grad_norm": 0.33043938875198364, + "learning_rate": 5e-06, + "loss": 0.0621, + "step": 1287 + }, + { + "epoch": 2.9339407744874713, + "grad_norm": 0.3892334997653961, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 1288 + }, + { + "epoch": 2.93621867881549, + "grad_norm": 0.3802751898765564, + "learning_rate": 5e-06, + "loss": 0.0612, + "step": 1289 + }, + { + "epoch": 2.938496583143508, + "grad_norm": 0.35937488079071045, + "learning_rate": 5e-06, + "loss": 0.0625, + "step": 1290 + }, + { + "epoch": 2.940774487471526, + "grad_norm": 0.36067768931388855, + "learning_rate": 5e-06, + "loss": 0.0625, + "step": 1291 + }, + { + "epoch": 2.9430523917995446, + "grad_norm": 0.33447036147117615, + "learning_rate": 5e-06, + "loss": 0.0639, + "step": 1292 + }, + { + "epoch": 2.9453302961275627, + "grad_norm": 0.36563393473625183, + "learning_rate": 5e-06, + "loss": 0.0623, + "step": 1293 + }, + { + "epoch": 2.947608200455581, + "grad_norm": 0.35415124893188477, + "learning_rate": 5e-06, + "loss": 0.0613, + "step": 1294 + }, + { + "epoch": 2.949886104783599, + "grad_norm": 0.32227981090545654, + "learning_rate": 5e-06, + "loss": 0.0596, + "step": 1295 + }, + { + "epoch": 2.9521640091116175, + "grad_norm": 0.36177361011505127, + "learning_rate": 5e-06, + "loss": 0.0609, + "step": 1296 + }, + { + "epoch": 2.9544419134396356, + "grad_norm": 0.3240223228931427, + "learning_rate": 5e-06, + "loss": 0.0626, + "step": 1297 + }, + { + "epoch": 2.9567198177676537, + "grad_norm": 0.4267748296260834, + "learning_rate": 5e-06, + "loss": 0.0618, + "step": 1298 + }, + { + "epoch": 2.9589977220956722, + "grad_norm": 0.3413817584514618, + "learning_rate": 5e-06, + "loss": 0.0649, + "step": 1299 + }, + { + "epoch": 2.9612756264236904, + "grad_norm": 0.3571028709411621, + "learning_rate": 5e-06, + "loss": 0.0613, + "step": 1300 + }, + { + "epoch": 2.9635535307517085, + "grad_norm": 0.3737579584121704, + "learning_rate": 5e-06, + "loss": 0.0637, + "step": 1301 + }, + { + "epoch": 2.9658314350797266, + "grad_norm": 0.34754571318626404, + "learning_rate": 5e-06, + "loss": 0.0659, + "step": 1302 + }, + { + "epoch": 2.9681093394077447, + "grad_norm": 0.3502677381038666, + "learning_rate": 5e-06, + "loss": 0.0623, + "step": 1303 + }, + { + "epoch": 2.9703872437357632, + "grad_norm": 0.3277047872543335, + "learning_rate": 5e-06, + "loss": 0.0651, + "step": 1304 + }, + { + "epoch": 2.9726651480637813, + "grad_norm": 0.3437495231628418, + "learning_rate": 5e-06, + "loss": 0.0644, + "step": 1305 + }, + { + "epoch": 2.9749430523917995, + "grad_norm": 0.33111703395843506, + "learning_rate": 5e-06, + "loss": 0.0646, + "step": 1306 + }, + { + "epoch": 2.977220956719818, + "grad_norm": 0.35934245586395264, + "learning_rate": 5e-06, + "loss": 0.0654, + "step": 1307 + }, + { + "epoch": 2.979498861047836, + "grad_norm": 0.331857293844223, + "learning_rate": 5e-06, + "loss": 0.0653, + "step": 1308 + }, + { + "epoch": 2.9817767653758542, + "grad_norm": 0.3608936667442322, + "learning_rate": 5e-06, + "loss": 0.0642, + "step": 1309 + }, + { + "epoch": 2.9840546697038723, + "grad_norm": 0.3277752697467804, + "learning_rate": 5e-06, + "loss": 0.0623, + "step": 1310 + }, + { + "epoch": 2.9863325740318905, + "grad_norm": 0.37028080224990845, + "learning_rate": 5e-06, + "loss": 0.0657, + "step": 1311 + }, + { + "epoch": 2.988610478359909, + "grad_norm": 0.3204702138900757, + "learning_rate": 5e-06, + "loss": 0.0616, + "step": 1312 + }, + { + "epoch": 2.990888382687927, + "grad_norm": 0.3566315770149231, + "learning_rate": 5e-06, + "loss": 0.062, + "step": 1313 + }, + { + "epoch": 2.9931662870159452, + "grad_norm": 0.32704809308052063, + "learning_rate": 5e-06, + "loss": 0.0622, + "step": 1314 + }, + { + "epoch": 2.995444191343964, + "grad_norm": 0.3317294716835022, + "learning_rate": 5e-06, + "loss": 0.0622, + "step": 1315 + }, + { + "epoch": 2.997722095671982, + "grad_norm": 0.3896945118904114, + "learning_rate": 5e-06, + "loss": 0.0684, + "step": 1316 + }, + { + "epoch": 3.0, + "grad_norm": 0.3358425796031952, + "learning_rate": 5e-06, + "loss": 0.0621, + "step": 1317 + }, + { + "epoch": 3.0, + "step": 1317, + "total_flos": 0.0, + "train_loss": 0.06867336523107444, + "train_runtime": 41172.7067, + "train_samples_per_second": 16.404, + "train_steps_per_second": 0.032 + } + ], + "logging_steps": 1, + "max_steps": 1317, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 132, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 0.0, + "train_batch_size": 64, + "trial_name": null, + "trial_params": null +} diff --git a/training_args.bin b/training_args.bin new file mode 100644 index 0000000..bbf61cc --- /dev/null +++ b/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b92e8d4d1d3d10e7a6115344fa522f539106f607d4cdb7e30b9ba7f8e047ad53 +size 9425