From 94a68644cd50e7f100b90724af8718616f819d74 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Mon, 13 Apr 2026 17:33:03 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: FlyPig23/Llama3.2-3B_Paper_Impact_SFT Source: Original Platform --- .gitattributes | 36 + README.md | 69 + all_results.json | 12 + chat_template.jinja | 93 + config.json | 36 + eval_results.json | 7 + generation_config.json | 14 + model-00001-of-00002.safetensors | 3 + model-00002-of-00002.safetensors | 3 + model.safetensors.index.json | 263 +++ special_tokens_map.json | 26 + tokenizer.json | 3 + tokenizer_config.json | 2068 +++++++++++++++++++++ train_results.json | 8 + trainer_log.jsonl | 420 +++++ trainer_state.json | 2980 ++++++++++++++++++++++++++++++ training_args.bin | 3 + 17 files changed, 6044 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 all_results.json create mode 100644 chat_template.jinja create mode 100644 config.json create mode 100644 eval_results.json create mode 100644 generation_config.json create mode 100644 model-00001-of-00002.safetensors create mode 100644 model-00002-of-00002.safetensors create mode 100644 model.safetensors.index.json create mode 100644 special_tokens_map.json create mode 100644 tokenizer.json create mode 100644 tokenizer_config.json create mode 100644 train_results.json create mode 100644 trainer_log.jsonl create mode 100644 trainer_state.json create mode 100644 training_args.bin diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..52373fe --- /dev/null +++ b/.gitattributes @@ -0,0 +1,36 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +tokenizer.json filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..f5d06a8 --- /dev/null +++ b/README.md @@ -0,0 +1,69 @@ +--- +library_name: transformers +license: other +base_model: meta-llama/Llama-3.2-3B-Instruct +tags: +- llama-factory +- full +- generated_from_trainer +model-index: +- name: Llama3.2-3B_Paper_Impact_SFT + results: [] +--- + + + +# Llama3.2-3B_Paper_Impact_SFT + +This model is a fine-tuned version of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) on the paper_impact_sft_train dataset. +It achieves the following results on the evaluation set: +- Loss: 0.1446 + +## Model description + +More information needed + +## Intended uses & limitations + +More information needed + +## Training and evaluation data + +More information needed + +## Training procedure + +### Training hyperparameters + +The following hyperparameters were used during training: +- learning_rate: 2e-05 +- train_batch_size: 8 +- eval_batch_size: 8 +- seed: 42 +- distributed_type: multi-GPU +- num_devices: 4 +- gradient_accumulation_steps: 4 +- total_train_batch_size: 128 +- total_eval_batch_size: 32 +- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments +- lr_scheduler_type: cosine +- lr_scheduler_warmup_ratio: 0.1 +- num_epochs: 3.0 + +### Training results + +| Training Loss | Epoch | Step | Validation Loss | +|:-------------:|:------:|:----:|:---------------:| +| 0.0607 | 0.7228 | 500 | 0.0733 | +| 0.029 | 1.4452 | 1000 | 0.0819 | +| 0.0058 | 2.1677 | 1500 | 0.1524 | +| 0.005 | 2.8905 | 2000 | 0.1443 | + + +### Framework versions + +- Transformers 4.57.1 +- Pytorch 2.6.0+cu124 +- Datasets 4.0.0 +- Tokenizers 0.22.1 diff --git a/all_results.json b/all_results.json new file mode 100644 index 0000000..3c1fc9d --- /dev/null +++ b/all_results.json @@ -0,0 +1,12 @@ +{ + "epoch": 3.0, + "eval_loss": 0.1446007937192917, + "eval_runtime": 1173.577, + "eval_samples_per_second": 54.71, + "eval_steps_per_second": 1.71, + "total_flos": 6.274047864041636e+18, + "train_loss": 0.03664169063040653, + "train_runtime": 42810.6565, + "train_samples_per_second": 6.203, + "train_steps_per_second": 0.048 +} \ No newline at end of file diff --git a/chat_template.jinja b/chat_template.jinja new file mode 100644 index 0000000..1bad6a0 --- /dev/null +++ b/chat_template.jinja @@ -0,0 +1,93 @@ +{{- bos_token }} +{%- if custom_tools is defined %} + {%- set tools = custom_tools %} +{%- endif %} +{%- if not tools_in_user_message is defined %} + {%- set tools_in_user_message = true %} +{%- endif %} +{%- if not date_string is defined %} + {%- if strftime_now is defined %} + {%- set date_string = strftime_now("%d %b %Y") %} + {%- else %} + {%- set date_string = "26 Jul 2024" %} + {%- endif %} +{%- endif %} +{%- if not tools is defined %} + {%- set tools = none %} +{%- endif %} + +{#- This block extracts the system message, so we can slot it into the right place. #} +{%- if messages[0]['role'] == 'system' %} + {%- set system_message = messages[0]['content']|trim %} + {%- set messages = messages[1:] %} +{%- else %} + {%- set system_message = "" %} +{%- endif %} + +{#- System message #} +{{- "<|start_header_id|>system<|end_header_id|>\n\n" }} +{%- if tools is not none %} + {{- "Environment: ipython\n" }} +{%- endif %} +{{- "Cutting Knowledge Date: December 2023\n" }} +{{- "Today Date: " + date_string + "\n\n" }} +{%- if tools is not none and not tools_in_user_message %} + {{- "You have access to the following functions. To call a function, please respond with JSON for a function call." }} + {{- 'Respond in the format {"name": function name, "parameters": dictionary of argument name and its value}.' }} + {{- "Do not use variables.\n\n" }} + {%- for t in tools %} + {{- t | tojson(indent=4) }} + {{- "\n\n" }} + {%- endfor %} +{%- endif %} +{{- system_message }} +{{- "<|eot_id|>" }} + +{#- Custom tools are passed in a user message with some extra guidance #} +{%- if tools_in_user_message and not tools is none %} + {#- Extract the first user message so we can plug it in here #} + {%- if messages | length != 0 %} + {%- set first_user_message = messages[0]['content']|trim %} + {%- set messages = messages[1:] %} + {%- else %} + {{- raise_exception("Cannot put tools in the first user message when there's no first user message!") }} +{%- endif %} + {{- '<|start_header_id|>user<|end_header_id|>\n\n' -}} + {{- "Given the following functions, please respond with a JSON for a function call " }} + {{- "with its proper arguments that best answers the given prompt.\n\n" }} + {{- 'Respond in the format {"name": function name, "parameters": dictionary of argument name and its value}.' }} + {{- "Do not use variables.\n\n" }} + {%- for t in tools %} + {{- t | tojson(indent=4) }} + {{- "\n\n" }} + {%- endfor %} + {{- first_user_message + "<|eot_id|>"}} +{%- endif %} + +{%- for message in messages %} + {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %} + {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' }} + {%- elif 'tool_calls' in message %} + {%- if not message.tool_calls|length == 1 %} + {{- raise_exception("This model only supports single tool-calls at once!") }} + {%- endif %} + {%- set tool_call = message.tool_calls[0].function %} + {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' -}} + {{- '{"name": "' + tool_call.name + '", ' }} + {{- '"parameters": ' }} + {{- tool_call.arguments | tojson }} + {{- "}" }} + {{- "<|eot_id|>" }} + {%- elif message.role == "tool" or message.role == "ipython" %} + {{- "<|start_header_id|>ipython<|end_header_id|>\n\n" }} + {%- if message.content is mapping or message.content is iterable %} + {{- message.content | tojson }} + {%- else %} + {{- message.content }} + {%- endif %} + {{- "<|eot_id|>" }} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|start_header_id|>assistant<|end_header_id|>\n\n' }} +{%- endif %} diff --git a/config.json b/config.json new file mode 100644 index 0000000..badb0f3 --- /dev/null +++ b/config.json @@ -0,0 +1,36 @@ +{ + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 128000, + "dtype": "bfloat16", + "eos_token_id": 128009, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 3072, + "initializer_range": 0.02, + "intermediate_size": 8192, + "max_position_embeddings": 131072, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 24, + "num_hidden_layers": 28, + "num_key_value_heads": 8, + "pad_token_id": 128009, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": { + "factor": 32.0, + "high_freq_factor": 4.0, + "low_freq_factor": 1.0, + "original_max_position_embeddings": 8192, + "rope_type": "llama3" + }, + "rope_theta": 500000.0, + "tie_word_embeddings": true, + "transformers_version": "4.57.1", + "use_cache": false, + "vocab_size": 128256 +} diff --git a/eval_results.json b/eval_results.json new file mode 100644 index 0000000..1c70b86 --- /dev/null +++ b/eval_results.json @@ -0,0 +1,7 @@ +{ + "epoch": 3.0, + "eval_loss": 0.1446007937192917, + "eval_runtime": 1173.577, + "eval_samples_per_second": 54.71, + "eval_steps_per_second": 1.71 +} \ No newline at end of file diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000..3459a6d --- /dev/null +++ b/generation_config.json @@ -0,0 +1,14 @@ +{ + "bos_token_id": 128000, + "do_sample": true, + "eos_token_id": [ + 128009, + 128001, + 128008, + 128009 + ], + "pad_token_id": 128009, + "temperature": 0.6, + "top_p": 0.9, + "transformers_version": "4.57.1" +} diff --git a/model-00001-of-00002.safetensors b/model-00001-of-00002.safetensors new file mode 100644 index 0000000..da55290 --- /dev/null +++ b/model-00001-of-00002.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:70b069115627dc3bfced60ba6c35f4554b9217422734022c58e316156a985e6e +size 4965799096 diff --git a/model-00002-of-00002.safetensors b/model-00002-of-00002.safetensors new file mode 100644 index 0000000..f785942 --- /dev/null +++ b/model-00002-of-00002.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:366bdbfd0baceb9b04da2118a8e07f4cddbb04473525ab28725d44070bd67d2c +size 2247734992 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000..887d493 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,263 @@ +{ + "metadata": { + "total_parameters": 3212749824, + "total_size": 7213504512 + }, + "weight_map": { + "lm_head.weight": "model-00002-of-00002.safetensors", + "model.embed_tokens.weight": "model-00001-of-00002.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.1.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.10.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.11.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.12.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.13.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.14.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.15.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.16.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.17.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.18.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.19.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.20.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.21.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.22.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.23.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.24.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.25.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.26.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.27.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.4.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.5.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.6.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.7.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.8.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.9.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.norm.weight": "model-00002-of-00002.safetensors" + } +} diff --git a/special_tokens_map.json b/special_tokens_map.json new file mode 100644 index 0000000..14daf45 --- /dev/null +++ b/special_tokens_map.json @@ -0,0 +1,26 @@ +{ + "additional_special_tokens": [ + { + "content": "<|eom_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } + ], + "bos_token": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": "<|eot_id|>" +} diff --git a/tokenizer.json b/tokenizer.json new file mode 100644 index 0000000..1c1d8d5 --- /dev/null +++ b/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b +size 17209920 diff --git a/tokenizer_config.json b/tokenizer_config.json new file mode 100644 index 0000000..d1e1ea9 --- /dev/null +++ b/tokenizer_config.json @@ -0,0 +1,2068 @@ +{ + "added_tokens_decoder": { + "128000": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128001": { + "content": "<|end_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128002": { + "content": "<|reserved_special_token_0|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128003": { + "content": "<|reserved_special_token_1|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128004": { + "content": "<|finetune_right_pad_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128005": { + "content": "<|reserved_special_token_2|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128006": { + "content": "<|start_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128007": { + "content": "<|end_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128008": { + "content": "<|eom_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128009": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128010": { + "content": "<|python_tag|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128011": { + "content": "<|reserved_special_token_3|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128012": { + "content": "<|reserved_special_token_4|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128013": { + "content": "<|reserved_special_token_5|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128014": { + "content": "<|reserved_special_token_6|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128015": { + "content": "<|reserved_special_token_7|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128016": { + "content": "<|reserved_special_token_8|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128017": { + "content": "<|reserved_special_token_9|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128018": { + "content": "<|reserved_special_token_10|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128019": { + "content": "<|reserved_special_token_11|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128020": { + "content": "<|reserved_special_token_12|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128021": { + "content": "<|reserved_special_token_13|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128022": { + "content": "<|reserved_special_token_14|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128023": { + "content": "<|reserved_special_token_15|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128024": { + "content": "<|reserved_special_token_16|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128025": { + "content": "<|reserved_special_token_17|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128026": { + "content": "<|reserved_special_token_18|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128027": { + "content": "<|reserved_special_token_19|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128028": { + "content": "<|reserved_special_token_20|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128029": { + "content": "<|reserved_special_token_21|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128030": { + "content": "<|reserved_special_token_22|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128031": { + "content": "<|reserved_special_token_23|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128032": { + "content": "<|reserved_special_token_24|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128033": { + "content": "<|reserved_special_token_25|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128034": { + "content": "<|reserved_special_token_26|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128035": { + "content": "<|reserved_special_token_27|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128036": { + "content": "<|reserved_special_token_28|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128037": { + "content": "<|reserved_special_token_29|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128038": { + "content": "<|reserved_special_token_30|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128039": { + "content": "<|reserved_special_token_31|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128040": { + "content": "<|reserved_special_token_32|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128041": { + "content": "<|reserved_special_token_33|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128042": { + "content": "<|reserved_special_token_34|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128043": { + "content": "<|reserved_special_token_35|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128044": { + "content": "<|reserved_special_token_36|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128045": { + "content": "<|reserved_special_token_37|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128046": { + "content": "<|reserved_special_token_38|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128047": { + "content": "<|reserved_special_token_39|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128048": { + "content": "<|reserved_special_token_40|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128049": { + "content": "<|reserved_special_token_41|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128050": { + "content": "<|reserved_special_token_42|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128051": { + "content": "<|reserved_special_token_43|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128052": { + "content": "<|reserved_special_token_44|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128053": { + "content": "<|reserved_special_token_45|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128054": { + "content": "<|reserved_special_token_46|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128055": { + "content": "<|reserved_special_token_47|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128056": { + "content": "<|reserved_special_token_48|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128057": { + "content": "<|reserved_special_token_49|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128058": { + "content": "<|reserved_special_token_50|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128059": { + "content": "<|reserved_special_token_51|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128060": { + "content": "<|reserved_special_token_52|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128061": { + "content": "<|reserved_special_token_53|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128062": { + "content": "<|reserved_special_token_54|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128063": { + "content": "<|reserved_special_token_55|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128064": { + "content": "<|reserved_special_token_56|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128065": { + "content": "<|reserved_special_token_57|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128066": { + "content": "<|reserved_special_token_58|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128067": { + "content": "<|reserved_special_token_59|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128068": { + "content": "<|reserved_special_token_60|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128069": { + "content": "<|reserved_special_token_61|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128070": { + "content": "<|reserved_special_token_62|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128071": { + "content": "<|reserved_special_token_63|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128072": { + "content": "<|reserved_special_token_64|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128073": { + "content": "<|reserved_special_token_65|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128074": { + "content": "<|reserved_special_token_66|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128075": { + "content": "<|reserved_special_token_67|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128076": { + "content": "<|reserved_special_token_68|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128077": { + "content": "<|reserved_special_token_69|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128078": { + "content": "<|reserved_special_token_70|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128079": { + "content": "<|reserved_special_token_71|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128080": { + "content": "<|reserved_special_token_72|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128081": { + "content": "<|reserved_special_token_73|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128082": { + "content": "<|reserved_special_token_74|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128083": { + "content": "<|reserved_special_token_75|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128084": { + "content": "<|reserved_special_token_76|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128085": { + "content": "<|reserved_special_token_77|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128086": { + "content": "<|reserved_special_token_78|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128087": { + "content": "<|reserved_special_token_79|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128088": { + "content": "<|reserved_special_token_80|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128089": { + "content": "<|reserved_special_token_81|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128090": { + "content": "<|reserved_special_token_82|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128091": { + "content": "<|reserved_special_token_83|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128092": { + "content": "<|reserved_special_token_84|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128093": { + "content": "<|reserved_special_token_85|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128094": { + "content": "<|reserved_special_token_86|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128095": { + "content": "<|reserved_special_token_87|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128096": { + "content": "<|reserved_special_token_88|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128097": { + "content": "<|reserved_special_token_89|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128098": { + "content": "<|reserved_special_token_90|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128099": { + "content": "<|reserved_special_token_91|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128100": { + "content": "<|reserved_special_token_92|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128101": { + "content": "<|reserved_special_token_93|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128102": { + "content": "<|reserved_special_token_94|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128103": { + "content": "<|reserved_special_token_95|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128104": { + "content": "<|reserved_special_token_96|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128105": { + "content": "<|reserved_special_token_97|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128106": { + "content": "<|reserved_special_token_98|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128107": { + "content": "<|reserved_special_token_99|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128108": { + "content": "<|reserved_special_token_100|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128109": { + "content": "<|reserved_special_token_101|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128110": { + "content": "<|reserved_special_token_102|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128111": { + "content": "<|reserved_special_token_103|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128112": { + "content": "<|reserved_special_token_104|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128113": { + "content": "<|reserved_special_token_105|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128114": { + "content": "<|reserved_special_token_106|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128115": { + "content": "<|reserved_special_token_107|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128116": { + "content": "<|reserved_special_token_108|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128117": { + "content": "<|reserved_special_token_109|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128118": { + "content": "<|reserved_special_token_110|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128119": { + "content": "<|reserved_special_token_111|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128120": { + "content": "<|reserved_special_token_112|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128121": { + "content": "<|reserved_special_token_113|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128122": { + "content": "<|reserved_special_token_114|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128123": { + "content": "<|reserved_special_token_115|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128124": { + "content": "<|reserved_special_token_116|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128125": { + "content": "<|reserved_special_token_117|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128126": { + "content": "<|reserved_special_token_118|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128127": { + "content": "<|reserved_special_token_119|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128128": { + "content": "<|reserved_special_token_120|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128129": { + "content": "<|reserved_special_token_121|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128130": { + "content": "<|reserved_special_token_122|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128131": { + "content": "<|reserved_special_token_123|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128132": { + "content": "<|reserved_special_token_124|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128133": { + "content": "<|reserved_special_token_125|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128134": { + "content": "<|reserved_special_token_126|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128135": { + "content": "<|reserved_special_token_127|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128136": { + "content": "<|reserved_special_token_128|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128137": { + "content": "<|reserved_special_token_129|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128138": { + "content": "<|reserved_special_token_130|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128139": { + "content": "<|reserved_special_token_131|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128140": { + "content": "<|reserved_special_token_132|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128141": { + "content": "<|reserved_special_token_133|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128142": { + "content": "<|reserved_special_token_134|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128143": { + "content": "<|reserved_special_token_135|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128144": { + "content": "<|reserved_special_token_136|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128145": { + "content": "<|reserved_special_token_137|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128146": { + "content": "<|reserved_special_token_138|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128147": { + "content": "<|reserved_special_token_139|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128148": { + "content": "<|reserved_special_token_140|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128149": { + "content": "<|reserved_special_token_141|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128150": { + "content": "<|reserved_special_token_142|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128151": { + "content": "<|reserved_special_token_143|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128152": { + "content": "<|reserved_special_token_144|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128153": { + "content": "<|reserved_special_token_145|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128154": { + "content": "<|reserved_special_token_146|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128155": { + "content": "<|reserved_special_token_147|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128156": { + "content": "<|reserved_special_token_148|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128157": { + "content": "<|reserved_special_token_149|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128158": { + "content": "<|reserved_special_token_150|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128159": { + "content": "<|reserved_special_token_151|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128160": { + "content": "<|reserved_special_token_152|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128161": { + "content": "<|reserved_special_token_153|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128162": { + "content": "<|reserved_special_token_154|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128163": { + "content": "<|reserved_special_token_155|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128164": { + "content": "<|reserved_special_token_156|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128165": { + "content": "<|reserved_special_token_157|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128166": { + "content": "<|reserved_special_token_158|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128167": { + "content": "<|reserved_special_token_159|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128168": { + "content": "<|reserved_special_token_160|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128169": { + "content": "<|reserved_special_token_161|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128170": { + "content": "<|reserved_special_token_162|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128171": { + "content": "<|reserved_special_token_163|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128172": { + "content": "<|reserved_special_token_164|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128173": { + "content": "<|reserved_special_token_165|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128174": { + "content": "<|reserved_special_token_166|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128175": { + "content": "<|reserved_special_token_167|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128176": { + "content": "<|reserved_special_token_168|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128177": { + "content": "<|reserved_special_token_169|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128178": { + "content": "<|reserved_special_token_170|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128179": { + "content": "<|reserved_special_token_171|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128180": { + "content": "<|reserved_special_token_172|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128181": { + "content": "<|reserved_special_token_173|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128182": { + "content": "<|reserved_special_token_174|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128183": { + "content": "<|reserved_special_token_175|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128184": { + "content": "<|reserved_special_token_176|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128185": { + "content": "<|reserved_special_token_177|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128186": { + "content": "<|reserved_special_token_178|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128187": { + "content": "<|reserved_special_token_179|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128188": { + "content": "<|reserved_special_token_180|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128189": { + "content": "<|reserved_special_token_181|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128190": { + "content": "<|reserved_special_token_182|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128191": { + "content": "<|reserved_special_token_183|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128192": { + "content": "<|reserved_special_token_184|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128193": { + "content": "<|reserved_special_token_185|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128194": { + "content": "<|reserved_special_token_186|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128195": { + "content": "<|reserved_special_token_187|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128196": { + "content": "<|reserved_special_token_188|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128197": { + "content": "<|reserved_special_token_189|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128198": { + "content": "<|reserved_special_token_190|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128199": { + "content": "<|reserved_special_token_191|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128200": { + "content": "<|reserved_special_token_192|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128201": { + "content": "<|reserved_special_token_193|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128202": { + "content": "<|reserved_special_token_194|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128203": { + "content": "<|reserved_special_token_195|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128204": { + "content": "<|reserved_special_token_196|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128205": { + "content": "<|reserved_special_token_197|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128206": { + "content": "<|reserved_special_token_198|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128207": { + "content": "<|reserved_special_token_199|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128208": { + "content": "<|reserved_special_token_200|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128209": { + "content": "<|reserved_special_token_201|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128210": { + "content": "<|reserved_special_token_202|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128211": { + "content": "<|reserved_special_token_203|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128212": { + "content": "<|reserved_special_token_204|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128213": { + "content": "<|reserved_special_token_205|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128214": { + "content": "<|reserved_special_token_206|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128215": { + "content": "<|reserved_special_token_207|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128216": { + "content": "<|reserved_special_token_208|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128217": { + "content": "<|reserved_special_token_209|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128218": { + "content": "<|reserved_special_token_210|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128219": { + "content": "<|reserved_special_token_211|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128220": { + "content": "<|reserved_special_token_212|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128221": { + "content": "<|reserved_special_token_213|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128222": { + "content": "<|reserved_special_token_214|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128223": { + "content": "<|reserved_special_token_215|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128224": { + "content": "<|reserved_special_token_216|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128225": { + "content": "<|reserved_special_token_217|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128226": { + "content": "<|reserved_special_token_218|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128227": { + "content": "<|reserved_special_token_219|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128228": { + "content": "<|reserved_special_token_220|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128229": { + "content": "<|reserved_special_token_221|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128230": { + "content": "<|reserved_special_token_222|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128231": { + "content": "<|reserved_special_token_223|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128232": { + "content": "<|reserved_special_token_224|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128233": { + "content": "<|reserved_special_token_225|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128234": { + "content": "<|reserved_special_token_226|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128235": { + "content": "<|reserved_special_token_227|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128236": { + "content": "<|reserved_special_token_228|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128237": { + "content": "<|reserved_special_token_229|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128238": { + "content": "<|reserved_special_token_230|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128239": { + "content": "<|reserved_special_token_231|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128240": { + "content": "<|reserved_special_token_232|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128241": { + "content": "<|reserved_special_token_233|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128242": { + "content": "<|reserved_special_token_234|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128243": { + "content": "<|reserved_special_token_235|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128244": { + "content": "<|reserved_special_token_236|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128245": { + "content": "<|reserved_special_token_237|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128246": { + "content": "<|reserved_special_token_238|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128247": { + "content": "<|reserved_special_token_239|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128248": { + "content": "<|reserved_special_token_240|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128249": { + "content": "<|reserved_special_token_241|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128250": { + "content": "<|reserved_special_token_242|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128251": { + "content": "<|reserved_special_token_243|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128252": { + "content": "<|reserved_special_token_244|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128253": { + "content": "<|reserved_special_token_245|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128254": { + "content": "<|reserved_special_token_246|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128255": { + "content": "<|reserved_special_token_247|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "additional_special_tokens": [ + "<|eom_id|>" + ], + "bos_token": "<|begin_of_text|>", + "clean_up_tokenization_spaces": true, + "eos_token": "<|eot_id|>", + "extra_special_tokens": {}, + "model_input_names": [ + "input_ids", + "attention_mask" + ], + "model_max_length": 131072, + "pad_token": "<|eot_id|>", + "padding_side": "right", + "split_special_tokens": false, + "tokenizer_class": "PreTrainedTokenizerFast" +} diff --git a/train_results.json b/train_results.json new file mode 100644 index 0000000..490f2cb --- /dev/null +++ b/train_results.json @@ -0,0 +1,8 @@ +{ + "epoch": 3.0, + "total_flos": 6.274047864041636e+18, + "train_loss": 0.03664169063040653, + "train_runtime": 42810.6565, + "train_samples_per_second": 6.203, + "train_steps_per_second": 0.048 +} \ No newline at end of file diff --git a/trainer_log.jsonl b/trainer_log.jsonl new file mode 100644 index 0000000..ca72108 --- /dev/null +++ b/trainer_log.jsonl @@ -0,0 +1,420 @@ +{"current_steps": 5, "total_steps": 2076, "loss": 0.2014, "lr": 3.846153846153847e-07, "epoch": 0.007228044813877846, "percentage": 0.24, "elapsed_time": "0:01:31", "remaining_time": "10:33:50"} +{"current_steps": 10, "total_steps": 2076, "loss": 0.1441, "lr": 8.653846153846154e-07, "epoch": 0.014456089627755691, "percentage": 0.48, "elapsed_time": "0:03:05", "remaining_time": "10:38:05"} +{"current_steps": 15, "total_steps": 2076, "loss": 0.0935, "lr": 1.3461538461538462e-06, "epoch": 0.02168413444163354, "percentage": 0.72, "elapsed_time": "0:04:39", "remaining_time": "10:40:05"} +{"current_steps": 20, "total_steps": 2076, "loss": 0.0841, "lr": 1.826923076923077e-06, "epoch": 0.028912179255511383, "percentage": 0.96, "elapsed_time": "0:06:11", "remaining_time": "10:35:52"} +{"current_steps": 25, "total_steps": 2076, "loss": 0.0844, "lr": 2.307692307692308e-06, "epoch": 0.03614022406938923, "percentage": 1.2, "elapsed_time": "0:07:38", "remaining_time": "10:26:48"} +{"current_steps": 30, "total_steps": 2076, "loss": 0.0809, "lr": 2.7884615384615386e-06, "epoch": 0.04336826888326708, "percentage": 1.45, "elapsed_time": "0:09:09", "remaining_time": "10:24:53"} +{"current_steps": 35, "total_steps": 2076, "loss": 0.0819, "lr": 3.2692307692307696e-06, "epoch": 0.05059631369714492, "percentage": 1.69, "elapsed_time": "0:10:40", "remaining_time": "10:22:48"} +{"current_steps": 40, "total_steps": 2076, "loss": 0.0803, "lr": 3.7500000000000005e-06, "epoch": 0.057824358511022765, "percentage": 1.93, "elapsed_time": "0:12:06", "remaining_time": "10:16:27"} +{"current_steps": 45, "total_steps": 2076, "loss": 0.0813, "lr": 4.230769230769231e-06, "epoch": 0.06505240332490062, "percentage": 2.17, "elapsed_time": "0:13:34", "remaining_time": "10:13:01"} +{"current_steps": 50, "total_steps": 2076, "loss": 0.0801, "lr": 4.711538461538462e-06, "epoch": 0.07228044813877846, "percentage": 2.41, "elapsed_time": "0:15:00", "remaining_time": "10:08:18"} +{"current_steps": 55, "total_steps": 2076, "loss": 0.0811, "lr": 5.192307692307693e-06, "epoch": 0.0795084929526563, "percentage": 2.65, "elapsed_time": "0:16:28", "remaining_time": "10:05:10"} +{"current_steps": 60, "total_steps": 2076, "loss": 0.0806, "lr": 5.6730769230769235e-06, "epoch": 0.08673653776653416, "percentage": 2.89, "elapsed_time": "0:17:56", "remaining_time": "10:02:47"} +{"current_steps": 65, "total_steps": 2076, "loss": 0.0808, "lr": 6.153846153846155e-06, "epoch": 0.093964582580412, "percentage": 3.13, "elapsed_time": "0:19:24", "remaining_time": "10:00:40"} +{"current_steps": 70, "total_steps": 2076, "loss": 0.0808, "lr": 6.6346153846153846e-06, "epoch": 0.10119262739428984, "percentage": 3.37, "elapsed_time": "0:20:53", "remaining_time": "9:58:55"} +{"current_steps": 75, "total_steps": 2076, "loss": 0.0805, "lr": 7.115384615384616e-06, "epoch": 0.10842067220816769, "percentage": 3.61, "elapsed_time": "0:22:16", "remaining_time": "9:54:11"} +{"current_steps": 80, "total_steps": 2076, "loss": 0.0807, "lr": 7.5961538461538465e-06, "epoch": 0.11564871702204553, "percentage": 3.85, "elapsed_time": "0:23:50", "remaining_time": "9:54:50"} +{"current_steps": 85, "total_steps": 2076, "loss": 0.0809, "lr": 8.076923076923077e-06, "epoch": 0.12287676183592339, "percentage": 4.09, "elapsed_time": "0:25:24", "remaining_time": "9:55:14"} +{"current_steps": 90, "total_steps": 2076, "loss": 0.0812, "lr": 8.557692307692308e-06, "epoch": 0.13010480664980123, "percentage": 4.34, "elapsed_time": "0:26:53", "remaining_time": "9:53:18"} +{"current_steps": 95, "total_steps": 2076, "loss": 0.08, "lr": 9.03846153846154e-06, "epoch": 0.13733285146367907, "percentage": 4.58, "elapsed_time": "0:28:25", "remaining_time": "9:52:37"} +{"current_steps": 100, "total_steps": 2076, "loss": 0.08, "lr": 9.51923076923077e-06, "epoch": 0.14456089627755692, "percentage": 4.82, "elapsed_time": "0:29:55", "remaining_time": "9:51:18"} +{"current_steps": 105, "total_steps": 2076, "loss": 0.0817, "lr": 1e-05, "epoch": 0.15178894109143476, "percentage": 5.06, "elapsed_time": "0:31:25", "remaining_time": "9:49:50"} +{"current_steps": 110, "total_steps": 2076, "loss": 0.0805, "lr": 1.0480769230769232e-05, "epoch": 0.1590169859053126, "percentage": 5.3, "elapsed_time": "0:32:57", "remaining_time": "9:49:00"} +{"current_steps": 115, "total_steps": 2076, "loss": 0.0797, "lr": 1.0961538461538464e-05, "epoch": 0.16624503071919045, "percentage": 5.54, "elapsed_time": "0:34:31", "remaining_time": "9:48:37"} +{"current_steps": 120, "total_steps": 2076, "loss": 0.0808, "lr": 1.1442307692307693e-05, "epoch": 0.17347307553306832, "percentage": 5.78, "elapsed_time": "0:36:03", "remaining_time": "9:47:51"} +{"current_steps": 125, "total_steps": 2076, "loss": 0.0805, "lr": 1.1923076923076925e-05, "epoch": 0.18070112034694616, "percentage": 6.02, "elapsed_time": "0:37:34", "remaining_time": "9:46:22"} +{"current_steps": 130, "total_steps": 2076, "loss": 0.0836, "lr": 1.2403846153846156e-05, "epoch": 0.187929165160824, "percentage": 6.26, "elapsed_time": "0:39:07", "remaining_time": "9:45:33"} +{"current_steps": 135, "total_steps": 2076, "loss": 0.0802, "lr": 1.2884615384615386e-05, "epoch": 0.19515720997470185, "percentage": 6.5, "elapsed_time": "0:40:41", "remaining_time": "9:45:06"} +{"current_steps": 140, "total_steps": 2076, "loss": 0.0811, "lr": 1.3365384615384615e-05, "epoch": 0.2023852547885797, "percentage": 6.74, "elapsed_time": "0:42:08", "remaining_time": "9:42:50"} +{"current_steps": 145, "total_steps": 2076, "loss": 0.0829, "lr": 1.3846153846153847e-05, "epoch": 0.20961329960245753, "percentage": 6.98, "elapsed_time": "0:43:42", "remaining_time": "9:41:59"} +{"current_steps": 150, "total_steps": 2076, "loss": 0.0816, "lr": 1.4326923076923078e-05, "epoch": 0.21684134441633537, "percentage": 7.23, "elapsed_time": "0:45:12", "remaining_time": "9:40:31"} +{"current_steps": 155, "total_steps": 2076, "loss": 0.0801, "lr": 1.480769230769231e-05, "epoch": 0.22406938923021322, "percentage": 7.47, "elapsed_time": "0:46:45", "remaining_time": "9:39:28"} +{"current_steps": 160, "total_steps": 2076, "loss": 0.0813, "lr": 1.528846153846154e-05, "epoch": 0.23129743404409106, "percentage": 7.71, "elapsed_time": "0:48:16", "remaining_time": "9:38:00"} +{"current_steps": 165, "total_steps": 2076, "loss": 0.0811, "lr": 1.576923076923077e-05, "epoch": 0.23852547885796893, "percentage": 7.95, "elapsed_time": "0:49:46", "remaining_time": "9:36:28"} +{"current_steps": 170, "total_steps": 2076, "loss": 0.081, "lr": 1.6250000000000002e-05, "epoch": 0.24575352367184677, "percentage": 8.19, "elapsed_time": "0:51:12", "remaining_time": "9:34:04"} +{"current_steps": 175, "total_steps": 2076, "loss": 0.0793, "lr": 1.673076923076923e-05, "epoch": 0.2529815684857246, "percentage": 8.43, "elapsed_time": "0:52:47", "remaining_time": "9:33:28"} +{"current_steps": 180, "total_steps": 2076, "loss": 0.0811, "lr": 1.7211538461538465e-05, "epoch": 0.26020961329960246, "percentage": 8.67, "elapsed_time": "0:54:21", "remaining_time": "9:32:37"} +{"current_steps": 185, "total_steps": 2076, "loss": 0.0819, "lr": 1.7692307692307694e-05, "epoch": 0.2674376581134803, "percentage": 8.91, "elapsed_time": "0:55:56", "remaining_time": "9:31:47"} +{"current_steps": 190, "total_steps": 2076, "loss": 0.0797, "lr": 1.8173076923076924e-05, "epoch": 0.27466570292735815, "percentage": 9.15, "elapsed_time": "0:57:35", "remaining_time": "9:31:40"} +{"current_steps": 195, "total_steps": 2076, "loss": 0.0792, "lr": 1.8653846153846157e-05, "epoch": 0.281893747741236, "percentage": 9.39, "elapsed_time": "0:59:00", "remaining_time": "9:29:16"} +{"current_steps": 200, "total_steps": 2076, "loss": 0.0948, "lr": 1.9134615384615387e-05, "epoch": 0.28912179255511383, "percentage": 9.63, "elapsed_time": "1:00:32", "remaining_time": "9:27:57"} +{"current_steps": 205, "total_steps": 2076, "loss": 0.0895, "lr": 1.9615384615384617e-05, "epoch": 0.2963498373689917, "percentage": 9.87, "elapsed_time": "1:02:06", "remaining_time": "9:26:54"} +{"current_steps": 210, "total_steps": 2076, "loss": 0.0818, "lr": 1.999998585783488e-05, "epoch": 0.3035778821828695, "percentage": 10.12, "elapsed_time": "1:03:39", "remaining_time": "9:25:37"} +{"current_steps": 215, "total_steps": 2076, "loss": 0.0812, "lr": 1.9999490886255767e-05, "epoch": 0.31080592699674736, "percentage": 10.36, "elapsed_time": "1:05:05", "remaining_time": "9:23:25"} +{"current_steps": 220, "total_steps": 2076, "loss": 0.0814, "lr": 1.999828884642042e-05, "epoch": 0.3180339718106252, "percentage": 10.6, "elapsed_time": "1:06:41", "remaining_time": "9:22:40"} +{"current_steps": 225, "total_steps": 2076, "loss": 0.0806, "lr": 1.9996379823325586e-05, "epoch": 0.32526201662450305, "percentage": 10.84, "elapsed_time": "1:08:10", "remaining_time": "9:20:53"} +{"current_steps": 230, "total_steps": 2076, "loss": 0.0815, "lr": 1.9993763951959107e-05, "epoch": 0.3324900614383809, "percentage": 11.08, "elapsed_time": "1:09:39", "remaining_time": "9:19:03"} +{"current_steps": 235, "total_steps": 2076, "loss": 0.0846, "lr": 1.9990441417290358e-05, "epoch": 0.3397181062522588, "percentage": 11.32, "elapsed_time": "1:11:13", "remaining_time": "9:18:01"} +{"current_steps": 240, "total_steps": 2076, "loss": 0.0849, "lr": 1.9986412454257178e-05, "epoch": 0.34694615106613663, "percentage": 11.56, "elapsed_time": "1:12:41", "remaining_time": "9:16:03"} +{"current_steps": 245, "total_steps": 2076, "loss": 0.0821, "lr": 1.998167734774926e-05, "epoch": 0.3541741958800145, "percentage": 11.8, "elapsed_time": "1:14:13", "remaining_time": "9:14:44"} +{"current_steps": 250, "total_steps": 2076, "loss": 0.0808, "lr": 1.9976236432588002e-05, "epoch": 0.3614022406938923, "percentage": 12.04, "elapsed_time": "1:15:43", "remaining_time": "9:13:02"} +{"current_steps": 255, "total_steps": 2076, "loss": 0.0805, "lr": 1.997009009350283e-05, "epoch": 0.36863028550777016, "percentage": 12.28, "elapsed_time": "1:17:16", "remaining_time": "9:11:51"} +{"current_steps": 260, "total_steps": 2076, "loss": 0.081, "lr": 1.996323876510399e-05, "epoch": 0.375858330321648, "percentage": 12.52, "elapsed_time": "1:18:47", "remaining_time": "9:10:21"} +{"current_steps": 265, "total_steps": 2076, "loss": 0.083, "lr": 1.9955682931851835e-05, "epoch": 0.38308637513552585, "percentage": 12.76, "elapsed_time": "1:20:17", "remaining_time": "9:08:41"} +{"current_steps": 270, "total_steps": 2076, "loss": 0.0815, "lr": 1.994742312802255e-05, "epoch": 0.3903144199494037, "percentage": 13.01, "elapsed_time": "1:21:46", "remaining_time": "9:06:57"} +{"current_steps": 275, "total_steps": 2076, "loss": 0.0801, "lr": 1.993845993767038e-05, "epoch": 0.39754246476328153, "percentage": 13.25, "elapsed_time": "1:23:16", "remaining_time": "9:05:23"} +{"current_steps": 280, "total_steps": 2076, "loss": 0.0857, "lr": 1.9928793994586323e-05, "epoch": 0.4047705095771594, "percentage": 13.49, "elapsed_time": "1:24:47", "remaining_time": "9:03:53"} +{"current_steps": 285, "total_steps": 2076, "loss": 0.0822, "lr": 1.9918425982253335e-05, "epoch": 0.4119985543910372, "percentage": 13.73, "elapsed_time": "1:26:20", "remaining_time": "9:02:32"} +{"current_steps": 290, "total_steps": 2076, "loss": 0.0813, "lr": 1.9907356633797978e-05, "epoch": 0.41922659920491506, "percentage": 13.97, "elapsed_time": "1:27:52", "remaining_time": "9:01:14"} +{"current_steps": 295, "total_steps": 2076, "loss": 0.0812, "lr": 1.9895586731938593e-05, "epoch": 0.4264546440187929, "percentage": 14.21, "elapsed_time": "1:29:22", "remaining_time": "8:59:37"} +{"current_steps": 300, "total_steps": 2076, "loss": 0.0825, "lr": 1.9883117108929947e-05, "epoch": 0.43368268883267075, "percentage": 14.45, "elapsed_time": "1:30:56", "remaining_time": "8:58:23"} +{"current_steps": 305, "total_steps": 2076, "loss": 0.0821, "lr": 1.986994864650439e-05, "epoch": 0.4409107336465486, "percentage": 14.69, "elapsed_time": "1:32:31", "remaining_time": "8:57:16"} +{"current_steps": 310, "total_steps": 2076, "loss": 0.0817, "lr": 1.9856082275809508e-05, "epoch": 0.44813877846042643, "percentage": 14.93, "elapsed_time": "1:34:05", "remaining_time": "8:56:01"} +{"current_steps": 315, "total_steps": 2076, "loss": 0.081, "lr": 1.9841518977342274e-05, "epoch": 0.4553668232743043, "percentage": 15.17, "elapsed_time": "1:35:33", "remaining_time": "8:54:14"} +{"current_steps": 320, "total_steps": 2076, "loss": 0.081, "lr": 1.9826259780879716e-05, "epoch": 0.4625948680881821, "percentage": 15.41, "elapsed_time": "1:37:04", "remaining_time": "8:52:39"} +{"current_steps": 325, "total_steps": 2076, "loss": 0.0806, "lr": 1.981030576540612e-05, "epoch": 0.46982291290206, "percentage": 15.66, "elapsed_time": "1:38:38", "remaining_time": "8:51:28"} +{"current_steps": 330, "total_steps": 2076, "loss": 0.0799, "lr": 1.9793658059036697e-05, "epoch": 0.47705095771593786, "percentage": 15.9, "elapsed_time": "1:40:06", "remaining_time": "8:49:41"} +{"current_steps": 335, "total_steps": 2076, "loss": 0.0813, "lr": 1.977631783893786e-05, "epoch": 0.4842790025298157, "percentage": 16.14, "elapsed_time": "1:41:35", "remaining_time": "8:48:00"} +{"current_steps": 340, "total_steps": 2076, "loss": 0.0805, "lr": 1.975828633124394e-05, "epoch": 0.49150704734369355, "percentage": 16.38, "elapsed_time": "1:42:56", "remaining_time": "8:45:38"} +{"current_steps": 345, "total_steps": 2076, "loss": 0.0804, "lr": 1.9739564810970534e-05, "epoch": 0.4987350921575714, "percentage": 16.62, "elapsed_time": "1:44:24", "remaining_time": "8:43:50"} +{"current_steps": 350, "total_steps": 2076, "loss": 0.0802, "lr": 1.9720154601924295e-05, "epoch": 0.5059631369714492, "percentage": 16.86, "elapsed_time": "1:45:56", "remaining_time": "8:42:25"} +{"current_steps": 355, "total_steps": 2076, "loss": 0.08, "lr": 1.9700057076609377e-05, "epoch": 0.5131911817853271, "percentage": 17.1, "elapsed_time": "1:47:24", "remaining_time": "8:40:43"} +{"current_steps": 360, "total_steps": 2076, "loss": 0.0802, "lr": 1.967927365613034e-05, "epoch": 0.5204192265992049, "percentage": 17.34, "elapsed_time": "1:48:59", "remaining_time": "8:39:31"} +{"current_steps": 365, "total_steps": 2076, "loss": 0.0799, "lr": 1.96578058100917e-05, "epoch": 0.5276472714130828, "percentage": 17.58, "elapsed_time": "1:50:30", "remaining_time": "8:38:00"} +{"current_steps": 370, "total_steps": 2076, "loss": 0.081, "lr": 1.963565505649398e-05, "epoch": 0.5348753162269606, "percentage": 17.82, "elapsed_time": "1:52:01", "remaining_time": "8:36:31"} +{"current_steps": 375, "total_steps": 2076, "loss": 0.0799, "lr": 1.961282296162639e-05, "epoch": 0.5421033610408384, "percentage": 18.06, "elapsed_time": "1:53:34", "remaining_time": "8:35:12"} +{"current_steps": 380, "total_steps": 2076, "loss": 0.0855, "lr": 1.9589311139956086e-05, "epoch": 0.5493314058547163, "percentage": 18.3, "elapsed_time": "1:55:08", "remaining_time": "8:33:53"} +{"current_steps": 385, "total_steps": 2076, "loss": 0.0809, "lr": 1.956512125401398e-05, "epoch": 0.5565594506685941, "percentage": 18.55, "elapsed_time": "1:56:44", "remaining_time": "8:32:46"} +{"current_steps": 390, "total_steps": 2076, "loss": 0.0796, "lr": 1.9540255014277198e-05, "epoch": 0.563787495482472, "percentage": 18.79, "elapsed_time": "1:58:11", "remaining_time": "8:30:56"} +{"current_steps": 395, "total_steps": 2076, "loss": 0.0792, "lr": 1.9514714179048138e-05, "epoch": 0.5710155402963498, "percentage": 19.03, "elapsed_time": "1:59:46", "remaining_time": "8:29:45"} +{"current_steps": 400, "total_steps": 2076, "loss": 0.0791, "lr": 1.9488500554330126e-05, "epoch": 0.5782435851102277, "percentage": 19.27, "elapsed_time": "2:01:23", "remaining_time": "8:28:39"} +{"current_steps": 405, "total_steps": 2076, "loss": 0.0794, "lr": 1.946161599369973e-05, "epoch": 0.5854716299241055, "percentage": 19.51, "elapsed_time": "2:03:00", "remaining_time": "8:27:31"} +{"current_steps": 410, "total_steps": 2076, "loss": 0.0776, "lr": 1.9434062398175667e-05, "epoch": 0.5926996747379834, "percentage": 19.75, "elapsed_time": "2:04:26", "remaining_time": "8:25:38"} +{"current_steps": 415, "total_steps": 2076, "loss": 0.0772, "lr": 1.9405841716084403e-05, "epoch": 0.5999277195518612, "percentage": 19.99, "elapsed_time": "2:06:01", "remaining_time": "8:24:25"} +{"current_steps": 420, "total_steps": 2076, "loss": 0.0765, "lr": 1.937695594292238e-05, "epoch": 0.607155764365739, "percentage": 20.23, "elapsed_time": "2:07:36", "remaining_time": "8:23:10"} +{"current_steps": 425, "total_steps": 2076, "loss": 0.075, "lr": 1.9347407121214917e-05, "epoch": 0.6143838091796169, "percentage": 20.47, "elapsed_time": "2:09:02", "remaining_time": "8:21:18"} +{"current_steps": 430, "total_steps": 2076, "loss": 0.0706, "lr": 1.9317197340371764e-05, "epoch": 0.6216118539934947, "percentage": 20.71, "elapsed_time": "2:10:28", "remaining_time": "8:19:25"} +{"current_steps": 435, "total_steps": 2076, "loss": 0.0827, "lr": 1.9286328736539385e-05, "epoch": 0.6288398988073726, "percentage": 20.95, "elapsed_time": "2:12:01", "remaining_time": "8:18:04"} +{"current_steps": 440, "total_steps": 2076, "loss": 0.0815, "lr": 1.9254803492449894e-05, "epoch": 0.6360679436212504, "percentage": 21.19, "elapsed_time": "2:13:31", "remaining_time": "8:16:27"} +{"current_steps": 445, "total_steps": 2076, "loss": 0.0816, "lr": 1.922262383726672e-05, "epoch": 0.6432959884351283, "percentage": 21.44, "elapsed_time": "2:15:02", "remaining_time": "8:14:56"} +{"current_steps": 450, "total_steps": 2076, "loss": 0.0795, "lr": 1.9189792046426972e-05, "epoch": 0.6505240332490061, "percentage": 21.68, "elapsed_time": "2:16:34", "remaining_time": "8:13:29"} +{"current_steps": 455, "total_steps": 2076, "loss": 0.0795, "lr": 1.9156310441480557e-05, "epoch": 0.6577520780628839, "percentage": 21.92, "elapsed_time": "2:18:06", "remaining_time": "8:12:01"} +{"current_steps": 460, "total_steps": 2076, "loss": 0.0772, "lr": 1.912218138992601e-05, "epoch": 0.6649801228767618, "percentage": 22.16, "elapsed_time": "2:19:34", "remaining_time": "8:10:19"} +{"current_steps": 465, "total_steps": 2076, "loss": 0.0732, "lr": 1.9087407305043085e-05, "epoch": 0.6722081676906397, "percentage": 22.4, "elapsed_time": "2:21:12", "remaining_time": "8:09:12"} +{"current_steps": 470, "total_steps": 2076, "loss": 0.0694, "lr": 1.9051990645722133e-05, "epoch": 0.6794362125045176, "percentage": 22.64, "elapsed_time": "2:22:43", "remaining_time": "8:07:42"} +{"current_steps": 475, "total_steps": 2076, "loss": 0.0676, "lr": 1.9015933916290202e-05, "epoch": 0.6866642573183954, "percentage": 22.88, "elapsed_time": "2:24:18", "remaining_time": "8:06:23"} +{"current_steps": 480, "total_steps": 2076, "loss": 0.0635, "lr": 1.8979239666333975e-05, "epoch": 0.6938923021322733, "percentage": 23.12, "elapsed_time": "2:25:43", "remaining_time": "8:04:33"} +{"current_steps": 485, "total_steps": 2076, "loss": 0.0636, "lr": 1.8941910490519483e-05, "epoch": 0.7011203469461511, "percentage": 23.36, "elapsed_time": "2:27:07", "remaining_time": "8:02:38"} +{"current_steps": 490, "total_steps": 2076, "loss": 0.0648, "lr": 1.8903949028408636e-05, "epoch": 0.708348391760029, "percentage": 23.6, "elapsed_time": "2:28:38", "remaining_time": "8:01:06"} +{"current_steps": 495, "total_steps": 2076, "loss": 0.062, "lr": 1.8865357964272576e-05, "epoch": 0.7155764365739068, "percentage": 23.84, "elapsed_time": "2:30:09", "remaining_time": "7:59:35"} +{"current_steps": 500, "total_steps": 2076, "loss": 0.0607, "lr": 1.8826140026901873e-05, "epoch": 0.7228044813877846, "percentage": 24.08, "elapsed_time": "2:31:46", "remaining_time": "7:58:24"} +{"current_steps": 500, "total_steps": 2076, "eval_loss": 0.07328393310308456, "epoch": 0.7228044813877846, "percentage": 24.08, "elapsed_time": "2:50:50", "remaining_time": "8:58:30"} +{"current_steps": 505, "total_steps": 2076, "loss": 0.0631, "lr": 1.878629798941357e-05, "epoch": 0.7300325262016625, "percentage": 24.33, "elapsed_time": "2:52:52", "remaining_time": "8:57:47"} +{"current_steps": 510, "total_steps": 2076, "loss": 0.064, "lr": 1.8745834669055085e-05, "epoch": 0.7372605710155403, "percentage": 24.57, "elapsed_time": "2:54:24", "remaining_time": "8:55:32"} +{"current_steps": 515, "total_steps": 2076, "loss": 0.0618, "lr": 1.8704752927005034e-05, "epoch": 0.7444886158294182, "percentage": 24.81, "elapsed_time": "2:55:51", "remaining_time": "8:53:02"} +{"current_steps": 520, "total_steps": 2076, "loss": 0.0613, "lr": 1.8663055668170873e-05, "epoch": 0.751716660643296, "percentage": 25.05, "elapsed_time": "2:57:27", "remaining_time": "8:50:59"} +{"current_steps": 525, "total_steps": 2076, "loss": 0.0589, "lr": 1.8620745840983522e-05, "epoch": 0.7589447054571739, "percentage": 25.29, "elapsed_time": "2:58:59", "remaining_time": "8:48:48"} +{"current_steps": 530, "total_steps": 2076, "loss": 0.0578, "lr": 1.857782643718887e-05, "epoch": 0.7661727502710517, "percentage": 25.53, "elapsed_time": "3:00:30", "remaining_time": "8:46:31"} +{"current_steps": 535, "total_steps": 2076, "loss": 0.0561, "lr": 1.8534300491636225e-05, "epoch": 0.7734007950849295, "percentage": 25.77, "elapsed_time": "3:01:58", "remaining_time": "8:44:10"} +{"current_steps": 540, "total_steps": 2076, "loss": 0.0615, "lr": 1.849017108206372e-05, "epoch": 0.7806288398988074, "percentage": 26.01, "elapsed_time": "3:03:30", "remaining_time": "8:41:57"} +{"current_steps": 545, "total_steps": 2076, "loss": 0.0613, "lr": 1.844544132888068e-05, "epoch": 0.7878568847126852, "percentage": 26.25, "elapsed_time": "3:05:08", "remaining_time": "8:40:05"} +{"current_steps": 550, "total_steps": 2076, "loss": 0.0546, "lr": 1.8400114394947003e-05, "epoch": 0.7950849295265631, "percentage": 26.49, "elapsed_time": "3:06:38", "remaining_time": "8:37:49"} +{"current_steps": 555, "total_steps": 2076, "loss": 0.0521, "lr": 1.8354193485349468e-05, "epoch": 0.8023129743404409, "percentage": 26.73, "elapsed_time": "3:08:10", "remaining_time": "8:35:40"} +{"current_steps": 560, "total_steps": 2076, "loss": 0.0581, "lr": 1.830768184717514e-05, "epoch": 0.8095410191543188, "percentage": 26.97, "elapsed_time": "3:09:34", "remaining_time": "8:33:13"} +{"current_steps": 565, "total_steps": 2076, "loss": 0.0596, "lr": 1.8260582769281747e-05, "epoch": 0.8167690639681966, "percentage": 27.22, "elapsed_time": "3:11:03", "remaining_time": "8:30:58"} +{"current_steps": 570, "total_steps": 2076, "loss": 0.0588, "lr": 1.821289958206513e-05, "epoch": 0.8239971087820744, "percentage": 27.46, "elapsed_time": "3:12:30", "remaining_time": "8:28:38"} +{"current_steps": 575, "total_steps": 2076, "loss": 0.0584, "lr": 1.8164635657223755e-05, "epoch": 0.8312251535959523, "percentage": 27.7, "elapsed_time": "3:13:57", "remaining_time": "8:26:20"} +{"current_steps": 580, "total_steps": 2076, "loss": 0.0558, "lr": 1.8115794407520287e-05, "epoch": 0.8384531984098301, "percentage": 27.94, "elapsed_time": "3:15:29", "remaining_time": "8:24:13"} +{"current_steps": 585, "total_steps": 2076, "loss": 0.0524, "lr": 1.8066379286540278e-05, "epoch": 0.845681243223708, "percentage": 28.18, "elapsed_time": "3:16:53", "remaining_time": "8:21:50"} +{"current_steps": 590, "total_steps": 2076, "loss": 0.0549, "lr": 1.8016393788447964e-05, "epoch": 0.8529092880375858, "percentage": 28.42, "elapsed_time": "3:18:36", "remaining_time": "8:20:12"} +{"current_steps": 595, "total_steps": 2076, "loss": 0.0491, "lr": 1.7965841447739185e-05, "epoch": 0.8601373328514637, "percentage": 28.66, "elapsed_time": "3:20:14", "remaining_time": "8:18:23"} +{"current_steps": 600, "total_steps": 2076, "loss": 0.0543, "lr": 1.7914725838991472e-05, "epoch": 0.8673653776653415, "percentage": 28.9, "elapsed_time": "3:21:43", "remaining_time": "8:16:14"} +{"current_steps": 605, "total_steps": 2076, "loss": 0.051, "lr": 1.7863050576611267e-05, "epoch": 0.8745934224792193, "percentage": 29.14, "elapsed_time": "3:23:10", "remaining_time": "8:14:01"} +{"current_steps": 610, "total_steps": 2076, "loss": 0.0553, "lr": 1.781081931457837e-05, "epoch": 0.8818214672930972, "percentage": 29.38, "elapsed_time": "3:24:49", "remaining_time": "8:12:14"} +{"current_steps": 615, "total_steps": 2076, "loss": 0.0517, "lr": 1.7758035746187553e-05, "epoch": 0.889049512106975, "percentage": 29.62, "elapsed_time": "3:26:20", "remaining_time": "8:10:11"} +{"current_steps": 620, "total_steps": 2076, "loss": 0.0545, "lr": 1.770470360378739e-05, "epoch": 0.8962775569208529, "percentage": 29.87, "elapsed_time": "3:27:53", "remaining_time": "8:08:12"} +{"current_steps": 625, "total_steps": 2076, "loss": 0.0548, "lr": 1.7650826658516375e-05, "epoch": 0.9035056017347307, "percentage": 30.11, "elapsed_time": "3:29:22", "remaining_time": "8:06:05"} +{"current_steps": 630, "total_steps": 2076, "loss": 0.0551, "lr": 1.7596408720036232e-05, "epoch": 0.9107336465486086, "percentage": 30.35, "elapsed_time": "3:30:51", "remaining_time": "8:03:57"} +{"current_steps": 635, "total_steps": 2076, "loss": 0.05, "lr": 1.754145363626256e-05, "epoch": 0.9179616913624864, "percentage": 30.59, "elapsed_time": "3:32:14", "remaining_time": "8:01:37"} +{"current_steps": 640, "total_steps": 2076, "loss": 0.0503, "lr": 1.748596529309271e-05, "epoch": 0.9251897361763642, "percentage": 30.83, "elapsed_time": "3:33:33", "remaining_time": "7:59:11"} +{"current_steps": 645, "total_steps": 2076, "loss": 0.0488, "lr": 1.742994761413105e-05, "epoch": 0.9324177809902421, "percentage": 31.07, "elapsed_time": "3:35:00", "remaining_time": "7:57:00"} +{"current_steps": 650, "total_steps": 2076, "loss": 0.0553, "lr": 1.73734045604115e-05, "epoch": 0.93964582580412, "percentage": 31.31, "elapsed_time": "3:36:36", "remaining_time": "7:55:12"} +{"current_steps": 655, "total_steps": 2076, "loss": 0.0499, "lr": 1.731634013011745e-05, "epoch": 0.9468738706179979, "percentage": 31.55, "elapsed_time": "3:38:10", "remaining_time": "7:53:18"} +{"current_steps": 660, "total_steps": 2076, "loss": 0.0521, "lr": 1.7258758358299053e-05, "epoch": 0.9541019154318757, "percentage": 31.79, "elapsed_time": "3:39:45", "remaining_time": "7:51:28"} +{"current_steps": 665, "total_steps": 2076, "loss": 0.0505, "lr": 1.7200663316587897e-05, "epoch": 0.9613299602457536, "percentage": 32.03, "elapsed_time": "3:41:21", "remaining_time": "7:49:40"} +{"current_steps": 670, "total_steps": 2076, "loss": 0.0537, "lr": 1.7142059112909107e-05, "epoch": 0.9685580050596314, "percentage": 32.27, "elapsed_time": "3:42:55", "remaining_time": "7:47:49"} +{"current_steps": 675, "total_steps": 2076, "loss": 0.0466, "lr": 1.708294989119087e-05, "epoch": 0.9757860498735093, "percentage": 32.51, "elapsed_time": "3:44:29", "remaining_time": "7:45:57"} +{"current_steps": 680, "total_steps": 2076, "loss": 0.0527, "lr": 1.7023339831071408e-05, "epoch": 0.9830140946873871, "percentage": 32.76, "elapsed_time": "3:46:00", "remaining_time": "7:43:59"} +{"current_steps": 685, "total_steps": 2076, "loss": 0.0483, "lr": 1.696323314760344e-05, "epoch": 0.9902421395012649, "percentage": 33.0, "elapsed_time": "3:47:26", "remaining_time": "7:41:50"} +{"current_steps": 690, "total_steps": 2076, "loss": 0.0483, "lr": 1.690263409095614e-05, "epoch": 0.9974701843151428, "percentage": 33.24, "elapsed_time": "3:48:59", "remaining_time": "7:39:59"} +{"current_steps": 695, "total_steps": 2076, "loss": 0.0417, "lr": 1.6841546946114586e-05, "epoch": 1.0043368268883268, "percentage": 33.48, "elapsed_time": "3:50:31", "remaining_time": "7:38:04"} +{"current_steps": 700, "total_steps": 2076, "loss": 0.0336, "lr": 1.6779976032576792e-05, "epoch": 1.0115648717022045, "percentage": 33.72, "elapsed_time": "3:52:06", "remaining_time": "7:36:15"} +{"current_steps": 705, "total_steps": 2076, "loss": 0.0416, "lr": 1.6717925704048256e-05, "epoch": 1.0187929165160825, "percentage": 33.96, "elapsed_time": "3:53:36", "remaining_time": "7:34:16"} +{"current_steps": 710, "total_steps": 2076, "loss": 0.0404, "lr": 1.6655400348134122e-05, "epoch": 1.0260209613299602, "percentage": 34.2, "elapsed_time": "3:55:12", "remaining_time": "7:32:32"} +{"current_steps": 715, "total_steps": 2076, "loss": 0.0431, "lr": 1.659240438602891e-05, "epoch": 1.0332490061438382, "percentage": 34.44, "elapsed_time": "3:56:56", "remaining_time": "7:31:01"} +{"current_steps": 720, "total_steps": 2076, "loss": 0.0419, "lr": 1.6528942272203912e-05, "epoch": 1.040477050957716, "percentage": 34.68, "elapsed_time": "3:58:29", "remaining_time": "7:29:09"} +{"current_steps": 725, "total_steps": 2076, "loss": 0.0401, "lr": 1.6465018494092213e-05, "epoch": 1.0477050957715939, "percentage": 34.92, "elapsed_time": "3:59:58", "remaining_time": "7:27:11"} +{"current_steps": 730, "total_steps": 2076, "loss": 0.0383, "lr": 1.6400637571771354e-05, "epoch": 1.0549331405854716, "percentage": 35.16, "elapsed_time": "4:01:35", "remaining_time": "7:25:27"} +{"current_steps": 735, "total_steps": 2076, "loss": 0.0355, "lr": 1.633580405764376e-05, "epoch": 1.0621611853993496, "percentage": 35.4, "elapsed_time": "4:03:10", "remaining_time": "7:23:40"} +{"current_steps": 740, "total_steps": 2076, "loss": 0.0393, "lr": 1.6270522536114813e-05, "epoch": 1.0693892302132273, "percentage": 35.65, "elapsed_time": "4:04:42", "remaining_time": "7:21:48"} +{"current_steps": 745, "total_steps": 2076, "loss": 0.0353, "lr": 1.6204797623268675e-05, "epoch": 1.0766172750271052, "percentage": 35.89, "elapsed_time": "4:06:10", "remaining_time": "7:19:47"} +{"current_steps": 750, "total_steps": 2076, "loss": 0.0395, "lr": 1.6138633966541905e-05, "epoch": 1.083845319840983, "percentage": 36.13, "elapsed_time": "4:07:44", "remaining_time": "7:17:59"} +{"current_steps": 755, "total_steps": 2076, "loss": 0.0409, "lr": 1.6072036244394836e-05, "epoch": 1.091073364654861, "percentage": 36.37, "elapsed_time": "4:09:15", "remaining_time": "7:16:07"} +{"current_steps": 760, "total_steps": 2076, "loss": 0.0382, "lr": 1.600500916598074e-05, "epoch": 1.0983014094687387, "percentage": 36.61, "elapsed_time": "4:10:47", "remaining_time": "7:14:16"} +{"current_steps": 765, "total_steps": 2076, "loss": 0.0369, "lr": 1.5937557470812852e-05, "epoch": 1.1055294542826166, "percentage": 36.85, "elapsed_time": "4:12:18", "remaining_time": "7:12:23"} +{"current_steps": 770, "total_steps": 2076, "loss": 0.0393, "lr": 1.5869685928429253e-05, "epoch": 1.1127574990964944, "percentage": 37.09, "elapsed_time": "4:13:50", "remaining_time": "7:10:32"} +{"current_steps": 775, "total_steps": 2076, "loss": 0.0352, "lr": 1.5801399338055584e-05, "epoch": 1.1199855439103723, "percentage": 37.33, "elapsed_time": "4:15:25", "remaining_time": "7:08:46"} +{"current_steps": 780, "total_steps": 2076, "loss": 0.0381, "lr": 1.5732702528265716e-05, "epoch": 1.12721358872425, "percentage": 37.57, "elapsed_time": "4:16:53", "remaining_time": "7:06:49"} +{"current_steps": 785, "total_steps": 2076, "loss": 0.0367, "lr": 1.5663600356640306e-05, "epoch": 1.134441633538128, "percentage": 37.81, "elapsed_time": "4:18:27", "remaining_time": "7:05:04"} +{"current_steps": 790, "total_steps": 2076, "loss": 0.0389, "lr": 1.5594097709423316e-05, "epoch": 1.1416696783520057, "percentage": 38.05, "elapsed_time": "4:20:01", "remaining_time": "7:03:17"} +{"current_steps": 795, "total_steps": 2076, "loss": 0.0381, "lr": 1.552419950117651e-05, "epoch": 1.1488977231658837, "percentage": 38.29, "elapsed_time": "4:21:33", "remaining_time": "7:01:26"} +{"current_steps": 800, "total_steps": 2076, "loss": 0.0343, "lr": 1.545391067443194e-05, "epoch": 1.1561257679797614, "percentage": 38.54, "elapsed_time": "4:22:58", "remaining_time": "6:59:27"} +{"current_steps": 805, "total_steps": 2076, "loss": 0.0389, "lr": 1.538323619934247e-05, "epoch": 1.1633538127936394, "percentage": 38.78, "elapsed_time": "4:24:26", "remaining_time": "6:57:30"} +{"current_steps": 810, "total_steps": 2076, "loss": 0.0366, "lr": 1.5312181073330295e-05, "epoch": 1.170581857607517, "percentage": 39.02, "elapsed_time": "4:26:03", "remaining_time": "6:55:50"} +{"current_steps": 815, "total_steps": 2076, "loss": 0.0373, "lr": 1.524075032073363e-05, "epoch": 1.177809902421395, "percentage": 39.26, "elapsed_time": "4:27:37", "remaining_time": "6:54:04"} +{"current_steps": 820, "total_steps": 2076, "loss": 0.0356, "lr": 1.5168948992451382e-05, "epoch": 1.1850379472352728, "percentage": 39.5, "elapsed_time": "4:29:09", "remaining_time": "6:52:16"} +{"current_steps": 825, "total_steps": 2076, "loss": 0.0365, "lr": 1.5096782165586037e-05, "epoch": 1.1922659920491507, "percentage": 39.74, "elapsed_time": "4:30:39", "remaining_time": "6:50:24"} +{"current_steps": 830, "total_steps": 2076, "loss": 0.0381, "lr": 1.5024254943084629e-05, "epoch": 1.1994940368630285, "percentage": 39.98, "elapsed_time": "4:32:06", "remaining_time": "6:48:29"} +{"current_steps": 835, "total_steps": 2076, "loss": 0.0367, "lr": 1.495137245337794e-05, "epoch": 1.2067220816769064, "percentage": 40.22, "elapsed_time": "4:33:38", "remaining_time": "6:46:41"} +{"current_steps": 840, "total_steps": 2076, "loss": 0.0356, "lr": 1.487813985001782e-05, "epoch": 1.2139501264907842, "percentage": 40.46, "elapsed_time": "4:35:08", "remaining_time": "6:44:50"} +{"current_steps": 845, "total_steps": 2076, "loss": 0.0352, "lr": 1.480456231131283e-05, "epoch": 1.221178171304662, "percentage": 40.7, "elapsed_time": "4:36:48", "remaining_time": "6:43:14"} +{"current_steps": 850, "total_steps": 2076, "loss": 0.0364, "lr": 1.4730645039962044e-05, "epoch": 1.2284062161185398, "percentage": 40.94, "elapsed_time": "4:38:27", "remaining_time": "6:41:38"} +{"current_steps": 855, "total_steps": 2076, "loss": 0.0368, "lr": 1.4656393262687172e-05, "epoch": 1.2356342609324178, "percentage": 41.18, "elapsed_time": "4:40:04", "remaining_time": "6:39:57"} +{"current_steps": 860, "total_steps": 2076, "loss": 0.0338, "lr": 1.4581812229862993e-05, "epoch": 1.2428623057462955, "percentage": 41.43, "elapsed_time": "4:41:35", "remaining_time": "6:38:09"} +{"current_steps": 865, "total_steps": 2076, "loss": 0.0328, "lr": 1.4506907215146075e-05, "epoch": 1.2500903505601735, "percentage": 41.67, "elapsed_time": "4:43:10", "remaining_time": "6:36:26"} +{"current_steps": 870, "total_steps": 2076, "loss": 0.0348, "lr": 1.443168351510189e-05, "epoch": 1.2573183953740514, "percentage": 41.91, "elapsed_time": "4:44:30", "remaining_time": "6:34:23"} +{"current_steps": 875, "total_steps": 2076, "loss": 0.0341, "lr": 1.4356146448830277e-05, "epoch": 1.2645464401879292, "percentage": 42.15, "elapsed_time": "4:46:00", "remaining_time": "6:32:33"} +{"current_steps": 880, "total_steps": 2076, "loss": 0.0379, "lr": 1.4280301357589349e-05, "epoch": 1.271774485001807, "percentage": 42.39, "elapsed_time": "4:47:34", "remaining_time": "6:30:50"} +{"current_steps": 885, "total_steps": 2076, "loss": 0.0339, "lr": 1.4204153604417775e-05, "epoch": 1.2790025298156849, "percentage": 42.63, "elapsed_time": "4:49:11", "remaining_time": "6:29:10"} +{"current_steps": 890, "total_steps": 2076, "loss": 0.0317, "lr": 1.4127708573755599e-05, "epoch": 1.2862305746295628, "percentage": 42.87, "elapsed_time": "4:50:47", "remaining_time": "6:27:30"} +{"current_steps": 895, "total_steps": 2076, "loss": 0.0341, "lr": 1.4050971671063464e-05, "epoch": 1.2934586194434405, "percentage": 43.11, "elapsed_time": "4:52:16", "remaining_time": "6:25:40"} +{"current_steps": 900, "total_steps": 2076, "loss": 0.0388, "lr": 1.3973948322440427e-05, "epoch": 1.3006866642573183, "percentage": 43.35, "elapsed_time": "4:53:49", "remaining_time": "6:23:55"} +{"current_steps": 905, "total_steps": 2076, "loss": 0.0309, "lr": 1.3896643974240245e-05, "epoch": 1.3079147090711962, "percentage": 43.59, "elapsed_time": "4:55:22", "remaining_time": "6:22:11"} +{"current_steps": 910, "total_steps": 2076, "loss": 0.0327, "lr": 1.3819064092686278e-05, "epoch": 1.3151427538850742, "percentage": 43.83, "elapsed_time": "4:56:52", "remaining_time": "6:20:23"} +{"current_steps": 915, "total_steps": 2076, "loss": 0.0345, "lr": 1.3741214163484968e-05, "epoch": 1.322370798698952, "percentage": 44.08, "elapsed_time": "4:58:20", "remaining_time": "6:18:33"} +{"current_steps": 920, "total_steps": 2076, "loss": 0.0336, "lr": 1.3663099691437945e-05, "epoch": 1.3295988435128296, "percentage": 44.32, "elapsed_time": "4:59:44", "remaining_time": "6:16:37"} +{"current_steps": 925, "total_steps": 2076, "loss": 0.038, "lr": 1.3584726200052767e-05, "epoch": 1.3368268883267076, "percentage": 44.56, "elapsed_time": "5:01:17", "remaining_time": "6:14:54"} +{"current_steps": 930, "total_steps": 2076, "loss": 0.0346, "lr": 1.3506099231152366e-05, "epoch": 1.3440549331405856, "percentage": 44.8, "elapsed_time": "5:02:51", "remaining_time": "6:13:12"} +{"current_steps": 935, "total_steps": 2076, "loss": 0.0296, "lr": 1.3427224344483178e-05, "epoch": 1.3512829779544633, "percentage": 45.04, "elapsed_time": "5:04:20", "remaining_time": "6:11:23"} +{"current_steps": 940, "total_steps": 2076, "loss": 0.0309, "lr": 1.3348107117322004e-05, "epoch": 1.3585110227683412, "percentage": 45.28, "elapsed_time": "5:05:47", "remaining_time": "6:09:33"} +{"current_steps": 945, "total_steps": 2076, "loss": 0.028, "lr": 1.3268753144081652e-05, "epoch": 1.365739067582219, "percentage": 45.52, "elapsed_time": "5:07:16", "remaining_time": "6:07:44"} +{"current_steps": 950, "total_steps": 2076, "loss": 0.0331, "lr": 1.3189168035915337e-05, "epoch": 1.372967112396097, "percentage": 45.76, "elapsed_time": "5:08:45", "remaining_time": "6:05:57"} +{"current_steps": 955, "total_steps": 2076, "loss": 0.031, "lr": 1.3109357420319933e-05, "epoch": 1.3801951572099747, "percentage": 46.0, "elapsed_time": "5:10:14", "remaining_time": "6:04:09"} +{"current_steps": 960, "total_steps": 2076, "loss": 0.0338, "lr": 1.3029326940738032e-05, "epoch": 1.3874232020238526, "percentage": 46.24, "elapsed_time": "5:11:45", "remaining_time": "6:02:24"} +{"current_steps": 965, "total_steps": 2076, "loss": 0.0313, "lr": 1.2949082256158904e-05, "epoch": 1.3946512468377303, "percentage": 46.48, "elapsed_time": "5:13:13", "remaining_time": "6:00:37"} +{"current_steps": 970, "total_steps": 2076, "loss": 0.0324, "lr": 1.286862904071835e-05, "epoch": 1.4018792916516083, "percentage": 46.72, "elapsed_time": "5:14:43", "remaining_time": "5:58:50"} +{"current_steps": 975, "total_steps": 2076, "loss": 0.0312, "lr": 1.2787972983297472e-05, "epoch": 1.409107336465486, "percentage": 46.97, "elapsed_time": "5:16:10", "remaining_time": "5:57:01"} +{"current_steps": 980, "total_steps": 2076, "loss": 0.0347, "lr": 1.2707119787120417e-05, "epoch": 1.416335381279364, "percentage": 47.21, "elapsed_time": "5:17:37", "remaining_time": "5:55:13"} +{"current_steps": 985, "total_steps": 2076, "loss": 0.0317, "lr": 1.26260751693511e-05, "epoch": 1.4235634260932417, "percentage": 47.45, "elapsed_time": "5:19:08", "remaining_time": "5:53:28"} +{"current_steps": 990, "total_steps": 2076, "loss": 0.0316, "lr": 1.254484486068893e-05, "epoch": 1.4307914709071197, "percentage": 47.69, "elapsed_time": "5:20:35", "remaining_time": "5:51:41"} +{"current_steps": 995, "total_steps": 2076, "loss": 0.0324, "lr": 1.24634346049636e-05, "epoch": 1.4380195157209974, "percentage": 47.93, "elapsed_time": "5:22:06", "remaining_time": "5:49:56"} +{"current_steps": 1000, "total_steps": 2076, "loss": 0.029, "lr": 1.2381850158728952e-05, "epoch": 1.4452475605348754, "percentage": 48.17, "elapsed_time": "5:23:42", "remaining_time": "5:48:18"} +{"current_steps": 1000, "total_steps": 2076, "eval_loss": 0.08190815895795822, "epoch": 1.4452475605348754, "percentage": 48.17, "elapsed_time": "5:42:44", "remaining_time": "6:08:47"} +{"current_steps": 1005, "total_steps": 2076, "loss": 0.0285, "lr": 1.2300097290855887e-05, "epoch": 1.452475605348753, "percentage": 48.41, "elapsed_time": "5:44:45", "remaining_time": "6:07:24"} +{"current_steps": 1010, "total_steps": 2076, "loss": 0.0316, "lr": 1.2218181782124496e-05, "epoch": 1.459703650162631, "percentage": 48.65, "elapsed_time": "5:46:13", "remaining_time": "6:05:25"} +{"current_steps": 1015, "total_steps": 2076, "loss": 0.0306, "lr": 1.2136109424815258e-05, "epoch": 1.466931694976509, "percentage": 48.89, "elapsed_time": "5:47:41", "remaining_time": "6:03:27"} +{"current_steps": 1020, "total_steps": 2076, "loss": 0.03, "lr": 1.205388602229949e-05, "epoch": 1.4741597397903867, "percentage": 49.13, "elapsed_time": "5:49:17", "remaining_time": "6:01:37"} +{"current_steps": 1025, "total_steps": 2076, "loss": 0.0341, "lr": 1.1971517388628972e-05, "epoch": 1.4813877846042645, "percentage": 49.37, "elapsed_time": "5:50:44", "remaining_time": "5:59:38"} +{"current_steps": 1030, "total_steps": 2076, "loss": 0.0303, "lr": 1.1889009348124857e-05, "epoch": 1.4886158294181424, "percentage": 49.61, "elapsed_time": "5:52:11", "remaining_time": "5:57:39"} +{"current_steps": 1035, "total_steps": 2076, "loss": 0.0276, "lr": 1.180636773496579e-05, "epoch": 1.4958438742320204, "percentage": 49.86, "elapsed_time": "5:53:38", "remaining_time": "5:55:41"} +{"current_steps": 1040, "total_steps": 2076, "loss": 0.0273, "lr": 1.1723598392775415e-05, "epoch": 1.503071919045898, "percentage": 50.1, "elapsed_time": "5:55:07", "remaining_time": "5:53:45"} +{"current_steps": 1045, "total_steps": 2076, "loss": 0.0299, "lr": 1.1640707174209147e-05, "epoch": 1.5102999638597758, "percentage": 50.34, "elapsed_time": "5:56:41", "remaining_time": "5:51:55"} +{"current_steps": 1050, "total_steps": 2076, "loss": 0.0308, "lr": 1.1557699940540321e-05, "epoch": 1.5175280086736538, "percentage": 50.58, "elapsed_time": "5:58:13", "remaining_time": "5:50:02"} +{"current_steps": 1055, "total_steps": 2076, "loss": 0.0278, "lr": 1.1474582561245767e-05, "epoch": 1.5247560534875317, "percentage": 50.82, "elapsed_time": "5:59:46", "remaining_time": "5:48:10"} +{"current_steps": 1060, "total_steps": 2076, "loss": 0.0342, "lr": 1.1391360913590736e-05, "epoch": 1.5319840983014095, "percentage": 51.06, "elapsed_time": "6:01:16", "remaining_time": "5:46:17"} +{"current_steps": 1065, "total_steps": 2076, "loss": 0.03, "lr": 1.1308040882213363e-05, "epoch": 1.5392121431152872, "percentage": 51.3, "elapsed_time": "6:02:48", "remaining_time": "5:44:24"} +{"current_steps": 1070, "total_steps": 2076, "loss": 0.0279, "lr": 1.122462835870852e-05, "epoch": 1.5464401879291652, "percentage": 51.54, "elapsed_time": "6:04:19", "remaining_time": "5:42:31"} +{"current_steps": 1075, "total_steps": 2076, "loss": 0.0263, "lr": 1.1141129241211246e-05, "epoch": 1.553668232743043, "percentage": 51.78, "elapsed_time": "6:05:56", "remaining_time": "5:40:45"} +{"current_steps": 1080, "total_steps": 2076, "loss": 0.0365, "lr": 1.1057549433979675e-05, "epoch": 1.5608962775569208, "percentage": 52.02, "elapsed_time": "6:07:26", "remaining_time": "5:38:51"} +{"current_steps": 1085, "total_steps": 2076, "loss": 0.0286, "lr": 1.0973894846977548e-05, "epoch": 1.5681243223707986, "percentage": 52.26, "elapsed_time": "6:09:00", "remaining_time": "5:37:02"} +{"current_steps": 1090, "total_steps": 2076, "loss": 0.0264, "lr": 1.089017139545631e-05, "epoch": 1.5753523671846765, "percentage": 52.5, "elapsed_time": "6:10:25", "remaining_time": "5:35:04"} +{"current_steps": 1095, "total_steps": 2076, "loss": 0.028, "lr": 1.0806384999536857e-05, "epoch": 1.5825804119985545, "percentage": 52.75, "elapsed_time": "6:11:52", "remaining_time": "5:33:09"} +{"current_steps": 1100, "total_steps": 2076, "loss": 0.0247, "lr": 1.0722541583790898e-05, "epoch": 1.5898084568124322, "percentage": 52.99, "elapsed_time": "6:13:22", "remaining_time": "5:31:16"} +{"current_steps": 1105, "total_steps": 2076, "loss": 0.0305, "lr": 1.0638647076822041e-05, "epoch": 1.59703650162631, "percentage": 53.23, "elapsed_time": "6:14:59", "remaining_time": "5:29:31"} +{"current_steps": 1110, "total_steps": 2076, "loss": 0.0294, "lr": 1.0554707410846585e-05, "epoch": 1.604264546440188, "percentage": 53.47, "elapsed_time": "6:16:30", "remaining_time": "5:27:39"} +{"current_steps": 1115, "total_steps": 2076, "loss": 0.028, "lr": 1.0470728521274028e-05, "epoch": 1.6114925912540659, "percentage": 53.71, "elapsed_time": "6:17:58", "remaining_time": "5:25:46"} +{"current_steps": 1120, "total_steps": 2076, "loss": 0.0278, "lr": 1.0386716346287398e-05, "epoch": 1.6187206360679436, "percentage": 53.95, "elapsed_time": "6:19:29", "remaining_time": "5:23:55"} +{"current_steps": 1125, "total_steps": 2076, "loss": 0.0312, "lr": 1.030267682642334e-05, "epoch": 1.6259486808818213, "percentage": 54.19, "elapsed_time": "6:20:57", "remaining_time": "5:22:02"} +{"current_steps": 1130, "total_steps": 2076, "loss": 0.027, "lr": 1.0218615904152067e-05, "epoch": 1.6331767256956993, "percentage": 54.43, "elapsed_time": "6:22:30", "remaining_time": "5:20:13"} +{"current_steps": 1135, "total_steps": 2076, "loss": 0.0301, "lr": 1.0134539523457172e-05, "epoch": 1.6404047705095772, "percentage": 54.67, "elapsed_time": "6:23:59", "remaining_time": "5:18:21"} +{"current_steps": 1140, "total_steps": 2076, "loss": 0.027, "lr": 1.0050453629415317e-05, "epoch": 1.647632815323455, "percentage": 54.91, "elapsed_time": "6:25:30", "remaining_time": "5:16:31"} +{"current_steps": 1145, "total_steps": 2076, "loss": 0.0307, "lr": 9.966364167775851e-06, "epoch": 1.6548608601373327, "percentage": 55.15, "elapsed_time": "6:27:03", "remaining_time": "5:14:43"} +{"current_steps": 1150, "total_steps": 2076, "loss": 0.0267, "lr": 9.882277084540399e-06, "epoch": 1.6620889049512106, "percentage": 55.39, "elapsed_time": "6:28:31", "remaining_time": "5:12:50"} +{"current_steps": 1155, "total_steps": 2076, "loss": 0.0255, "lr": 9.798198325542399e-06, "epoch": 1.6693169497650886, "percentage": 55.64, "elapsed_time": "6:29:57", "remaining_time": "5:10:56"} +{"current_steps": 1160, "total_steps": 2076, "loss": 0.0277, "lr": 9.714133836026687e-06, "epoch": 1.6765449945789666, "percentage": 55.88, "elapsed_time": "6:31:21", "remaining_time": "5:09:01"} +{"current_steps": 1165, "total_steps": 2076, "loss": 0.0248, "lr": 9.630089560229088e-06, "epoch": 1.6837730393928443, "percentage": 56.12, "elapsed_time": "6:32:58", "remaining_time": "5:07:17"} +{"current_steps": 1170, "total_steps": 2076, "loss": 0.0298, "lr": 9.546071440956115e-06, "epoch": 1.691001084206722, "percentage": 56.36, "elapsed_time": "6:34:34", "remaining_time": "5:05:32"} +{"current_steps": 1175, "total_steps": 2076, "loss": 0.0308, "lr": 9.46208541916474e-06, "epoch": 1.6982291290206, "percentage": 56.6, "elapsed_time": "6:36:07", "remaining_time": "5:03:44"} +{"current_steps": 1180, "total_steps": 2076, "loss": 0.0293, "lr": 9.378137433542305e-06, "epoch": 1.705457173834478, "percentage": 56.84, "elapsed_time": "6:37:40", "remaining_time": "5:01:57"} +{"current_steps": 1185, "total_steps": 2076, "loss": 0.0271, "lr": 9.294233420086604e-06, "epoch": 1.7126852186483557, "percentage": 57.08, "elapsed_time": "6:39:09", "remaining_time": "5:00:07"} +{"current_steps": 1190, "total_steps": 2076, "loss": 0.0293, "lr": 9.210379311686129e-06, "epoch": 1.7199132634622334, "percentage": 57.32, "elapsed_time": "6:40:48", "remaining_time": "4:58:24"} +{"current_steps": 1195, "total_steps": 2076, "loss": 0.0289, "lr": 9.12658103770058e-06, "epoch": 1.7271413082761113, "percentage": 57.56, "elapsed_time": "6:42:20", "remaining_time": "4:56:37"} +{"current_steps": 1200, "total_steps": 2076, "loss": 0.0286, "lr": 9.042844523541572e-06, "epoch": 1.7343693530899893, "percentage": 57.8, "elapsed_time": "6:43:49", "remaining_time": "4:54:47"} +{"current_steps": 1205, "total_steps": 2076, "loss": 0.0275, "lr": 8.95917569025366e-06, "epoch": 1.741597397903867, "percentage": 58.04, "elapsed_time": "6:45:19", "remaining_time": "4:52:58"} +{"current_steps": 1210, "total_steps": 2076, "loss": 0.0239, "lr": 8.875580454095651e-06, "epoch": 1.7488254427177448, "percentage": 58.29, "elapsed_time": "6:46:44", "remaining_time": "4:51:06"} +{"current_steps": 1215, "total_steps": 2076, "loss": 0.0262, "lr": 8.792064726122275e-06, "epoch": 1.7560534875316227, "percentage": 58.53, "elapsed_time": "6:48:14", "remaining_time": "4:49:17"} +{"current_steps": 1220, "total_steps": 2076, "loss": 0.0277, "lr": 8.708634411766195e-06, "epoch": 1.7632815323455007, "percentage": 58.77, "elapsed_time": "6:49:44", "remaining_time": "4:47:29"} +{"current_steps": 1225, "total_steps": 2076, "loss": 0.0235, "lr": 8.625295410420451e-06, "epoch": 1.7705095771593784, "percentage": 59.01, "elapsed_time": "6:51:12", "remaining_time": "4:45:40"} +{"current_steps": 1230, "total_steps": 2076, "loss": 0.0271, "lr": 8.542053615021291e-06, "epoch": 1.7777376219732561, "percentage": 59.25, "elapsed_time": "6:52:40", "remaining_time": "4:43:50"} +{"current_steps": 1235, "total_steps": 2076, "loss": 0.025, "lr": 8.4589149116315e-06, "epoch": 1.784965666787134, "percentage": 59.49, "elapsed_time": "6:54:08", "remaining_time": "4:42:01"} +{"current_steps": 1240, "total_steps": 2076, "loss": 0.0294, "lr": 8.375885179024175e-06, "epoch": 1.792193711601012, "percentage": 59.73, "elapsed_time": "6:55:29", "remaining_time": "4:40:07"} +{"current_steps": 1245, "total_steps": 2076, "loss": 0.0239, "lr": 8.292970288267043e-06, "epoch": 1.7994217564148898, "percentage": 59.97, "elapsed_time": "6:57:00", "remaining_time": "4:38:20"} +{"current_steps": 1250, "total_steps": 2076, "loss": 0.0312, "lr": 8.21017610230732e-06, "epoch": 1.8066498012287675, "percentage": 60.21, "elapsed_time": "6:58:36", "remaining_time": "4:36:37"} +{"current_steps": 1255, "total_steps": 2076, "loss": 0.0288, "lr": 8.12750847555713e-06, "epoch": 1.8138778460426455, "percentage": 60.45, "elapsed_time": "7:00:13", "remaining_time": "4:34:54"} +{"current_steps": 1260, "total_steps": 2076, "loss": 0.0242, "lr": 8.044973253479544e-06, "epoch": 1.8211058908565234, "percentage": 60.69, "elapsed_time": "7:01:42", "remaining_time": "4:33:06"} +{"current_steps": 1265, "total_steps": 2076, "loss": 0.0292, "lr": 7.96257627217524e-06, "epoch": 1.8283339356704011, "percentage": 60.93, "elapsed_time": "7:03:09", "remaining_time": "4:31:17"} +{"current_steps": 1270, "total_steps": 2076, "loss": 0.0239, "lr": 7.880323357969838e-06, "epoch": 1.8355619804842789, "percentage": 61.18, "elapsed_time": "7:04:40", "remaining_time": "4:29:30"} +{"current_steps": 1275, "total_steps": 2076, "loss": 0.0245, "lr": 7.798220327001898e-06, "epoch": 1.8427900252981568, "percentage": 61.42, "elapsed_time": "7:06:10", "remaining_time": "4:27:44"} +{"current_steps": 1280, "total_steps": 2076, "loss": 0.0261, "lr": 7.716272984811688e-06, "epoch": 1.8500180701120348, "percentage": 61.66, "elapsed_time": "7:07:49", "remaining_time": "4:26:03"} +{"current_steps": 1285, "total_steps": 2076, "loss": 0.0259, "lr": 7.634487125930649e-06, "epoch": 1.8572461149259125, "percentage": 61.9, "elapsed_time": "7:09:15", "remaining_time": "4:24:14"} +{"current_steps": 1290, "total_steps": 2076, "loss": 0.0237, "lr": 7.55286853347167e-06, "epoch": 1.8644741597397902, "percentage": 62.14, "elapsed_time": "7:10:42", "remaining_time": "4:22:26"} +{"current_steps": 1295, "total_steps": 2076, "loss": 0.0244, "lr": 7.471422978720162e-06, "epoch": 1.8717022045536682, "percentage": 62.38, "elapsed_time": "7:12:15", "remaining_time": "4:20:41"} +{"current_steps": 1300, "total_steps": 2076, "loss": 0.0245, "lr": 7.3901562207259555e-06, "epoch": 1.8789302493675462, "percentage": 62.62, "elapsed_time": "7:13:46", "remaining_time": "4:18:55"} +{"current_steps": 1305, "total_steps": 2076, "loss": 0.0216, "lr": 7.309074005896103e-06, "epoch": 1.8861582941814239, "percentage": 62.86, "elapsed_time": "7:15:19", "remaining_time": "4:17:11"} +{"current_steps": 1310, "total_steps": 2076, "loss": 0.0278, "lr": 7.228182067588518e-06, "epoch": 1.8933863389953016, "percentage": 63.1, "elapsed_time": "7:16:51", "remaining_time": "4:15:26"} +{"current_steps": 1315, "total_steps": 2076, "loss": 0.0258, "lr": 7.1474861257065866e-06, "epoch": 1.9006143838091796, "percentage": 63.34, "elapsed_time": "7:18:25", "remaining_time": "4:13:43"} +{"current_steps": 1320, "total_steps": 2076, "loss": 0.0227, "lr": 7.066991886294702e-06, "epoch": 1.9078424286230575, "percentage": 63.58, "elapsed_time": "7:19:55", "remaining_time": "4:11:57"} +{"current_steps": 1325, "total_steps": 2076, "loss": 0.0333, "lr": 6.9867050411347955e-06, "epoch": 1.9150704734369353, "percentage": 63.82, "elapsed_time": "7:21:28", "remaining_time": "4:10:13"} +{"current_steps": 1330, "total_steps": 2076, "loss": 0.0235, "lr": 6.906631267343849e-06, "epoch": 1.922298518250813, "percentage": 64.07, "elapsed_time": "7:23:03", "remaining_time": "4:08:30"} +{"current_steps": 1335, "total_steps": 2076, "loss": 0.0247, "lr": 6.826776226972489e-06, "epoch": 1.929526563064691, "percentage": 64.31, "elapsed_time": "7:24:31", "remaining_time": "4:06:44"} +{"current_steps": 1340, "total_steps": 2076, "loss": 0.0281, "lr": 6.747145566604605e-06, "epoch": 1.936754607878569, "percentage": 64.55, "elapsed_time": "7:26:01", "remaining_time": "4:04:59"} +{"current_steps": 1345, "total_steps": 2076, "loss": 0.0242, "lr": 6.667744916958085e-06, "epoch": 1.9439826526924469, "percentage": 64.79, "elapsed_time": "7:27:30", "remaining_time": "4:03:13"} +{"current_steps": 1350, "total_steps": 2076, "loss": 0.0254, "lr": 6.588579892486657e-06, "epoch": 1.9512106975063246, "percentage": 65.03, "elapsed_time": "7:29:05", "remaining_time": "4:01:30"} +{"current_steps": 1355, "total_steps": 2076, "loss": 0.0206, "lr": 6.5096560909828855e-06, "epoch": 1.9584387423202023, "percentage": 65.27, "elapsed_time": "7:30:30", "remaining_time": "3:59:43"} +{"current_steps": 1360, "total_steps": 2076, "loss": 0.0223, "lr": 6.430979093182372e-06, "epoch": 1.9656667871340803, "percentage": 65.51, "elapsed_time": "7:32:06", "remaining_time": "3:58:01"} +{"current_steps": 1365, "total_steps": 2076, "loss": 0.0281, "lr": 6.352554462369112e-06, "epoch": 1.9728948319479582, "percentage": 65.75, "elapsed_time": "7:33:42", "remaining_time": "3:56:19"} +{"current_steps": 1370, "total_steps": 2076, "loss": 0.0234, "lr": 6.274387743982127e-06, "epoch": 1.980122876761836, "percentage": 65.99, "elapsed_time": "7:35:09", "remaining_time": "3:54:33"} +{"current_steps": 1375, "total_steps": 2076, "loss": 0.0218, "lr": 6.196484465223343e-06, "epoch": 1.9873509215757137, "percentage": 66.23, "elapsed_time": "7:36:41", "remaining_time": "3:52:49"} +{"current_steps": 1380, "total_steps": 2076, "loss": 0.0267, "lr": 6.1188501346667536e-06, "epoch": 1.9945789663895916, "percentage": 66.47, "elapsed_time": "7:38:15", "remaining_time": "3:51:07"} +{"current_steps": 1385, "total_steps": 2076, "loss": 0.0196, "lr": 6.04149024186891e-06, "epoch": 2.0014456089627757, "percentage": 66.71, "elapsed_time": "7:39:39", "remaining_time": "3:49:20"} +{"current_steps": 1390, "total_steps": 2076, "loss": 0.0085, "lr": 5.964410256980762e-06, "epoch": 2.0086736537766536, "percentage": 66.96, "elapsed_time": "7:41:09", "remaining_time": "3:47:35"} +{"current_steps": 1395, "total_steps": 2076, "loss": 0.0064, "lr": 5.887615630360836e-06, "epoch": 2.015901698590531, "percentage": 67.2, "elapsed_time": "7:42:44", "remaining_time": "3:45:53"} +{"current_steps": 1400, "total_steps": 2076, "loss": 0.0058, "lr": 5.811111792189873e-06, "epoch": 2.023129743404409, "percentage": 67.44, "elapsed_time": "7:44:14", "remaining_time": "3:44:09"} +{"current_steps": 1405, "total_steps": 2076, "loss": 0.0052, "lr": 5.734904152086829e-06, "epoch": 2.030357788218287, "percentage": 67.68, "elapsed_time": "7:45:48", "remaining_time": "3:42:27"} +{"current_steps": 1410, "total_steps": 2076, "loss": 0.0054, "lr": 5.658998098726361e-06, "epoch": 2.037585833032165, "percentage": 67.92, "elapsed_time": "7:47:15", "remaining_time": "3:40:42"} +{"current_steps": 1415, "total_steps": 2076, "loss": 0.0043, "lr": 5.583398999457812e-06, "epoch": 2.0448138778460425, "percentage": 68.16, "elapsed_time": "7:48:41", "remaining_time": "3:38:56"} +{"current_steps": 1420, "total_steps": 2076, "loss": 0.0041, "lr": 5.508112199925659e-06, "epoch": 2.0520419226599205, "percentage": 68.4, "elapsed_time": "7:50:01", "remaining_time": "3:37:08"} +{"current_steps": 1425, "total_steps": 2076, "loss": 0.0038, "lr": 5.433143023691547e-06, "epoch": 2.0592699674737984, "percentage": 68.64, "elapsed_time": "7:51:33", "remaining_time": "3:35:25"} +{"current_steps": 1430, "total_steps": 2076, "loss": 0.0044, "lr": 5.358496771857831e-06, "epoch": 2.0664980122876764, "percentage": 68.88, "elapsed_time": "7:53:05", "remaining_time": "3:33:43"} +{"current_steps": 1435, "total_steps": 2076, "loss": 0.0052, "lr": 5.284178722692743e-06, "epoch": 2.073726057101554, "percentage": 69.12, "elapsed_time": "7:54:46", "remaining_time": "3:32:04"} +{"current_steps": 1440, "total_steps": 2076, "loss": 0.0037, "lr": 5.2101941312571724e-06, "epoch": 2.080954101915432, "percentage": 69.36, "elapsed_time": "7:56:22", "remaining_time": "3:30:23"} +{"current_steps": 1445, "total_steps": 2076, "loss": 0.0047, "lr": 5.136548229033065e-06, "epoch": 2.08818214672931, "percentage": 69.61, "elapsed_time": "7:58:00", "remaining_time": "3:28:44"} +{"current_steps": 1450, "total_steps": 2076, "loss": 0.0052, "lr": 5.063246223553509e-06, "epoch": 2.0954101915431878, "percentage": 69.85, "elapsed_time": "7:59:26", "remaining_time": "3:26:59"} +{"current_steps": 1455, "total_steps": 2076, "loss": 0.0045, "lr": 4.990293298034505e-06, "epoch": 2.1026382363570653, "percentage": 70.09, "elapsed_time": "8:01:10", "remaining_time": "3:25:22"} +{"current_steps": 1460, "total_steps": 2076, "loss": 0.0041, "lr": 4.917694611008477e-06, "epoch": 2.109866281170943, "percentage": 70.33, "elapsed_time": "8:02:42", "remaining_time": "3:23:39"} +{"current_steps": 1465, "total_steps": 2076, "loss": 0.0051, "lr": 4.845455295959468e-06, "epoch": 2.117094325984821, "percentage": 70.57, "elapsed_time": "8:04:16", "remaining_time": "3:21:58"} +{"current_steps": 1470, "total_steps": 2076, "loss": 0.0065, "lr": 4.773580460960195e-06, "epoch": 2.124322370798699, "percentage": 70.81, "elapsed_time": "8:05:49", "remaining_time": "3:20:16"} +{"current_steps": 1475, "total_steps": 2076, "loss": 0.0052, "lr": 4.702075188310826e-06, "epoch": 2.1315504156125766, "percentage": 71.05, "elapsed_time": "8:07:26", "remaining_time": "3:18:36"} +{"current_steps": 1480, "total_steps": 2076, "loss": 0.0048, "lr": 4.6309445341796286e-06, "epoch": 2.1387784604264546, "percentage": 71.29, "elapsed_time": "8:09:00", "remaining_time": "3:16:55"} +{"current_steps": 1485, "total_steps": 2076, "loss": 0.0035, "lr": 4.5601935282454255e-06, "epoch": 2.1460065052403325, "percentage": 71.53, "elapsed_time": "8:10:26", "remaining_time": "3:15:11"} +{"current_steps": 1490, "total_steps": 2076, "loss": 0.0046, "lr": 4.489827173341957e-06, "epoch": 2.1532345500542105, "percentage": 71.77, "elapsed_time": "8:11:56", "remaining_time": "3:13:28"} +{"current_steps": 1495, "total_steps": 2076, "loss": 0.0066, "lr": 4.419850445104126e-06, "epoch": 2.160462594868088, "percentage": 72.01, "elapsed_time": "8:13:33", "remaining_time": "3:11:48"} +{"current_steps": 1500, "total_steps": 2076, "loss": 0.0058, "lr": 4.350268291616166e-06, "epoch": 2.167690639681966, "percentage": 72.25, "elapsed_time": "8:15:05", "remaining_time": "3:10:06"} +{"current_steps": 1500, "total_steps": 2076, "eval_loss": 0.15236619114875793, "epoch": 2.167690639681966, "percentage": 72.25, "elapsed_time": "8:34:08", "remaining_time": "3:17:25"} +{"current_steps": 1505, "total_steps": 2076, "loss": 0.0047, "lr": 4.281085633061764e-06, "epoch": 2.174918684495844, "percentage": 72.5, "elapsed_time": "8:36:01", "remaining_time": "3:15:46"} +{"current_steps": 1510, "total_steps": 2076, "loss": 0.0078, "lr": 4.212307361376146e-06, "epoch": 2.182146729309722, "percentage": 72.74, "elapsed_time": "8:37:32", "remaining_time": "3:13:59"} +{"current_steps": 1515, "total_steps": 2076, "loss": 0.0043, "lr": 4.1439383399001865e-06, "epoch": 2.1893747741235994, "percentage": 72.98, "elapsed_time": "8:39:07", "remaining_time": "3:12:13"} +{"current_steps": 1520, "total_steps": 2076, "loss": 0.0045, "lr": 4.075983403036479e-06, "epoch": 2.1966028189374773, "percentage": 73.22, "elapsed_time": "8:40:41", "remaining_time": "3:10:27"} +{"current_steps": 1525, "total_steps": 2076, "loss": 0.0073, "lr": 4.0084473559075335e-06, "epoch": 2.2038308637513553, "percentage": 73.46, "elapsed_time": "8:42:14", "remaining_time": "3:08:41"} +{"current_steps": 1530, "total_steps": 2076, "loss": 0.006, "lr": 3.941334974015981e-06, "epoch": 2.2110589085652332, "percentage": 73.7, "elapsed_time": "8:43:44", "remaining_time": "3:06:54"} +{"current_steps": 1535, "total_steps": 2076, "loss": 0.0047, "lr": 3.874651002906915e-06, "epoch": 2.2182869533791107, "percentage": 73.94, "elapsed_time": "8:45:16", "remaining_time": "3:05:07"} +{"current_steps": 1540, "total_steps": 2076, "loss": 0.0049, "lr": 3.8084001578323093e-06, "epoch": 2.2255149981929887, "percentage": 74.18, "elapsed_time": "8:46:47", "remaining_time": "3:03:21"} +{"current_steps": 1545, "total_steps": 2076, "loss": 0.0069, "lr": 3.7425871234176134e-06, "epoch": 2.2327430430068667, "percentage": 74.42, "elapsed_time": "8:48:26", "remaining_time": "3:01:37"} +{"current_steps": 1550, "total_steps": 2076, "loss": 0.0034, "lr": 3.6772165533305024e-06, "epoch": 2.2399710878207446, "percentage": 74.66, "elapsed_time": "8:49:55", "remaining_time": "2:59:49"} +{"current_steps": 1555, "total_steps": 2076, "loss": 0.0045, "lr": 3.6122930699518057e-06, "epoch": 2.2471991326346226, "percentage": 74.9, "elapsed_time": "8:51:18", "remaining_time": "2:58:00"} +{"current_steps": 1560, "total_steps": 2076, "loss": 0.0049, "lr": 3.5478212640486652e-06, "epoch": 2.2544271774485, "percentage": 75.14, "elapsed_time": "8:52:52", "remaining_time": "2:56:15"} +{"current_steps": 1565, "total_steps": 2076, "loss": 0.0034, "lr": 3.483805694449913e-06, "epoch": 2.261655222262378, "percentage": 75.39, "elapsed_time": "8:54:16", "remaining_time": "2:54:26"} +{"current_steps": 1570, "total_steps": 2076, "loss": 0.0044, "lr": 3.420250887723722e-06, "epoch": 2.268883267076256, "percentage": 75.63, "elapsed_time": "8:55:49", "remaining_time": "2:52:41"} +{"current_steps": 1575, "total_steps": 2076, "loss": 0.0037, "lr": 3.357161337857523e-06, "epoch": 2.2761113118901335, "percentage": 75.87, "elapsed_time": "8:57:18", "remaining_time": "2:50:54"} +{"current_steps": 1580, "total_steps": 2076, "loss": 0.0051, "lr": 3.2945415059402363e-06, "epoch": 2.2833393567040114, "percentage": 76.11, "elapsed_time": "8:58:52", "remaining_time": "2:49:09"} +{"current_steps": 1585, "total_steps": 2076, "loss": 0.0036, "lr": 3.232395819846824e-06, "epoch": 2.2905674015178894, "percentage": 76.35, "elapsed_time": "9:00:22", "remaining_time": "2:47:23"} +{"current_steps": 1590, "total_steps": 2076, "loss": 0.003, "lr": 3.170728673925206e-06, "epoch": 2.2977954463317674, "percentage": 76.59, "elapsed_time": "9:01:47", "remaining_time": "2:45:36"} +{"current_steps": 1595, "total_steps": 2076, "loss": 0.0056, "lr": 3.1095444286855112e-06, "epoch": 2.3050234911456453, "percentage": 76.83, "elapsed_time": "9:03:13", "remaining_time": "2:43:49"} +{"current_steps": 1600, "total_steps": 2076, "loss": 0.0047, "lr": 3.04884741049176e-06, "epoch": 2.312251535959523, "percentage": 77.07, "elapsed_time": "9:04:44", "remaining_time": "2:42:03"} +{"current_steps": 1605, "total_steps": 2076, "loss": 0.005, "lr": 2.9886419112559396e-06, "epoch": 2.3194795807734008, "percentage": 77.31, "elapsed_time": "9:06:18", "remaining_time": "2:40:19"} +{"current_steps": 1610, "total_steps": 2076, "loss": 0.0065, "lr": 2.9289321881345257e-06, "epoch": 2.3267076255872787, "percentage": 77.55, "elapsed_time": "9:07:51", "remaining_time": "2:38:34"} +{"current_steps": 1615, "total_steps": 2076, "loss": 0.0029, "lr": 2.86972246322745e-06, "epoch": 2.3339356704011567, "percentage": 77.79, "elapsed_time": "9:09:11", "remaining_time": "2:36:46"} +{"current_steps": 1620, "total_steps": 2076, "loss": 0.0029, "lr": 2.8110169232795615e-06, "epoch": 2.341163715215034, "percentage": 78.03, "elapsed_time": "9:10:37", "remaining_time": "2:34:59"} +{"current_steps": 1625, "total_steps": 2076, "loss": 0.0059, "lr": 2.752819719384573e-06, "epoch": 2.348391760028912, "percentage": 78.28, "elapsed_time": "9:12:01", "remaining_time": "2:33:12"} +{"current_steps": 1630, "total_steps": 2076, "loss": 0.0046, "lr": 2.6951349666915404e-06, "epoch": 2.35561980484279, "percentage": 78.52, "elapsed_time": "9:13:31", "remaining_time": "2:31:27"} +{"current_steps": 1635, "total_steps": 2076, "loss": 0.005, "lr": 2.637966744113877e-06, "epoch": 2.362847849656668, "percentage": 78.76, "elapsed_time": "9:15:05", "remaining_time": "2:29:43"} +{"current_steps": 1640, "total_steps": 2076, "loss": 0.0046, "lr": 2.581319094040927e-06, "epoch": 2.3700758944705456, "percentage": 79.0, "elapsed_time": "9:16:29", "remaining_time": "2:27:56"} +{"current_steps": 1645, "total_steps": 2076, "loss": 0.0036, "lr": 2.5251960220521422e-06, "epoch": 2.3773039392844235, "percentage": 79.24, "elapsed_time": "9:17:53", "remaining_time": "2:26:10"} +{"current_steps": 1650, "total_steps": 2076, "loss": 0.0043, "lr": 2.4696014966338267e-06, "epoch": 2.3845319840983015, "percentage": 79.48, "elapsed_time": "9:19:25", "remaining_time": "2:24:25"} +{"current_steps": 1655, "total_steps": 2076, "loss": 0.0031, "lr": 2.4145394488985307e-06, "epoch": 2.3917600289121794, "percentage": 79.72, "elapsed_time": "9:20:56", "remaining_time": "2:22:41"} +{"current_steps": 1660, "total_steps": 2076, "loss": 0.0039, "lr": 2.360013772307086e-06, "epoch": 2.398988073726057, "percentage": 79.96, "elapsed_time": "9:22:31", "remaining_time": "2:20:58"} +{"current_steps": 1665, "total_steps": 2076, "loss": 0.0041, "lr": 2.3060283223932876e-06, "epoch": 2.406216118539935, "percentage": 80.2, "elapsed_time": "9:24:09", "remaining_time": "2:19:15"} +{"current_steps": 1670, "total_steps": 2076, "loss": 0.0062, "lr": 2.252586916491275e-06, "epoch": 2.413444163353813, "percentage": 80.44, "elapsed_time": "9:25:47", "remaining_time": "2:17:33"} +{"current_steps": 1675, "total_steps": 2076, "loss": 0.0048, "lr": 2.1996933334656044e-06, "epoch": 2.420672208167691, "percentage": 80.68, "elapsed_time": "9:27:16", "remaining_time": "2:15:48"} +{"current_steps": 1680, "total_steps": 2076, "loss": 0.0064, "lr": 2.1473513134440425e-06, "epoch": 2.4279002529815683, "percentage": 80.92, "elapsed_time": "9:29:03", "remaining_time": "2:14:07"} +{"current_steps": 1685, "total_steps": 2076, "loss": 0.005, "lr": 2.0955645575531e-06, "epoch": 2.4351282977954463, "percentage": 81.17, "elapsed_time": "9:30:35", "remaining_time": "2:12:24"} +{"current_steps": 1690, "total_steps": 2076, "loss": 0.0056, "lr": 2.0443367276563277e-06, "epoch": 2.442356342609324, "percentage": 81.41, "elapsed_time": "9:32:17", "remaining_time": "2:10:42"} +{"current_steps": 1695, "total_steps": 2076, "loss": 0.0058, "lr": 1.9936714460953743e-06, "epoch": 2.449584387423202, "percentage": 81.65, "elapsed_time": "9:33:56", "remaining_time": "2:09:00"} +{"current_steps": 1700, "total_steps": 2076, "loss": 0.0038, "lr": 1.9435722954338675e-06, "epoch": 2.4568124322370797, "percentage": 81.89, "elapsed_time": "9:35:20", "remaining_time": "2:07:15"} +{"current_steps": 1705, "total_steps": 2076, "loss": 0.0042, "lr": 1.8940428182040715e-06, "epoch": 2.4640404770509576, "percentage": 82.13, "elapsed_time": "9:36:50", "remaining_time": "2:05:31"} +{"current_steps": 1710, "total_steps": 2076, "loss": 0.0029, "lr": 1.8450865166564003e-06, "epoch": 2.4712685218648356, "percentage": 82.37, "elapsed_time": "9:38:20", "remaining_time": "2:03:47"} +{"current_steps": 1715, "total_steps": 2076, "loss": 0.0041, "lr": 1.7967068525117658e-06, "epoch": 2.4784965666787135, "percentage": 82.61, "elapsed_time": "9:40:06", "remaining_time": "2:02:06"} +{"current_steps": 1720, "total_steps": 2076, "loss": 0.0039, "lr": 1.7489072467168166e-06, "epoch": 2.485724611492591, "percentage": 82.85, "elapsed_time": "9:41:37", "remaining_time": "2:00:22"} +{"current_steps": 1725, "total_steps": 2076, "loss": 0.0062, "lr": 1.7016910792020191e-06, "epoch": 2.492952656306469, "percentage": 83.09, "elapsed_time": "9:43:12", "remaining_time": "1:58:40"} +{"current_steps": 1730, "total_steps": 2076, "loss": 0.0033, "lr": 1.6550616886426718e-06, "epoch": 2.500180701120347, "percentage": 83.33, "elapsed_time": "9:44:46", "remaining_time": "1:56:57"} +{"current_steps": 1735, "total_steps": 2076, "loss": 0.005, "lr": 1.609022372222827e-06, "epoch": 2.507408745934225, "percentage": 83.57, "elapsed_time": "9:46:16", "remaining_time": "1:55:13"} +{"current_steps": 1740, "total_steps": 2076, "loss": 0.004, "lr": 1.5635763854021424e-06, "epoch": 2.514636790748103, "percentage": 83.82, "elapsed_time": "9:47:52", "remaining_time": "1:53:31"} +{"current_steps": 1745, "total_steps": 2076, "loss": 0.0044, "lr": 1.5187269416856875e-06, "epoch": 2.5218648355619804, "percentage": 84.06, "elapsed_time": "9:49:28", "remaining_time": "1:51:48"} +{"current_steps": 1750, "total_steps": 2076, "loss": 0.003, "lr": 1.474477212396712e-06, "epoch": 2.5290928803758583, "percentage": 84.3, "elapsed_time": "9:51:06", "remaining_time": "1:50:06"} +{"current_steps": 1755, "total_steps": 2076, "loss": 0.0046, "lr": 1.4308303264524115e-06, "epoch": 2.5363209251897363, "percentage": 84.54, "elapsed_time": "9:52:38", "remaining_time": "1:48:23"} +{"current_steps": 1760, "total_steps": 2076, "loss": 0.0037, "lr": 1.3877893701426637e-06, "epoch": 2.543548970003614, "percentage": 84.78, "elapsed_time": "9:54:16", "remaining_time": "1:46:41"} +{"current_steps": 1765, "total_steps": 2076, "loss": 0.0056, "lr": 1.3453573869118097e-06, "epoch": 2.5507770148174918, "percentage": 85.02, "elapsed_time": "9:55:39", "remaining_time": "1:44:57"} +{"current_steps": 1770, "total_steps": 2076, "loss": 0.0063, "lr": 1.3035373771434356e-06, "epoch": 2.5580050596313697, "percentage": 85.26, "elapsed_time": "9:57:13", "remaining_time": "1:43:14"} +{"current_steps": 1775, "total_steps": 2076, "loss": 0.0049, "lr": 1.2623322979482355e-06, "epoch": 2.5652331044452477, "percentage": 85.5, "elapsed_time": "9:58:49", "remaining_time": "1:41:32"} +{"current_steps": 1780, "total_steps": 2076, "loss": 0.0042, "lr": 1.2217450629548955e-06, "epoch": 2.5724611492591256, "percentage": 85.74, "elapsed_time": "10:00:14", "remaining_time": "1:39:48"} +{"current_steps": 1785, "total_steps": 2076, "loss": 0.0025, "lr": 1.181778542104075e-06, "epoch": 2.579689194073003, "percentage": 85.98, "elapsed_time": "10:01:53", "remaining_time": "1:38:07"} +{"current_steps": 1790, "total_steps": 2076, "loss": 0.0047, "lr": 1.1424355614454718e-06, "epoch": 2.586917238886881, "percentage": 86.22, "elapsed_time": "10:03:30", "remaining_time": "1:36:25"} +{"current_steps": 1795, "total_steps": 2076, "loss": 0.0078, "lr": 1.1037189029379925e-06, "epoch": 2.594145283700759, "percentage": 86.46, "elapsed_time": "10:05:11", "remaining_time": "1:34:44"} +{"current_steps": 1800, "total_steps": 2076, "loss": 0.0051, "lr": 1.0656313042530376e-06, "epoch": 2.6013733285146365, "percentage": 86.71, "elapsed_time": "10:06:46", "remaining_time": "1:33:02"} +{"current_steps": 1805, "total_steps": 2076, "loss": 0.0057, "lr": 1.028175458580918e-06, "epoch": 2.6086013733285145, "percentage": 86.95, "elapsed_time": "10:08:19", "remaining_time": "1:31:20"} +{"current_steps": 1810, "total_steps": 2076, "loss": 0.0029, "lr": 9.913540144404254e-07, "epoch": 2.6158294181423924, "percentage": 87.19, "elapsed_time": "10:09:52", "remaining_time": "1:29:37"} +{"current_steps": 1815, "total_steps": 2076, "loss": 0.0058, "lr": 9.551695754915447e-07, "epoch": 2.6230574629562704, "percentage": 87.43, "elapsed_time": "10:11:27", "remaining_time": "1:27:55"} +{"current_steps": 1820, "total_steps": 2076, "loss": 0.0056, "lr": 9.196247003513537e-07, "epoch": 2.6302855077701484, "percentage": 87.67, "elapsed_time": "10:13:01", "remaining_time": "1:26:13"} +{"current_steps": 1825, "total_steps": 2076, "loss": 0.0042, "lr": 8.84721902413097e-07, "epoch": 2.637513552584026, "percentage": 87.91, "elapsed_time": "10:14:36", "remaining_time": "1:24:31"} +{"current_steps": 1830, "total_steps": 2076, "loss": 0.0043, "lr": 8.50463649668477e-07, "epoch": 2.644741597397904, "percentage": 88.15, "elapsed_time": "10:16:01", "remaining_time": "1:22:48"} +{"current_steps": 1835, "total_steps": 2076, "loss": 0.0047, "lr": 8.168523645331216e-07, "epoch": 2.6519696422117818, "percentage": 88.39, "elapsed_time": "10:17:28", "remaining_time": "1:21:05"} +{"current_steps": 1840, "total_steps": 2076, "loss": 0.0041, "lr": 7.838904236753087e-07, "epoch": 2.6591976870256593, "percentage": 88.63, "elapsed_time": "10:19:04", "remaining_time": "1:19:24"} +{"current_steps": 1845, "total_steps": 2076, "loss": 0.0065, "lr": 7.515801578479032e-07, "epoch": 2.6664257318395372, "percentage": 88.87, "elapsed_time": "10:20:37", "remaining_time": "1:17:42"} +{"current_steps": 1850, "total_steps": 2076, "loss": 0.003, "lr": 7.199238517235541e-07, "epoch": 2.673653776653415, "percentage": 89.11, "elapsed_time": "10:22:09", "remaining_time": "1:16:00"} +{"current_steps": 1855, "total_steps": 2076, "loss": 0.0036, "lr": 6.889237437331398e-07, "epoch": 2.680881821467293, "percentage": 89.35, "elapsed_time": "10:23:43", "remaining_time": "1:14:18"} +{"current_steps": 1860, "total_steps": 2076, "loss": 0.005, "lr": 6.585820259074882e-07, "epoch": 2.688109866281171, "percentage": 89.6, "elapsed_time": "10:25:30", "remaining_time": "1:12:38"} +{"current_steps": 1865, "total_steps": 2076, "loss": 0.0061, "lr": 6.289008437223798e-07, "epoch": 2.6953379110950486, "percentage": 89.84, "elapsed_time": "10:27:03", "remaining_time": "1:10:56"} +{"current_steps": 1870, "total_steps": 2076, "loss": 0.0044, "lr": 5.998822959468409e-07, "epoch": 2.7025659559089266, "percentage": 90.08, "elapsed_time": "10:28:38", "remaining_time": "1:09:15"} +{"current_steps": 1875, "total_steps": 2076, "loss": 0.0052, "lr": 5.715284344947358e-07, "epoch": 2.7097940007228045, "percentage": 90.32, "elapsed_time": "10:30:14", "remaining_time": "1:07:33"} +{"current_steps": 1880, "total_steps": 2076, "loss": 0.004, "lr": 5.438412642796686e-07, "epoch": 2.7170220455366825, "percentage": 90.56, "elapsed_time": "10:31:55", "remaining_time": "1:05:52"} +{"current_steps": 1885, "total_steps": 2076, "loss": 0.0046, "lr": 5.168227430732353e-07, "epoch": 2.7242500903505604, "percentage": 90.8, "elapsed_time": "10:33:31", "remaining_time": "1:04:11"} +{"current_steps": 1890, "total_steps": 2076, "loss": 0.005, "lr": 4.904747813665656e-07, "epoch": 2.731478135164438, "percentage": 91.04, "elapsed_time": "10:35:05", "remaining_time": "1:02:30"} +{"current_steps": 1895, "total_steps": 2076, "loss": 0.0056, "lr": 4.6479924223524655e-07, "epoch": 2.738706179978316, "percentage": 91.28, "elapsed_time": "10:36:39", "remaining_time": "1:00:48"} +{"current_steps": 1900, "total_steps": 2076, "loss": 0.004, "lr": 4.39797941207577e-07, "epoch": 2.745934224792194, "percentage": 91.52, "elapsed_time": "10:38:10", "remaining_time": "0:59:06"} +{"current_steps": 1905, "total_steps": 2076, "loss": 0.0052, "lr": 4.1547264613619243e-07, "epoch": 2.7531622696060714, "percentage": 91.76, "elapsed_time": "10:39:46", "remaining_time": "0:57:25"} +{"current_steps": 1910, "total_steps": 2076, "loss": 0.0055, "lr": 3.9182507707305915e-07, "epoch": 2.7603903144199493, "percentage": 92.0, "elapsed_time": "10:41:17", "remaining_time": "0:55:44"} +{"current_steps": 1915, "total_steps": 2076, "loss": 0.0043, "lr": 3.6885690614785197e-07, "epoch": 2.7676183592338273, "percentage": 92.24, "elapsed_time": "10:42:58", "remaining_time": "0:54:03"} +{"current_steps": 1920, "total_steps": 2076, "loss": 0.0046, "lr": 3.4656975744970846e-07, "epoch": 2.774846404047705, "percentage": 92.49, "elapsed_time": "10:44:29", "remaining_time": "0:52:21"} +{"current_steps": 1925, "total_steps": 2076, "loss": 0.0028, "lr": 3.249652069124032e-07, "epoch": 2.782074448861583, "percentage": 92.73, "elapsed_time": "10:45:57", "remaining_time": "0:50:40"} +{"current_steps": 1930, "total_steps": 2076, "loss": 0.0048, "lr": 3.040447822028958e-07, "epoch": 2.7893024936754607, "percentage": 92.97, "elapsed_time": "10:47:29", "remaining_time": "0:48:58"} +{"current_steps": 1935, "total_steps": 2076, "loss": 0.004, "lr": 2.838099626133206e-07, "epoch": 2.7965305384893386, "percentage": 93.21, "elapsed_time": "10:49:01", "remaining_time": "0:47:17"} +{"current_steps": 1940, "total_steps": 2076, "loss": 0.0069, "lr": 2.642621789563848e-07, "epoch": 2.8037585833032166, "percentage": 93.45, "elapsed_time": "10:50:38", "remaining_time": "0:45:36"} +{"current_steps": 1945, "total_steps": 2076, "loss": 0.0037, "lr": 2.4540281346418946e-07, "epoch": 2.810986628117094, "percentage": 93.69, "elapsed_time": "10:52:12", "remaining_time": "0:43:55"} +{"current_steps": 1950, "total_steps": 2076, "loss": 0.0036, "lr": 2.2723319969049307e-07, "epoch": 2.818214672930972, "percentage": 93.93, "elapsed_time": "10:53:49", "remaining_time": "0:42:14"} +{"current_steps": 1955, "total_steps": 2076, "loss": 0.0024, "lr": 2.0975462241642042e-07, "epoch": 2.82544271774485, "percentage": 94.17, "elapsed_time": "10:55:16", "remaining_time": "0:40:33"} +{"current_steps": 1960, "total_steps": 2076, "loss": 0.0052, "lr": 1.9296831755960753e-07, "epoch": 2.832670762558728, "percentage": 94.41, "elapsed_time": "10:56:49", "remaining_time": "0:38:52"} +{"current_steps": 1965, "total_steps": 2076, "loss": 0.0037, "lr": 1.76875472086816e-07, "epoch": 2.839898807372606, "percentage": 94.65, "elapsed_time": "10:58:28", "remaining_time": "0:37:11"} +{"current_steps": 1970, "total_steps": 2076, "loss": 0.0038, "lr": 1.6147722392999887e-07, "epoch": 2.8471268521864834, "percentage": 94.89, "elapsed_time": "11:00:00", "remaining_time": "0:35:30"} +{"current_steps": 1975, "total_steps": 2076, "loss": 0.0049, "lr": 1.467746619058341e-07, "epoch": 2.8543548970003614, "percentage": 95.13, "elapsed_time": "11:01:31", "remaining_time": "0:33:49"} +{"current_steps": 1980, "total_steps": 2076, "loss": 0.0042, "lr": 1.327688256387416e-07, "epoch": 2.8615829418142393, "percentage": 95.38, "elapsed_time": "11:03:04", "remaining_time": "0:32:08"} +{"current_steps": 1985, "total_steps": 2076, "loss": 0.0035, "lr": 1.1946070548736532e-07, "epoch": 2.868810986628117, "percentage": 95.62, "elapsed_time": "11:04:33", "remaining_time": "0:30:27"} +{"current_steps": 1990, "total_steps": 2076, "loss": 0.0038, "lr": 1.0685124247454159e-07, "epoch": 2.876039031441995, "percentage": 95.86, "elapsed_time": "11:06:01", "remaining_time": "0:28:46"} +{"current_steps": 1995, "total_steps": 2076, "loss": 0.0043, "lr": 9.494132822077007e-08, "epoch": 2.8832670762558728, "percentage": 96.1, "elapsed_time": "11:07:38", "remaining_time": "0:27:06"} +{"current_steps": 2000, "total_steps": 2076, "loss": 0.005, "lr": 8.373180488115529e-08, "epoch": 2.8904951210697507, "percentage": 96.34, "elapsed_time": "11:09:15", "remaining_time": "0:25:25"} +{"current_steps": 2000, "total_steps": 2076, "eval_loss": 0.1443248987197876, "epoch": 2.8904951210697507, "percentage": 96.34, "elapsed_time": "11:29:15", "remaining_time": "0:26:11"} +{"current_steps": 2005, "total_steps": 2076, "loss": 0.0043, "lr": 7.322346508586209e-08, "epoch": 2.8977231658836287, "percentage": 96.58, "elapsed_time": "11:31:13", "remaining_time": "0:24:28"} +{"current_steps": 2010, "total_steps": 2076, "loss": 0.0048, "lr": 6.341705188407043e-08, "epoch": 2.904951210697506, "percentage": 96.82, "elapsed_time": "11:32:43", "remaining_time": "0:22:44"} +{"current_steps": 2015, "total_steps": 2076, "loss": 0.0054, "lr": 5.431325869143189e-08, "epoch": 2.912179255511384, "percentage": 97.06, "elapsed_time": "11:34:18", "remaining_time": "0:21:01"} +{"current_steps": 2020, "total_steps": 2076, "loss": 0.0038, "lr": 4.5912729241036624e-08, "epoch": 2.919407300325262, "percentage": 97.3, "elapsed_time": "11:35:59", "remaining_time": "0:19:17"} +{"current_steps": 2025, "total_steps": 2076, "loss": 0.0038, "lr": 3.821605753789648e-08, "epoch": 2.9266353451391396, "percentage": 97.54, "elapsed_time": "11:37:28", "remaining_time": "0:17:33"} +{"current_steps": 2030, "total_steps": 2076, "loss": 0.0028, "lr": 3.122378781694524e-08, "epoch": 2.933863389953018, "percentage": 97.78, "elapsed_time": "11:38:49", "remaining_time": "0:15:50"} +{"current_steps": 2035, "total_steps": 2076, "loss": 0.0041, "lr": 2.493641450454942e-08, "epoch": 2.9410914347668955, "percentage": 98.03, "elapsed_time": "11:40:23", "remaining_time": "0:14:06"} +{"current_steps": 2040, "total_steps": 2076, "loss": 0.0037, "lr": 1.93543821835529e-08, "epoch": 2.9483194795807735, "percentage": 98.27, "elapsed_time": "11:42:03", "remaining_time": "0:12:23"} +{"current_steps": 2045, "total_steps": 2076, "loss": 0.0041, "lr": 1.4478085561835387e-08, "epoch": 2.9555475243946514, "percentage": 98.51, "elapsed_time": "11:43:35", "remaining_time": "0:10:39"} +{"current_steps": 2050, "total_steps": 2076, "loss": 0.0043, "lr": 1.0307869444406981e-08, "epoch": 2.962775569208529, "percentage": 98.75, "elapsed_time": "11:45:14", "remaining_time": "0:08:56"} +{"current_steps": 2055, "total_steps": 2076, "loss": 0.0053, "lr": 6.844028709024342e-09, "epoch": 2.970003614022407, "percentage": 98.99, "elapsed_time": "11:46:47", "remaining_time": "0:07:13"} +{"current_steps": 2060, "total_steps": 2076, "loss": 0.0043, "lr": 4.086808285338472e-09, "epoch": 2.977231658836285, "percentage": 99.23, "elapsed_time": "11:48:24", "remaining_time": "0:05:30"} +{"current_steps": 2065, "total_steps": 2076, "loss": 0.0034, "lr": 2.0364031375819104e-09, "epoch": 2.9844597036501628, "percentage": 99.47, "elapsed_time": "11:49:46", "remaining_time": "0:03:46"} +{"current_steps": 2070, "total_steps": 2076, "loss": 0.0055, "lr": 6.929582507719801e-10, "epoch": 2.9916877484640407, "percentage": 99.71, "elapsed_time": "11:51:16", "remaining_time": "0:02:03"} +{"current_steps": 2075, "total_steps": 2076, "loss": 0.0047, "lr": 5.6568620471209035e-11, "epoch": 2.9989157932779182, "percentage": 99.95, "elapsed_time": "11:52:47", "remaining_time": "0:00:20"} +{"current_steps": 2076, "total_steps": 2076, "epoch": 3.0, "percentage": 100.0, "elapsed_time": "11:53:29", "remaining_time": "0:00:00"} diff --git a/trainer_state.json b/trainer_state.json new file mode 100644 index 0000000..f7a266a --- /dev/null +++ b/trainer_state.json @@ -0,0 +1,2980 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 3.0, + "eval_steps": 500, + "global_step": 2076, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.007228044813877846, + "grad_norm": 2.43762469291687, + "learning_rate": 3.846153846153847e-07, + "loss": 0.2014, + "step": 5 + }, + { + "epoch": 0.014456089627755691, + "grad_norm": 0.7071607112884521, + "learning_rate": 8.653846153846154e-07, + "loss": 0.1441, + "step": 10 + }, + { + "epoch": 0.02168413444163354, + "grad_norm": 0.25739526748657227, + "learning_rate": 1.3461538461538462e-06, + "loss": 0.0935, + "step": 15 + }, + { + "epoch": 0.028912179255511383, + "grad_norm": 0.1548013836145401, + "learning_rate": 1.826923076923077e-06, + "loss": 0.0841, + "step": 20 + }, + { + "epoch": 0.03614022406938923, + "grad_norm": 0.1367483139038086, + "learning_rate": 2.307692307692308e-06, + "loss": 0.0844, + "step": 25 + }, + { + "epoch": 0.04336826888326708, + "grad_norm": 0.012530342675745487, + "learning_rate": 2.7884615384615386e-06, + "loss": 0.0809, + "step": 30 + }, + { + "epoch": 0.05059631369714492, + "grad_norm": 0.049705736339092255, + "learning_rate": 3.2692307692307696e-06, + "loss": 0.0819, + "step": 35 + }, + { + "epoch": 0.057824358511022765, + "grad_norm": 0.051718585193157196, + "learning_rate": 3.7500000000000005e-06, + "loss": 0.0803, + "step": 40 + }, + { + "epoch": 0.06505240332490062, + "grad_norm": 0.01605582982301712, + "learning_rate": 4.230769230769231e-06, + "loss": 0.0813, + "step": 45 + }, + { + "epoch": 0.07228044813877846, + "grad_norm": 0.06171448528766632, + "learning_rate": 4.711538461538462e-06, + "loss": 0.0801, + "step": 50 + }, + { + "epoch": 0.0795084929526563, + "grad_norm": 0.05586954951286316, + "learning_rate": 5.192307692307693e-06, + "loss": 0.0811, + "step": 55 + }, + { + "epoch": 0.08673653776653416, + "grad_norm": 0.028248045593500137, + "learning_rate": 5.6730769230769235e-06, + "loss": 0.0806, + "step": 60 + }, + { + "epoch": 0.093964582580412, + "grad_norm": 0.04312776029109955, + "learning_rate": 6.153846153846155e-06, + "loss": 0.0808, + "step": 65 + }, + { + "epoch": 0.10119262739428984, + "grad_norm": 0.10684467852115631, + "learning_rate": 6.6346153846153846e-06, + "loss": 0.0808, + "step": 70 + }, + { + "epoch": 0.10842067220816769, + "grad_norm": 0.07184753566980362, + "learning_rate": 7.115384615384616e-06, + "loss": 0.0805, + "step": 75 + }, + { + "epoch": 0.11564871702204553, + "grad_norm": 0.09277470409870148, + "learning_rate": 7.5961538461538465e-06, + "loss": 0.0807, + "step": 80 + }, + { + "epoch": 0.12287676183592339, + "grad_norm": 0.02109931781888008, + "learning_rate": 8.076923076923077e-06, + "loss": 0.0809, + "step": 85 + }, + { + "epoch": 0.13010480664980123, + "grad_norm": 0.03073902055621147, + "learning_rate": 8.557692307692308e-06, + "loss": 0.0812, + "step": 90 + }, + { + "epoch": 0.13733285146367907, + "grad_norm": 0.06802671402692795, + "learning_rate": 9.03846153846154e-06, + "loss": 0.08, + "step": 95 + }, + { + "epoch": 0.14456089627755692, + "grad_norm": 0.03209488093852997, + "learning_rate": 9.51923076923077e-06, + "loss": 0.08, + "step": 100 + }, + { + "epoch": 0.15178894109143476, + "grad_norm": 0.07769843190908432, + "learning_rate": 1e-05, + "loss": 0.0817, + "step": 105 + }, + { + "epoch": 0.1590169859053126, + "grad_norm": 0.06089721992611885, + "learning_rate": 1.0480769230769232e-05, + "loss": 0.0805, + "step": 110 + }, + { + "epoch": 0.16624503071919045, + "grad_norm": 0.021150365471839905, + "learning_rate": 1.0961538461538464e-05, + "loss": 0.0797, + "step": 115 + }, + { + "epoch": 0.17347307553306832, + "grad_norm": 0.035523343831300735, + "learning_rate": 1.1442307692307693e-05, + "loss": 0.0808, + "step": 120 + }, + { + "epoch": 0.18070112034694616, + "grad_norm": 0.019112691283226013, + "learning_rate": 1.1923076923076925e-05, + "loss": 0.0805, + "step": 125 + }, + { + "epoch": 0.187929165160824, + "grad_norm": 0.23690395057201385, + "learning_rate": 1.2403846153846156e-05, + "loss": 0.0836, + "step": 130 + }, + { + "epoch": 0.19515720997470185, + "grad_norm": 0.08766212314367294, + "learning_rate": 1.2884615384615386e-05, + "loss": 0.0802, + "step": 135 + }, + { + "epoch": 0.2023852547885797, + "grad_norm": 0.04037011042237282, + "learning_rate": 1.3365384615384615e-05, + "loss": 0.0811, + "step": 140 + }, + { + "epoch": 0.20961329960245753, + "grad_norm": 0.15161241590976715, + "learning_rate": 1.3846153846153847e-05, + "loss": 0.0829, + "step": 145 + }, + { + "epoch": 0.21684134441633537, + "grad_norm": 0.09708157926797867, + "learning_rate": 1.4326923076923078e-05, + "loss": 0.0816, + "step": 150 + }, + { + "epoch": 0.22406938923021322, + "grad_norm": 0.09547246992588043, + "learning_rate": 1.480769230769231e-05, + "loss": 0.0801, + "step": 155 + }, + { + "epoch": 0.23129743404409106, + "grad_norm": 0.015538723208010197, + "learning_rate": 1.528846153846154e-05, + "loss": 0.0813, + "step": 160 + }, + { + "epoch": 0.23852547885796893, + "grad_norm": 0.053264446556568146, + "learning_rate": 1.576923076923077e-05, + "loss": 0.0811, + "step": 165 + }, + { + "epoch": 0.24575352367184677, + "grad_norm": 0.0034629153087735176, + "learning_rate": 1.6250000000000002e-05, + "loss": 0.081, + "step": 170 + }, + { + "epoch": 0.2529815684857246, + "grad_norm": 0.06591220200061798, + "learning_rate": 1.673076923076923e-05, + "loss": 0.0793, + "step": 175 + }, + { + "epoch": 0.26020961329960246, + "grad_norm": 0.02226085402071476, + "learning_rate": 1.7211538461538465e-05, + "loss": 0.0811, + "step": 180 + }, + { + "epoch": 0.2674376581134803, + "grad_norm": 0.031285785138607025, + "learning_rate": 1.7692307692307694e-05, + "loss": 0.0819, + "step": 185 + }, + { + "epoch": 0.27466570292735815, + "grad_norm": 0.014217695221304893, + "learning_rate": 1.8173076923076924e-05, + "loss": 0.0797, + "step": 190 + }, + { + "epoch": 0.281893747741236, + "grad_norm": 0.061094146221876144, + "learning_rate": 1.8653846153846157e-05, + "loss": 0.0792, + "step": 195 + }, + { + "epoch": 0.28912179255511383, + "grad_norm": 0.1535295695066452, + "learning_rate": 1.9134615384615387e-05, + "loss": 0.0948, + "step": 200 + }, + { + "epoch": 0.2963498373689917, + "grad_norm": 0.15612953901290894, + "learning_rate": 1.9615384615384617e-05, + "loss": 0.0895, + "step": 205 + }, + { + "epoch": 0.3035778821828695, + "grad_norm": 0.015299557708203793, + "learning_rate": 1.999998585783488e-05, + "loss": 0.0818, + "step": 210 + }, + { + "epoch": 0.31080592699674736, + "grad_norm": 0.07040851563215256, + "learning_rate": 1.9999490886255767e-05, + "loss": 0.0812, + "step": 215 + }, + { + "epoch": 0.3180339718106252, + "grad_norm": 0.008683345280587673, + "learning_rate": 1.999828884642042e-05, + "loss": 0.0814, + "step": 220 + }, + { + "epoch": 0.32526201662450305, + "grad_norm": 0.10729002952575684, + "learning_rate": 1.9996379823325586e-05, + "loss": 0.0806, + "step": 225 + }, + { + "epoch": 0.3324900614383809, + "grad_norm": 0.07564987987279892, + "learning_rate": 1.9993763951959107e-05, + "loss": 0.0815, + "step": 230 + }, + { + "epoch": 0.3397181062522588, + "grad_norm": 0.10431886464357376, + "learning_rate": 1.9990441417290358e-05, + "loss": 0.0846, + "step": 235 + }, + { + "epoch": 0.34694615106613663, + "grad_norm": 0.11196550726890564, + "learning_rate": 1.9986412454257178e-05, + "loss": 0.0849, + "step": 240 + }, + { + "epoch": 0.3541741958800145, + "grad_norm": 0.04622909799218178, + "learning_rate": 1.998167734774926e-05, + "loss": 0.0821, + "step": 245 + }, + { + "epoch": 0.3614022406938923, + "grad_norm": 0.06048440560698509, + "learning_rate": 1.9976236432588002e-05, + "loss": 0.0808, + "step": 250 + }, + { + "epoch": 0.36863028550777016, + "grad_norm": 0.01802447997033596, + "learning_rate": 1.997009009350283e-05, + "loss": 0.0805, + "step": 255 + }, + { + "epoch": 0.375858330321648, + "grad_norm": 0.04636721312999725, + "learning_rate": 1.996323876510399e-05, + "loss": 0.081, + "step": 260 + }, + { + "epoch": 0.38308637513552585, + "grad_norm": 0.062323447316884995, + "learning_rate": 1.9955682931851835e-05, + "loss": 0.083, + "step": 265 + }, + { + "epoch": 0.3903144199494037, + "grad_norm": 0.0894196629524231, + "learning_rate": 1.994742312802255e-05, + "loss": 0.0815, + "step": 270 + }, + { + "epoch": 0.39754246476328153, + "grad_norm": 0.0037153863813728094, + "learning_rate": 1.993845993767038e-05, + "loss": 0.0801, + "step": 275 + }, + { + "epoch": 0.4047705095771594, + "grad_norm": 0.14254964888095856, + "learning_rate": 1.9928793994586323e-05, + "loss": 0.0857, + "step": 280 + }, + { + "epoch": 0.4119985543910372, + "grad_norm": 0.08120843023061752, + "learning_rate": 1.9918425982253335e-05, + "loss": 0.0822, + "step": 285 + }, + { + "epoch": 0.41922659920491506, + "grad_norm": 0.029054520651698112, + "learning_rate": 1.9907356633797978e-05, + "loss": 0.0813, + "step": 290 + }, + { + "epoch": 0.4264546440187929, + "grad_norm": 0.03537634015083313, + "learning_rate": 1.9895586731938593e-05, + "loss": 0.0812, + "step": 295 + }, + { + "epoch": 0.43368268883267075, + "grad_norm": 0.015365133993327618, + "learning_rate": 1.9883117108929947e-05, + "loss": 0.0825, + "step": 300 + }, + { + "epoch": 0.4409107336465486, + "grad_norm": 0.029979810118675232, + "learning_rate": 1.986994864650439e-05, + "loss": 0.0821, + "step": 305 + }, + { + "epoch": 0.44813877846042643, + "grad_norm": 0.021777283400297165, + "learning_rate": 1.9856082275809508e-05, + "loss": 0.0817, + "step": 310 + }, + { + "epoch": 0.4553668232743043, + "grad_norm": 0.005012670066207647, + "learning_rate": 1.9841518977342274e-05, + "loss": 0.081, + "step": 315 + }, + { + "epoch": 0.4625948680881821, + "grad_norm": 0.07129844278097153, + "learning_rate": 1.9826259780879716e-05, + "loss": 0.081, + "step": 320 + }, + { + "epoch": 0.46982291290206, + "grad_norm": 0.07923093438148499, + "learning_rate": 1.981030576540612e-05, + "loss": 0.0806, + "step": 325 + }, + { + "epoch": 0.47705095771593786, + "grad_norm": 0.04692668095231056, + "learning_rate": 1.9793658059036697e-05, + "loss": 0.0799, + "step": 330 + }, + { + "epoch": 0.4842790025298157, + "grad_norm": 0.04923313483595848, + "learning_rate": 1.977631783893786e-05, + "loss": 0.0813, + "step": 335 + }, + { + "epoch": 0.49150704734369355, + "grad_norm": 0.04896867647767067, + "learning_rate": 1.975828633124394e-05, + "loss": 0.0805, + "step": 340 + }, + { + "epoch": 0.4987350921575714, + "grad_norm": 0.004160281270742416, + "learning_rate": 1.9739564810970534e-05, + "loss": 0.0804, + "step": 345 + }, + { + "epoch": 0.5059631369714492, + "grad_norm": 0.07328899949789047, + "learning_rate": 1.9720154601924295e-05, + "loss": 0.0802, + "step": 350 + }, + { + "epoch": 0.5131911817853271, + "grad_norm": 0.007687863428145647, + "learning_rate": 1.9700057076609377e-05, + "loss": 0.08, + "step": 355 + }, + { + "epoch": 0.5204192265992049, + "grad_norm": 0.011853563599288464, + "learning_rate": 1.967927365613034e-05, + "loss": 0.0802, + "step": 360 + }, + { + "epoch": 0.5276472714130828, + "grad_norm": 0.007507723290473223, + "learning_rate": 1.96578058100917e-05, + "loss": 0.0799, + "step": 365 + }, + { + "epoch": 0.5348753162269606, + "grad_norm": 0.0316060446202755, + "learning_rate": 1.963565505649398e-05, + "loss": 0.081, + "step": 370 + }, + { + "epoch": 0.5421033610408384, + "grad_norm": 0.03697923943400383, + "learning_rate": 1.961282296162639e-05, + "loss": 0.0799, + "step": 375 + }, + { + "epoch": 0.5493314058547163, + "grad_norm": 0.007007018197327852, + "learning_rate": 1.9589311139956086e-05, + "loss": 0.0855, + "step": 380 + }, + { + "epoch": 0.5565594506685941, + "grad_norm": 0.024093549698591232, + "learning_rate": 1.956512125401398e-05, + "loss": 0.0809, + "step": 385 + }, + { + "epoch": 0.563787495482472, + "grad_norm": 0.007132918573915958, + "learning_rate": 1.9540255014277198e-05, + "loss": 0.0796, + "step": 390 + }, + { + "epoch": 0.5710155402963498, + "grad_norm": 0.055107131600379944, + "learning_rate": 1.9514714179048138e-05, + "loss": 0.0792, + "step": 395 + }, + { + "epoch": 0.5782435851102277, + "grad_norm": 0.036154747009277344, + "learning_rate": 1.9488500554330126e-05, + "loss": 0.0791, + "step": 400 + }, + { + "epoch": 0.5854716299241055, + "grad_norm": 0.02283984236419201, + "learning_rate": 1.946161599369973e-05, + "loss": 0.0794, + "step": 405 + }, + { + "epoch": 0.5926996747379834, + "grad_norm": 0.08648855239152908, + "learning_rate": 1.9434062398175667e-05, + "loss": 0.0776, + "step": 410 + }, + { + "epoch": 0.5999277195518612, + "grad_norm": 0.042574405670166016, + "learning_rate": 1.9405841716084403e-05, + "loss": 0.0772, + "step": 415 + }, + { + "epoch": 0.607155764365739, + "grad_norm": 0.035125792026519775, + "learning_rate": 1.937695594292238e-05, + "loss": 0.0765, + "step": 420 + }, + { + "epoch": 0.6143838091796169, + "grad_norm": 0.05990980565547943, + "learning_rate": 1.9347407121214917e-05, + "loss": 0.075, + "step": 425 + }, + { + "epoch": 0.6216118539934947, + "grad_norm": 0.13196605443954468, + "learning_rate": 1.9317197340371764e-05, + "loss": 0.0706, + "step": 430 + }, + { + "epoch": 0.6288398988073726, + "grad_norm": 0.04624694585800171, + "learning_rate": 1.9286328736539385e-05, + "loss": 0.0827, + "step": 435 + }, + { + "epoch": 0.6360679436212504, + "grad_norm": 0.03560846298933029, + "learning_rate": 1.9254803492449894e-05, + "loss": 0.0815, + "step": 440 + }, + { + "epoch": 0.6432959884351283, + "grad_norm": 0.0264581311494112, + "learning_rate": 1.922262383726672e-05, + "loss": 0.0816, + "step": 445 + }, + { + "epoch": 0.6505240332490061, + "grad_norm": 0.013737207278609276, + "learning_rate": 1.9189792046426972e-05, + "loss": 0.0795, + "step": 450 + }, + { + "epoch": 0.6577520780628839, + "grad_norm": 0.008679666556417942, + "learning_rate": 1.9156310441480557e-05, + "loss": 0.0795, + "step": 455 + }, + { + "epoch": 0.6649801228767618, + "grad_norm": 0.011995796114206314, + "learning_rate": 1.912218138992601e-05, + "loss": 0.0772, + "step": 460 + }, + { + "epoch": 0.6722081676906397, + "grad_norm": 0.025770675390958786, + "learning_rate": 1.9087407305043085e-05, + "loss": 0.0732, + "step": 465 + }, + { + "epoch": 0.6794362125045176, + "grad_norm": 0.11803118139505386, + "learning_rate": 1.9051990645722133e-05, + "loss": 0.0694, + "step": 470 + }, + { + "epoch": 0.6866642573183954, + "grad_norm": 0.0760372206568718, + "learning_rate": 1.9015933916290202e-05, + "loss": 0.0676, + "step": 475 + }, + { + "epoch": 0.6938923021322733, + "grad_norm": 0.03851527348160744, + "learning_rate": 1.8979239666333975e-05, + "loss": 0.0635, + "step": 480 + }, + { + "epoch": 0.7011203469461511, + "grad_norm": 0.08836951106786728, + "learning_rate": 1.8941910490519483e-05, + "loss": 0.0636, + "step": 485 + }, + { + "epoch": 0.708348391760029, + "grad_norm": 0.06829584389925003, + "learning_rate": 1.8903949028408636e-05, + "loss": 0.0648, + "step": 490 + }, + { + "epoch": 0.7155764365739068, + "grad_norm": 0.02876531518995762, + "learning_rate": 1.8865357964272576e-05, + "loss": 0.062, + "step": 495 + }, + { + "epoch": 0.7228044813877846, + "grad_norm": 0.039769161492586136, + "learning_rate": 1.8826140026901873e-05, + "loss": 0.0607, + "step": 500 + }, + { + "epoch": 0.7228044813877846, + "eval_loss": 0.07328393310308456, + "eval_runtime": 1144.2594, + "eval_samples_per_second": 56.111, + "eval_steps_per_second": 1.754, + "step": 500 + }, + { + "epoch": 0.7300325262016625, + "grad_norm": 0.04059358313679695, + "learning_rate": 1.878629798941357e-05, + "loss": 0.0631, + "step": 505 + }, + { + "epoch": 0.7372605710155403, + "grad_norm": 0.11944068223237991, + "learning_rate": 1.8745834669055085e-05, + "loss": 0.064, + "step": 510 + }, + { + "epoch": 0.7444886158294182, + "grad_norm": 0.04337216168642044, + "learning_rate": 1.8704752927005034e-05, + "loss": 0.0618, + "step": 515 + }, + { + "epoch": 0.751716660643296, + "grad_norm": 0.04100070148706436, + "learning_rate": 1.8663055668170873e-05, + "loss": 0.0613, + "step": 520 + }, + { + "epoch": 0.7589447054571739, + "grad_norm": 0.0634031817317009, + "learning_rate": 1.8620745840983522e-05, + "loss": 0.0589, + "step": 525 + }, + { + "epoch": 0.7661727502710517, + "grad_norm": 0.08053874224424362, + "learning_rate": 1.857782643718887e-05, + "loss": 0.0578, + "step": 530 + }, + { + "epoch": 0.7734007950849295, + "grad_norm": 0.037648145109415054, + "learning_rate": 1.8534300491636225e-05, + "loss": 0.0561, + "step": 535 + }, + { + "epoch": 0.7806288398988074, + "grad_norm": 0.07604615390300751, + "learning_rate": 1.849017108206372e-05, + "loss": 0.0615, + "step": 540 + }, + { + "epoch": 0.7878568847126852, + "grad_norm": 0.07877160608768463, + "learning_rate": 1.844544132888068e-05, + "loss": 0.0613, + "step": 545 + }, + { + "epoch": 0.7950849295265631, + "grad_norm": 0.09688904136419296, + "learning_rate": 1.8400114394947003e-05, + "loss": 0.0546, + "step": 550 + }, + { + "epoch": 0.8023129743404409, + "grad_norm": 0.0808292031288147, + "learning_rate": 1.8354193485349468e-05, + "loss": 0.0521, + "step": 555 + }, + { + "epoch": 0.8095410191543188, + "grad_norm": 0.05046294629573822, + "learning_rate": 1.830768184717514e-05, + "loss": 0.0581, + "step": 560 + }, + { + "epoch": 0.8167690639681966, + "grad_norm": 0.045729391276836395, + "learning_rate": 1.8260582769281747e-05, + "loss": 0.0596, + "step": 565 + }, + { + "epoch": 0.8239971087820744, + "grad_norm": 0.06313765794038773, + "learning_rate": 1.821289958206513e-05, + "loss": 0.0588, + "step": 570 + }, + { + "epoch": 0.8312251535959523, + "grad_norm": 0.04902196675539017, + "learning_rate": 1.8164635657223755e-05, + "loss": 0.0584, + "step": 575 + }, + { + "epoch": 0.8384531984098301, + "grad_norm": 0.08288609981536865, + "learning_rate": 1.8115794407520287e-05, + "loss": 0.0558, + "step": 580 + }, + { + "epoch": 0.845681243223708, + "grad_norm": 0.04202403128147125, + "learning_rate": 1.8066379286540278e-05, + "loss": 0.0524, + "step": 585 + }, + { + "epoch": 0.8529092880375858, + "grad_norm": 0.046127066016197205, + "learning_rate": 1.8016393788447964e-05, + "loss": 0.0549, + "step": 590 + }, + { + "epoch": 0.8601373328514637, + "grad_norm": 0.04576544463634491, + "learning_rate": 1.7965841447739185e-05, + "loss": 0.0491, + "step": 595 + }, + { + "epoch": 0.8673653776653415, + "grad_norm": 0.03939468041062355, + "learning_rate": 1.7914725838991472e-05, + "loss": 0.0543, + "step": 600 + }, + { + "epoch": 0.8745934224792193, + "grad_norm": 0.03499499708414078, + "learning_rate": 1.7863050576611267e-05, + "loss": 0.051, + "step": 605 + }, + { + "epoch": 0.8818214672930972, + "grad_norm": 0.04491008073091507, + "learning_rate": 1.781081931457837e-05, + "loss": 0.0553, + "step": 610 + }, + { + "epoch": 0.889049512106975, + "grad_norm": 0.06234387680888176, + "learning_rate": 1.7758035746187553e-05, + "loss": 0.0517, + "step": 615 + }, + { + "epoch": 0.8962775569208529, + "grad_norm": 0.04003310948610306, + "learning_rate": 1.770470360378739e-05, + "loss": 0.0545, + "step": 620 + }, + { + "epoch": 0.9035056017347307, + "grad_norm": 0.08948640525341034, + "learning_rate": 1.7650826658516375e-05, + "loss": 0.0548, + "step": 625 + }, + { + "epoch": 0.9107336465486086, + "grad_norm": 0.04328719154000282, + "learning_rate": 1.7596408720036232e-05, + "loss": 0.0551, + "step": 630 + }, + { + "epoch": 0.9179616913624864, + "grad_norm": 0.03560628369450569, + "learning_rate": 1.754145363626256e-05, + "loss": 0.05, + "step": 635 + }, + { + "epoch": 0.9251897361763642, + "grad_norm": 0.04258381202816963, + "learning_rate": 1.748596529309271e-05, + "loss": 0.0503, + "step": 640 + }, + { + "epoch": 0.9324177809902421, + "grad_norm": 0.07129397243261337, + "learning_rate": 1.742994761413105e-05, + "loss": 0.0488, + "step": 645 + }, + { + "epoch": 0.93964582580412, + "grad_norm": 0.03454764559864998, + "learning_rate": 1.73734045604115e-05, + "loss": 0.0553, + "step": 650 + }, + { + "epoch": 0.9468738706179979, + "grad_norm": 0.053935691714286804, + "learning_rate": 1.731634013011745e-05, + "loss": 0.0499, + "step": 655 + }, + { + "epoch": 0.9541019154318757, + "grad_norm": 0.08872876316308975, + "learning_rate": 1.7258758358299053e-05, + "loss": 0.0521, + "step": 660 + }, + { + "epoch": 0.9613299602457536, + "grad_norm": 0.057639699429273605, + "learning_rate": 1.7200663316587897e-05, + "loss": 0.0505, + "step": 665 + }, + { + "epoch": 0.9685580050596314, + "grad_norm": 0.08045148104429245, + "learning_rate": 1.7142059112909107e-05, + "loss": 0.0537, + "step": 670 + }, + { + "epoch": 0.9757860498735093, + "grad_norm": 0.05597732216119766, + "learning_rate": 1.708294989119087e-05, + "loss": 0.0466, + "step": 675 + }, + { + "epoch": 0.9830140946873871, + "grad_norm": 0.07576154917478561, + "learning_rate": 1.7023339831071408e-05, + "loss": 0.0527, + "step": 680 + }, + { + "epoch": 0.9902421395012649, + "grad_norm": 0.03821377828717232, + "learning_rate": 1.696323314760344e-05, + "loss": 0.0483, + "step": 685 + }, + { + "epoch": 0.9974701843151428, + "grad_norm": 0.07789347320795059, + "learning_rate": 1.690263409095614e-05, + "loss": 0.0483, + "step": 690 + }, + { + "epoch": 1.0043368268883268, + "grad_norm": 0.03913086652755737, + "learning_rate": 1.6841546946114586e-05, + "loss": 0.0417, + "step": 695 + }, + { + "epoch": 1.0115648717022045, + "grad_norm": 0.054379936307668686, + "learning_rate": 1.6779976032576792e-05, + "loss": 0.0336, + "step": 700 + }, + { + "epoch": 1.0187929165160825, + "grad_norm": 0.08715476840734482, + "learning_rate": 1.6717925704048256e-05, + "loss": 0.0416, + "step": 705 + }, + { + "epoch": 1.0260209613299602, + "grad_norm": 0.0766800120472908, + "learning_rate": 1.6655400348134122e-05, + "loss": 0.0404, + "step": 710 + }, + { + "epoch": 1.0332490061438382, + "grad_norm": 0.06571623682975769, + "learning_rate": 1.659240438602891e-05, + "loss": 0.0431, + "step": 715 + }, + { + "epoch": 1.040477050957716, + "grad_norm": 0.09106060862541199, + "learning_rate": 1.6528942272203912e-05, + "loss": 0.0419, + "step": 720 + }, + { + "epoch": 1.0477050957715939, + "grad_norm": 0.0675068348646164, + "learning_rate": 1.6465018494092213e-05, + "loss": 0.0401, + "step": 725 + }, + { + "epoch": 1.0549331405854716, + "grad_norm": 0.06592784821987152, + "learning_rate": 1.6400637571771354e-05, + "loss": 0.0383, + "step": 730 + }, + { + "epoch": 1.0621611853993496, + "grad_norm": 0.08610466867685318, + "learning_rate": 1.633580405764376e-05, + "loss": 0.0355, + "step": 735 + }, + { + "epoch": 1.0693892302132273, + "grad_norm": 0.09420937299728394, + "learning_rate": 1.6270522536114813e-05, + "loss": 0.0393, + "step": 740 + }, + { + "epoch": 1.0766172750271052, + "grad_norm": 0.066034696996212, + "learning_rate": 1.6204797623268675e-05, + "loss": 0.0353, + "step": 745 + }, + { + "epoch": 1.083845319840983, + "grad_norm": 0.05469588562846184, + "learning_rate": 1.6138633966541905e-05, + "loss": 0.0395, + "step": 750 + }, + { + "epoch": 1.091073364654861, + "grad_norm": 0.05333936959505081, + "learning_rate": 1.6072036244394836e-05, + "loss": 0.0409, + "step": 755 + }, + { + "epoch": 1.0983014094687387, + "grad_norm": 0.06300196051597595, + "learning_rate": 1.600500916598074e-05, + "loss": 0.0382, + "step": 760 + }, + { + "epoch": 1.1055294542826166, + "grad_norm": 0.08063532412052155, + "learning_rate": 1.5937557470812852e-05, + "loss": 0.0369, + "step": 765 + }, + { + "epoch": 1.1127574990964944, + "grad_norm": 0.07369716465473175, + "learning_rate": 1.5869685928429253e-05, + "loss": 0.0393, + "step": 770 + }, + { + "epoch": 1.1199855439103723, + "grad_norm": 0.058140210807323456, + "learning_rate": 1.5801399338055584e-05, + "loss": 0.0352, + "step": 775 + }, + { + "epoch": 1.12721358872425, + "grad_norm": 0.060627613216638565, + "learning_rate": 1.5732702528265716e-05, + "loss": 0.0381, + "step": 780 + }, + { + "epoch": 1.134441633538128, + "grad_norm": 0.07465813308954239, + "learning_rate": 1.5663600356640306e-05, + "loss": 0.0367, + "step": 785 + }, + { + "epoch": 1.1416696783520057, + "grad_norm": 0.047345198690891266, + "learning_rate": 1.5594097709423316e-05, + "loss": 0.0389, + "step": 790 + }, + { + "epoch": 1.1488977231658837, + "grad_norm": 0.06834863871335983, + "learning_rate": 1.552419950117651e-05, + "loss": 0.0381, + "step": 795 + }, + { + "epoch": 1.1561257679797614, + "grad_norm": 0.047312233597040176, + "learning_rate": 1.545391067443194e-05, + "loss": 0.0343, + "step": 800 + }, + { + "epoch": 1.1633538127936394, + "grad_norm": 0.06944846361875534, + "learning_rate": 1.538323619934247e-05, + "loss": 0.0389, + "step": 805 + }, + { + "epoch": 1.170581857607517, + "grad_norm": 0.04954347014427185, + "learning_rate": 1.5312181073330295e-05, + "loss": 0.0366, + "step": 810 + }, + { + "epoch": 1.177809902421395, + "grad_norm": 0.06755795329809189, + "learning_rate": 1.524075032073363e-05, + "loss": 0.0373, + "step": 815 + }, + { + "epoch": 1.1850379472352728, + "grad_norm": 0.08281169086694717, + "learning_rate": 1.5168948992451382e-05, + "loss": 0.0356, + "step": 820 + }, + { + "epoch": 1.1922659920491507, + "grad_norm": 0.08935344219207764, + "learning_rate": 1.5096782165586037e-05, + "loss": 0.0365, + "step": 825 + }, + { + "epoch": 1.1994940368630285, + "grad_norm": 0.04098968952894211, + "learning_rate": 1.5024254943084629e-05, + "loss": 0.0381, + "step": 830 + }, + { + "epoch": 1.2067220816769064, + "grad_norm": 0.05055451765656471, + "learning_rate": 1.495137245337794e-05, + "loss": 0.0367, + "step": 835 + }, + { + "epoch": 1.2139501264907842, + "grad_norm": 0.06718173623085022, + "learning_rate": 1.487813985001782e-05, + "loss": 0.0356, + "step": 840 + }, + { + "epoch": 1.221178171304662, + "grad_norm": 0.05843829736113548, + "learning_rate": 1.480456231131283e-05, + "loss": 0.0352, + "step": 845 + }, + { + "epoch": 1.2284062161185398, + "grad_norm": 0.052432768046855927, + "learning_rate": 1.4730645039962044e-05, + "loss": 0.0364, + "step": 850 + }, + { + "epoch": 1.2356342609324178, + "grad_norm": 0.05346972495317459, + "learning_rate": 1.4656393262687172e-05, + "loss": 0.0368, + "step": 855 + }, + { + "epoch": 1.2428623057462955, + "grad_norm": 0.06525395065546036, + "learning_rate": 1.4581812229862993e-05, + "loss": 0.0338, + "step": 860 + }, + { + "epoch": 1.2500903505601735, + "grad_norm": 0.07090573757886887, + "learning_rate": 1.4506907215146075e-05, + "loss": 0.0328, + "step": 865 + }, + { + "epoch": 1.2573183953740514, + "grad_norm": 0.05279651656746864, + "learning_rate": 1.443168351510189e-05, + "loss": 0.0348, + "step": 870 + }, + { + "epoch": 1.2645464401879292, + "grad_norm": 0.05886390060186386, + "learning_rate": 1.4356146448830277e-05, + "loss": 0.0341, + "step": 875 + }, + { + "epoch": 1.271774485001807, + "grad_norm": 0.0537516325712204, + "learning_rate": 1.4280301357589349e-05, + "loss": 0.0379, + "step": 880 + }, + { + "epoch": 1.2790025298156849, + "grad_norm": 0.049141135066747665, + "learning_rate": 1.4204153604417775e-05, + "loss": 0.0339, + "step": 885 + }, + { + "epoch": 1.2862305746295628, + "grad_norm": 0.05724327638745308, + "learning_rate": 1.4127708573755599e-05, + "loss": 0.0317, + "step": 890 + }, + { + "epoch": 1.2934586194434405, + "grad_norm": 0.05847681313753128, + "learning_rate": 1.4050971671063464e-05, + "loss": 0.0341, + "step": 895 + }, + { + "epoch": 1.3006866642573183, + "grad_norm": 0.04777985066175461, + "learning_rate": 1.3973948322440427e-05, + "loss": 0.0388, + "step": 900 + }, + { + "epoch": 1.3079147090711962, + "grad_norm": 0.062013089656829834, + "learning_rate": 1.3896643974240245e-05, + "loss": 0.0309, + "step": 905 + }, + { + "epoch": 1.3151427538850742, + "grad_norm": 0.08561990410089493, + "learning_rate": 1.3819064092686278e-05, + "loss": 0.0327, + "step": 910 + }, + { + "epoch": 1.322370798698952, + "grad_norm": 0.08605846017599106, + "learning_rate": 1.3741214163484968e-05, + "loss": 0.0345, + "step": 915 + }, + { + "epoch": 1.3295988435128296, + "grad_norm": 0.06043161824345589, + "learning_rate": 1.3663099691437945e-05, + "loss": 0.0336, + "step": 920 + }, + { + "epoch": 1.3368268883267076, + "grad_norm": 0.04175262525677681, + "learning_rate": 1.3584726200052767e-05, + "loss": 0.038, + "step": 925 + }, + { + "epoch": 1.3440549331405856, + "grad_norm": 0.04762093350291252, + "learning_rate": 1.3506099231152366e-05, + "loss": 0.0346, + "step": 930 + }, + { + "epoch": 1.3512829779544633, + "grad_norm": 0.06360676139593124, + "learning_rate": 1.3427224344483178e-05, + "loss": 0.0296, + "step": 935 + }, + { + "epoch": 1.3585110227683412, + "grad_norm": 0.06761486828327179, + "learning_rate": 1.3348107117322004e-05, + "loss": 0.0309, + "step": 940 + }, + { + "epoch": 1.365739067582219, + "grad_norm": 0.06754028797149658, + "learning_rate": 1.3268753144081652e-05, + "loss": 0.028, + "step": 945 + }, + { + "epoch": 1.372967112396097, + "grad_norm": 0.06639332324266434, + "learning_rate": 1.3189168035915337e-05, + "loss": 0.0331, + "step": 950 + }, + { + "epoch": 1.3801951572099747, + "grad_norm": 0.05263343080878258, + "learning_rate": 1.3109357420319933e-05, + "loss": 0.031, + "step": 955 + }, + { + "epoch": 1.3874232020238526, + "grad_norm": 0.07213468849658966, + "learning_rate": 1.3029326940738032e-05, + "loss": 0.0338, + "step": 960 + }, + { + "epoch": 1.3946512468377303, + "grad_norm": 0.05976350978016853, + "learning_rate": 1.2949082256158904e-05, + "loss": 0.0313, + "step": 965 + }, + { + "epoch": 1.4018792916516083, + "grad_norm": 0.054479606449604034, + "learning_rate": 1.286862904071835e-05, + "loss": 0.0324, + "step": 970 + }, + { + "epoch": 1.409107336465486, + "grad_norm": 0.07411843538284302, + "learning_rate": 1.2787972983297472e-05, + "loss": 0.0312, + "step": 975 + }, + { + "epoch": 1.416335381279364, + "grad_norm": 0.05356777831912041, + "learning_rate": 1.2707119787120417e-05, + "loss": 0.0347, + "step": 980 + }, + { + "epoch": 1.4235634260932417, + "grad_norm": 0.05905517190694809, + "learning_rate": 1.26260751693511e-05, + "loss": 0.0317, + "step": 985 + }, + { + "epoch": 1.4307914709071197, + "grad_norm": 0.07836019992828369, + "learning_rate": 1.254484486068893e-05, + "loss": 0.0316, + "step": 990 + }, + { + "epoch": 1.4380195157209974, + "grad_norm": 0.06900329887866974, + "learning_rate": 1.24634346049636e-05, + "loss": 0.0324, + "step": 995 + }, + { + "epoch": 1.4452475605348754, + "grad_norm": 0.05929545313119888, + "learning_rate": 1.2381850158728952e-05, + "loss": 0.029, + "step": 1000 + }, + { + "epoch": 1.4452475605348754, + "eval_loss": 0.08190815895795822, + "eval_runtime": 1141.9961, + "eval_samples_per_second": 56.223, + "eval_steps_per_second": 1.757, + "step": 1000 + }, + { + "epoch": 1.452475605348753, + "grad_norm": 0.0604124590754509, + "learning_rate": 1.2300097290855887e-05, + "loss": 0.0285, + "step": 1005 + }, + { + "epoch": 1.459703650162631, + "grad_norm": 0.06895657628774643, + "learning_rate": 1.2218181782124496e-05, + "loss": 0.0316, + "step": 1010 + }, + { + "epoch": 1.466931694976509, + "grad_norm": 0.048645876348018646, + "learning_rate": 1.2136109424815258e-05, + "loss": 0.0306, + "step": 1015 + }, + { + "epoch": 1.4741597397903867, + "grad_norm": 0.06193140521645546, + "learning_rate": 1.205388602229949e-05, + "loss": 0.03, + "step": 1020 + }, + { + "epoch": 1.4813877846042645, + "grad_norm": 0.07050759345293045, + "learning_rate": 1.1971517388628972e-05, + "loss": 0.0341, + "step": 1025 + }, + { + "epoch": 1.4886158294181424, + "grad_norm": 0.05533516779541969, + "learning_rate": 1.1889009348124857e-05, + "loss": 0.0303, + "step": 1030 + }, + { + "epoch": 1.4958438742320204, + "grad_norm": 0.04415017366409302, + "learning_rate": 1.180636773496579e-05, + "loss": 0.0276, + "step": 1035 + }, + { + "epoch": 1.503071919045898, + "grad_norm": 0.059612423181533813, + "learning_rate": 1.1723598392775415e-05, + "loss": 0.0273, + "step": 1040 + }, + { + "epoch": 1.5102999638597758, + "grad_norm": 0.06513796001672745, + "learning_rate": 1.1640707174209147e-05, + "loss": 0.0299, + "step": 1045 + }, + { + "epoch": 1.5175280086736538, + "grad_norm": 0.056087836623191833, + "learning_rate": 1.1557699940540321e-05, + "loss": 0.0308, + "step": 1050 + }, + { + "epoch": 1.5247560534875317, + "grad_norm": 0.06388755887746811, + "learning_rate": 1.1474582561245767e-05, + "loss": 0.0278, + "step": 1055 + }, + { + "epoch": 1.5319840983014095, + "grad_norm": 0.06793609261512756, + "learning_rate": 1.1391360913590736e-05, + "loss": 0.0342, + "step": 1060 + }, + { + "epoch": 1.5392121431152872, + "grad_norm": 0.04267344996333122, + "learning_rate": 1.1308040882213363e-05, + "loss": 0.03, + "step": 1065 + }, + { + "epoch": 1.5464401879291652, + "grad_norm": 0.06034848093986511, + "learning_rate": 1.122462835870852e-05, + "loss": 0.0279, + "step": 1070 + }, + { + "epoch": 1.553668232743043, + "grad_norm": 0.06860997527837753, + "learning_rate": 1.1141129241211246e-05, + "loss": 0.0263, + "step": 1075 + }, + { + "epoch": 1.5608962775569208, + "grad_norm": 0.04508688300848007, + "learning_rate": 1.1057549433979675e-05, + "loss": 0.0365, + "step": 1080 + }, + { + "epoch": 1.5681243223707986, + "grad_norm": 0.04600263386964798, + "learning_rate": 1.0973894846977548e-05, + "loss": 0.0286, + "step": 1085 + }, + { + "epoch": 1.5753523671846765, + "grad_norm": 0.05820371210575104, + "learning_rate": 1.089017139545631e-05, + "loss": 0.0264, + "step": 1090 + }, + { + "epoch": 1.5825804119985545, + "grad_norm": 0.0669277012348175, + "learning_rate": 1.0806384999536857e-05, + "loss": 0.028, + "step": 1095 + }, + { + "epoch": 1.5898084568124322, + "grad_norm": 0.05904907360672951, + "learning_rate": 1.0722541583790898e-05, + "loss": 0.0247, + "step": 1100 + }, + { + "epoch": 1.59703650162631, + "grad_norm": 0.05929577723145485, + "learning_rate": 1.0638647076822041e-05, + "loss": 0.0305, + "step": 1105 + }, + { + "epoch": 1.604264546440188, + "grad_norm": 0.058572858572006226, + "learning_rate": 1.0554707410846585e-05, + "loss": 0.0294, + "step": 1110 + }, + { + "epoch": 1.6114925912540659, + "grad_norm": 0.05682854354381561, + "learning_rate": 1.0470728521274028e-05, + "loss": 0.028, + "step": 1115 + }, + { + "epoch": 1.6187206360679436, + "grad_norm": 0.0703597441315651, + "learning_rate": 1.0386716346287398e-05, + "loss": 0.0278, + "step": 1120 + }, + { + "epoch": 1.6259486808818213, + "grad_norm": 0.07455068826675415, + "learning_rate": 1.030267682642334e-05, + "loss": 0.0312, + "step": 1125 + }, + { + "epoch": 1.6331767256956993, + "grad_norm": 0.06019241735339165, + "learning_rate": 1.0218615904152067e-05, + "loss": 0.027, + "step": 1130 + }, + { + "epoch": 1.6404047705095772, + "grad_norm": 0.05638565123081207, + "learning_rate": 1.0134539523457172e-05, + "loss": 0.0301, + "step": 1135 + }, + { + "epoch": 1.647632815323455, + "grad_norm": 0.06251167505979538, + "learning_rate": 1.0050453629415317e-05, + "loss": 0.027, + "step": 1140 + }, + { + "epoch": 1.6548608601373327, + "grad_norm": 0.08058342337608337, + "learning_rate": 9.966364167775851e-06, + "loss": 0.0307, + "step": 1145 + }, + { + "epoch": 1.6620889049512106, + "grad_norm": 0.05652245879173279, + "learning_rate": 9.882277084540399e-06, + "loss": 0.0267, + "step": 1150 + }, + { + "epoch": 1.6693169497650886, + "grad_norm": 0.06871891021728516, + "learning_rate": 9.798198325542399e-06, + "loss": 0.0255, + "step": 1155 + }, + { + "epoch": 1.6765449945789666, + "grad_norm": 0.07430125027894974, + "learning_rate": 9.714133836026687e-06, + "loss": 0.0277, + "step": 1160 + }, + { + "epoch": 1.6837730393928443, + "grad_norm": 0.058816712349653244, + "learning_rate": 9.630089560229088e-06, + "loss": 0.0248, + "step": 1165 + }, + { + "epoch": 1.691001084206722, + "grad_norm": 0.06506705284118652, + "learning_rate": 9.546071440956115e-06, + "loss": 0.0298, + "step": 1170 + }, + { + "epoch": 1.6982291290206, + "grad_norm": 0.06538432091474533, + "learning_rate": 9.46208541916474e-06, + "loss": 0.0308, + "step": 1175 + }, + { + "epoch": 1.705457173834478, + "grad_norm": 0.057376306504011154, + "learning_rate": 9.378137433542305e-06, + "loss": 0.0293, + "step": 1180 + }, + { + "epoch": 1.7126852186483557, + "grad_norm": 0.04726172983646393, + "learning_rate": 9.294233420086604e-06, + "loss": 0.0271, + "step": 1185 + }, + { + "epoch": 1.7199132634622334, + "grad_norm": 0.05788370966911316, + "learning_rate": 9.210379311686129e-06, + "loss": 0.0293, + "step": 1190 + }, + { + "epoch": 1.7271413082761113, + "grad_norm": 0.04595355689525604, + "learning_rate": 9.12658103770058e-06, + "loss": 0.0289, + "step": 1195 + }, + { + "epoch": 1.7343693530899893, + "grad_norm": 0.06266051530838013, + "learning_rate": 9.042844523541572e-06, + "loss": 0.0286, + "step": 1200 + }, + { + "epoch": 1.741597397903867, + "grad_norm": 0.049365997314453125, + "learning_rate": 8.95917569025366e-06, + "loss": 0.0275, + "step": 1205 + }, + { + "epoch": 1.7488254427177448, + "grad_norm": 0.056487612426280975, + "learning_rate": 8.875580454095651e-06, + "loss": 0.0239, + "step": 1210 + }, + { + "epoch": 1.7560534875316227, + "grad_norm": 0.04812345653772354, + "learning_rate": 8.792064726122275e-06, + "loss": 0.0262, + "step": 1215 + }, + { + "epoch": 1.7632815323455007, + "grad_norm": 0.06868524849414825, + "learning_rate": 8.708634411766195e-06, + "loss": 0.0277, + "step": 1220 + }, + { + "epoch": 1.7705095771593784, + "grad_norm": 0.07294084876775742, + "learning_rate": 8.625295410420451e-06, + "loss": 0.0235, + "step": 1225 + }, + { + "epoch": 1.7777376219732561, + "grad_norm": 0.05644133314490318, + "learning_rate": 8.542053615021291e-06, + "loss": 0.0271, + "step": 1230 + }, + { + "epoch": 1.784965666787134, + "grad_norm": 0.059861283749341965, + "learning_rate": 8.4589149116315e-06, + "loss": 0.025, + "step": 1235 + }, + { + "epoch": 1.792193711601012, + "grad_norm": 0.06358060985803604, + "learning_rate": 8.375885179024175e-06, + "loss": 0.0294, + "step": 1240 + }, + { + "epoch": 1.7994217564148898, + "grad_norm": 0.03532201051712036, + "learning_rate": 8.292970288267043e-06, + "loss": 0.0239, + "step": 1245 + }, + { + "epoch": 1.8066498012287675, + "grad_norm": 0.047285765409469604, + "learning_rate": 8.21017610230732e-06, + "loss": 0.0312, + "step": 1250 + }, + { + "epoch": 1.8138778460426455, + "grad_norm": 0.044171951711177826, + "learning_rate": 8.12750847555713e-06, + "loss": 0.0288, + "step": 1255 + }, + { + "epoch": 1.8211058908565234, + "grad_norm": 0.05230150744318962, + "learning_rate": 8.044973253479544e-06, + "loss": 0.0242, + "step": 1260 + }, + { + "epoch": 1.8283339356704011, + "grad_norm": 0.04772350192070007, + "learning_rate": 7.96257627217524e-06, + "loss": 0.0292, + "step": 1265 + }, + { + "epoch": 1.8355619804842789, + "grad_norm": 0.04245223104953766, + "learning_rate": 7.880323357969838e-06, + "loss": 0.0239, + "step": 1270 + }, + { + "epoch": 1.8427900252981568, + "grad_norm": 0.05859874188899994, + "learning_rate": 7.798220327001898e-06, + "loss": 0.0245, + "step": 1275 + }, + { + "epoch": 1.8500180701120348, + "grad_norm": 0.06144941225647926, + "learning_rate": 7.716272984811688e-06, + "loss": 0.0261, + "step": 1280 + }, + { + "epoch": 1.8572461149259125, + "grad_norm": 0.03744060546159744, + "learning_rate": 7.634487125930649e-06, + "loss": 0.0259, + "step": 1285 + }, + { + "epoch": 1.8644741597397902, + "grad_norm": 0.06158106401562691, + "learning_rate": 7.55286853347167e-06, + "loss": 0.0237, + "step": 1290 + }, + { + "epoch": 1.8717022045536682, + "grad_norm": 0.05013835057616234, + "learning_rate": 7.471422978720162e-06, + "loss": 0.0244, + "step": 1295 + }, + { + "epoch": 1.8789302493675462, + "grad_norm": 0.06363669037818909, + "learning_rate": 7.3901562207259555e-06, + "loss": 0.0245, + "step": 1300 + }, + { + "epoch": 1.8861582941814239, + "grad_norm": 0.05522134155035019, + "learning_rate": 7.309074005896103e-06, + "loss": 0.0216, + "step": 1305 + }, + { + "epoch": 1.8933863389953016, + "grad_norm": 0.06466201692819595, + "learning_rate": 7.228182067588518e-06, + "loss": 0.0278, + "step": 1310 + }, + { + "epoch": 1.9006143838091796, + "grad_norm": 0.047263894230127335, + "learning_rate": 7.1474861257065866e-06, + "loss": 0.0258, + "step": 1315 + }, + { + "epoch": 1.9078424286230575, + "grad_norm": 0.051960770040750504, + "learning_rate": 7.066991886294702e-06, + "loss": 0.0227, + "step": 1320 + }, + { + "epoch": 1.9150704734369353, + "grad_norm": 0.06168799102306366, + "learning_rate": 6.9867050411347955e-06, + "loss": 0.0333, + "step": 1325 + }, + { + "epoch": 1.922298518250813, + "grad_norm": 0.03766432777047157, + "learning_rate": 6.906631267343849e-06, + "loss": 0.0235, + "step": 1330 + }, + { + "epoch": 1.929526563064691, + "grad_norm": 0.04953250661492348, + "learning_rate": 6.826776226972489e-06, + "loss": 0.0247, + "step": 1335 + }, + { + "epoch": 1.936754607878569, + "grad_norm": 0.047898851335048676, + "learning_rate": 6.747145566604605e-06, + "loss": 0.0281, + "step": 1340 + }, + { + "epoch": 1.9439826526924469, + "grad_norm": 0.062446679919958115, + "learning_rate": 6.667744916958085e-06, + "loss": 0.0242, + "step": 1345 + }, + { + "epoch": 1.9512106975063246, + "grad_norm": 0.050179507583379745, + "learning_rate": 6.588579892486657e-06, + "loss": 0.0254, + "step": 1350 + }, + { + "epoch": 1.9584387423202023, + "grad_norm": 0.052683789283037186, + "learning_rate": 6.5096560909828855e-06, + "loss": 0.0206, + "step": 1355 + }, + { + "epoch": 1.9656667871340803, + "grad_norm": 0.06998462975025177, + "learning_rate": 6.430979093182372e-06, + "loss": 0.0223, + "step": 1360 + }, + { + "epoch": 1.9728948319479582, + "grad_norm": 0.07918884605169296, + "learning_rate": 6.352554462369112e-06, + "loss": 0.0281, + "step": 1365 + }, + { + "epoch": 1.980122876761836, + "grad_norm": 0.06278680264949799, + "learning_rate": 6.274387743982127e-06, + "loss": 0.0234, + "step": 1370 + }, + { + "epoch": 1.9873509215757137, + "grad_norm": 0.04667511582374573, + "learning_rate": 6.196484465223343e-06, + "loss": 0.0218, + "step": 1375 + }, + { + "epoch": 1.9945789663895916, + "grad_norm": 0.05683530122041702, + "learning_rate": 6.1188501346667536e-06, + "loss": 0.0267, + "step": 1380 + }, + { + "epoch": 2.0014456089627757, + "grad_norm": 0.021900292485952377, + "learning_rate": 6.04149024186891e-06, + "loss": 0.0196, + "step": 1385 + }, + { + "epoch": 2.0086736537766536, + "grad_norm": 0.014646291732788086, + "learning_rate": 5.964410256980762e-06, + "loss": 0.0085, + "step": 1390 + }, + { + "epoch": 2.015901698590531, + "grad_norm": 0.018468832597136497, + "learning_rate": 5.887615630360836e-06, + "loss": 0.0064, + "step": 1395 + }, + { + "epoch": 2.023129743404409, + "grad_norm": 0.02347305603325367, + "learning_rate": 5.811111792189873e-06, + "loss": 0.0058, + "step": 1400 + }, + { + "epoch": 2.030357788218287, + "grad_norm": 0.022464651614427567, + "learning_rate": 5.734904152086829e-06, + "loss": 0.0052, + "step": 1405 + }, + { + "epoch": 2.037585833032165, + "grad_norm": 0.027769049629569054, + "learning_rate": 5.658998098726361e-06, + "loss": 0.0054, + "step": 1410 + }, + { + "epoch": 2.0448138778460425, + "grad_norm": 0.03556771203875542, + "learning_rate": 5.583398999457812e-06, + "loss": 0.0043, + "step": 1415 + }, + { + "epoch": 2.0520419226599205, + "grad_norm": 0.030191823840141296, + "learning_rate": 5.508112199925659e-06, + "loss": 0.0041, + "step": 1420 + }, + { + "epoch": 2.0592699674737984, + "grad_norm": 0.03760818764567375, + "learning_rate": 5.433143023691547e-06, + "loss": 0.0038, + "step": 1425 + }, + { + "epoch": 2.0664980122876764, + "grad_norm": 0.014797261916100979, + "learning_rate": 5.358496771857831e-06, + "loss": 0.0044, + "step": 1430 + }, + { + "epoch": 2.073726057101554, + "grad_norm": 0.06623335927724838, + "learning_rate": 5.284178722692743e-06, + "loss": 0.0052, + "step": 1435 + }, + { + "epoch": 2.080954101915432, + "grad_norm": 0.016557743772864342, + "learning_rate": 5.2101941312571724e-06, + "loss": 0.0037, + "step": 1440 + }, + { + "epoch": 2.08818214672931, + "grad_norm": 0.03200926259160042, + "learning_rate": 5.136548229033065e-06, + "loss": 0.0047, + "step": 1445 + }, + { + "epoch": 2.0954101915431878, + "grad_norm": 0.03867388516664505, + "learning_rate": 5.063246223553509e-06, + "loss": 0.0052, + "step": 1450 + }, + { + "epoch": 2.1026382363570653, + "grad_norm": 0.009015249088406563, + "learning_rate": 4.990293298034505e-06, + "loss": 0.0045, + "step": 1455 + }, + { + "epoch": 2.109866281170943, + "grad_norm": 0.01907913200557232, + "learning_rate": 4.917694611008477e-06, + "loss": 0.0041, + "step": 1460 + }, + { + "epoch": 2.117094325984821, + "grad_norm": 0.02901625819504261, + "learning_rate": 4.845455295959468e-06, + "loss": 0.0051, + "step": 1465 + }, + { + "epoch": 2.124322370798699, + "grad_norm": 0.03763509541749954, + "learning_rate": 4.773580460960195e-06, + "loss": 0.0065, + "step": 1470 + }, + { + "epoch": 2.1315504156125766, + "grad_norm": 0.02262153849005699, + "learning_rate": 4.702075188310826e-06, + "loss": 0.0052, + "step": 1475 + }, + { + "epoch": 2.1387784604264546, + "grad_norm": 0.02351069077849388, + "learning_rate": 4.6309445341796286e-06, + "loss": 0.0048, + "step": 1480 + }, + { + "epoch": 2.1460065052403325, + "grad_norm": 0.009482895024120808, + "learning_rate": 4.5601935282454255e-06, + "loss": 0.0035, + "step": 1485 + }, + { + "epoch": 2.1532345500542105, + "grad_norm": 0.039236586540937424, + "learning_rate": 4.489827173341957e-06, + "loss": 0.0046, + "step": 1490 + }, + { + "epoch": 2.160462594868088, + "grad_norm": 0.029299462214112282, + "learning_rate": 4.419850445104126e-06, + "loss": 0.0066, + "step": 1495 + }, + { + "epoch": 2.167690639681966, + "grad_norm": 0.038081999868154526, + "learning_rate": 4.350268291616166e-06, + "loss": 0.0058, + "step": 1500 + }, + { + "epoch": 2.167690639681966, + "eval_loss": 0.15236619114875793, + "eval_runtime": 1142.5896, + "eval_samples_per_second": 56.193, + "eval_steps_per_second": 1.757, + "step": 1500 + }, + { + "epoch": 2.174918684495844, + "grad_norm": 0.05242437124252319, + "learning_rate": 4.281085633061764e-06, + "loss": 0.0047, + "step": 1505 + }, + { + "epoch": 2.182146729309722, + "grad_norm": 0.04368291050195694, + "learning_rate": 4.212307361376146e-06, + "loss": 0.0078, + "step": 1510 + }, + { + "epoch": 2.1893747741235994, + "grad_norm": 0.033737700432538986, + "learning_rate": 4.1439383399001865e-06, + "loss": 0.0043, + "step": 1515 + }, + { + "epoch": 2.1966028189374773, + "grad_norm": 0.05659673735499382, + "learning_rate": 4.075983403036479e-06, + "loss": 0.0045, + "step": 1520 + }, + { + "epoch": 2.2038308637513553, + "grad_norm": 0.0330926850438118, + "learning_rate": 4.0084473559075335e-06, + "loss": 0.0073, + "step": 1525 + }, + { + "epoch": 2.2110589085652332, + "grad_norm": 0.01673804223537445, + "learning_rate": 3.941334974015981e-06, + "loss": 0.006, + "step": 1530 + }, + { + "epoch": 2.2182869533791107, + "grad_norm": 0.013828652910888195, + "learning_rate": 3.874651002906915e-06, + "loss": 0.0047, + "step": 1535 + }, + { + "epoch": 2.2255149981929887, + "grad_norm": 0.02410770393908024, + "learning_rate": 3.8084001578323093e-06, + "loss": 0.0049, + "step": 1540 + }, + { + "epoch": 2.2327430430068667, + "grad_norm": 0.04236437752842903, + "learning_rate": 3.7425871234176134e-06, + "loss": 0.0069, + "step": 1545 + }, + { + "epoch": 2.2399710878207446, + "grad_norm": 0.017214614897966385, + "learning_rate": 3.6772165533305024e-06, + "loss": 0.0034, + "step": 1550 + }, + { + "epoch": 2.2471991326346226, + "grad_norm": 0.007290941663086414, + "learning_rate": 3.6122930699518057e-06, + "loss": 0.0045, + "step": 1555 + }, + { + "epoch": 2.2544271774485, + "grad_norm": 0.02728499099612236, + "learning_rate": 3.5478212640486652e-06, + "loss": 0.0049, + "step": 1560 + }, + { + "epoch": 2.261655222262378, + "grad_norm": 0.023531029000878334, + "learning_rate": 3.483805694449913e-06, + "loss": 0.0034, + "step": 1565 + }, + { + "epoch": 2.268883267076256, + "grad_norm": 0.01806485652923584, + "learning_rate": 3.420250887723722e-06, + "loss": 0.0044, + "step": 1570 + }, + { + "epoch": 2.2761113118901335, + "grad_norm": 0.022033169865608215, + "learning_rate": 3.357161337857523e-06, + "loss": 0.0037, + "step": 1575 + }, + { + "epoch": 2.2833393567040114, + "grad_norm": 0.03150279447436333, + "learning_rate": 3.2945415059402363e-06, + "loss": 0.0051, + "step": 1580 + }, + { + "epoch": 2.2905674015178894, + "grad_norm": 0.01539881806820631, + "learning_rate": 3.232395819846824e-06, + "loss": 0.0036, + "step": 1585 + }, + { + "epoch": 2.2977954463317674, + "grad_norm": 0.022633062675595284, + "learning_rate": 3.170728673925206e-06, + "loss": 0.003, + "step": 1590 + }, + { + "epoch": 2.3050234911456453, + "grad_norm": 0.07029638439416885, + "learning_rate": 3.1095444286855112e-06, + "loss": 0.0056, + "step": 1595 + }, + { + "epoch": 2.312251535959523, + "grad_norm": 0.02109723724424839, + "learning_rate": 3.04884741049176e-06, + "loss": 0.0047, + "step": 1600 + }, + { + "epoch": 2.3194795807734008, + "grad_norm": 0.03571590408682823, + "learning_rate": 2.9886419112559396e-06, + "loss": 0.005, + "step": 1605 + }, + { + "epoch": 2.3267076255872787, + "grad_norm": 0.047896191477775574, + "learning_rate": 2.9289321881345257e-06, + "loss": 0.0065, + "step": 1610 + }, + { + "epoch": 2.3339356704011567, + "grad_norm": 0.018771937116980553, + "learning_rate": 2.86972246322745e-06, + "loss": 0.0029, + "step": 1615 + }, + { + "epoch": 2.341163715215034, + "grad_norm": 0.013248492032289505, + "learning_rate": 2.8110169232795615e-06, + "loss": 0.0029, + "step": 1620 + }, + { + "epoch": 2.348391760028912, + "grad_norm": 0.0571102574467659, + "learning_rate": 2.752819719384573e-06, + "loss": 0.0059, + "step": 1625 + }, + { + "epoch": 2.35561980484279, + "grad_norm": 0.037497229874134064, + "learning_rate": 2.6951349666915404e-06, + "loss": 0.0046, + "step": 1630 + }, + { + "epoch": 2.362847849656668, + "grad_norm": 0.050996676087379456, + "learning_rate": 2.637966744113877e-06, + "loss": 0.005, + "step": 1635 + }, + { + "epoch": 2.3700758944705456, + "grad_norm": 0.017970601096749306, + "learning_rate": 2.581319094040927e-06, + "loss": 0.0046, + "step": 1640 + }, + { + "epoch": 2.3773039392844235, + "grad_norm": 0.012019157409667969, + "learning_rate": 2.5251960220521422e-06, + "loss": 0.0036, + "step": 1645 + }, + { + "epoch": 2.3845319840983015, + "grad_norm": 0.04049897938966751, + "learning_rate": 2.4696014966338267e-06, + "loss": 0.0043, + "step": 1650 + }, + { + "epoch": 2.3917600289121794, + "grad_norm": 0.011841571889817715, + "learning_rate": 2.4145394488985307e-06, + "loss": 0.0031, + "step": 1655 + }, + { + "epoch": 2.398988073726057, + "grad_norm": 0.05339455232024193, + "learning_rate": 2.360013772307086e-06, + "loss": 0.0039, + "step": 1660 + }, + { + "epoch": 2.406216118539935, + "grad_norm": 0.017674589529633522, + "learning_rate": 2.3060283223932876e-06, + "loss": 0.0041, + "step": 1665 + }, + { + "epoch": 2.413444163353813, + "grad_norm": 0.036407146602869034, + "learning_rate": 2.252586916491275e-06, + "loss": 0.0062, + "step": 1670 + }, + { + "epoch": 2.420672208167691, + "grad_norm": 0.05213891342282295, + "learning_rate": 2.1996933334656044e-06, + "loss": 0.0048, + "step": 1675 + }, + { + "epoch": 2.4279002529815683, + "grad_norm": 0.027244996279478073, + "learning_rate": 2.1473513134440425e-06, + "loss": 0.0064, + "step": 1680 + }, + { + "epoch": 2.4351282977954463, + "grad_norm": 0.020630542188882828, + "learning_rate": 2.0955645575531e-06, + "loss": 0.005, + "step": 1685 + }, + { + "epoch": 2.442356342609324, + "grad_norm": 0.01828751713037491, + "learning_rate": 2.0443367276563277e-06, + "loss": 0.0056, + "step": 1690 + }, + { + "epoch": 2.449584387423202, + "grad_norm": 0.07830678671598434, + "learning_rate": 1.9936714460953743e-06, + "loss": 0.0058, + "step": 1695 + }, + { + "epoch": 2.4568124322370797, + "grad_norm": 0.04249007627367973, + "learning_rate": 1.9435722954338675e-06, + "loss": 0.0038, + "step": 1700 + }, + { + "epoch": 2.4640404770509576, + "grad_norm": 0.04109486937522888, + "learning_rate": 1.8940428182040715e-06, + "loss": 0.0042, + "step": 1705 + }, + { + "epoch": 2.4712685218648356, + "grad_norm": 0.011558642610907555, + "learning_rate": 1.8450865166564003e-06, + "loss": 0.0029, + "step": 1710 + }, + { + "epoch": 2.4784965666787135, + "grad_norm": 0.0303326603025198, + "learning_rate": 1.7967068525117658e-06, + "loss": 0.0041, + "step": 1715 + }, + { + "epoch": 2.485724611492591, + "grad_norm": 0.016660748049616814, + "learning_rate": 1.7489072467168166e-06, + "loss": 0.0039, + "step": 1720 + }, + { + "epoch": 2.492952656306469, + "grad_norm": 0.044775962829589844, + "learning_rate": 1.7016910792020191e-06, + "loss": 0.0062, + "step": 1725 + }, + { + "epoch": 2.500180701120347, + "grad_norm": 0.018128257244825363, + "learning_rate": 1.6550616886426718e-06, + "loss": 0.0033, + "step": 1730 + }, + { + "epoch": 2.507408745934225, + "grad_norm": 0.04440128430724144, + "learning_rate": 1.609022372222827e-06, + "loss": 0.005, + "step": 1735 + }, + { + "epoch": 2.514636790748103, + "grad_norm": 0.013020209036767483, + "learning_rate": 1.5635763854021424e-06, + "loss": 0.004, + "step": 1740 + }, + { + "epoch": 2.5218648355619804, + "grad_norm": 0.04560156539082527, + "learning_rate": 1.5187269416856875e-06, + "loss": 0.0044, + "step": 1745 + }, + { + "epoch": 2.5290928803758583, + "grad_norm": 0.020239338278770447, + "learning_rate": 1.474477212396712e-06, + "loss": 0.003, + "step": 1750 + }, + { + "epoch": 2.5363209251897363, + "grad_norm": 0.020898908376693726, + "learning_rate": 1.4308303264524115e-06, + "loss": 0.0046, + "step": 1755 + }, + { + "epoch": 2.543548970003614, + "grad_norm": 0.01411470677703619, + "learning_rate": 1.3877893701426637e-06, + "loss": 0.0037, + "step": 1760 + }, + { + "epoch": 2.5507770148174918, + "grad_norm": 0.027346884831786156, + "learning_rate": 1.3453573869118097e-06, + "loss": 0.0056, + "step": 1765 + }, + { + "epoch": 2.5580050596313697, + "grad_norm": 0.017516661435365677, + "learning_rate": 1.3035373771434356e-06, + "loss": 0.0063, + "step": 1770 + }, + { + "epoch": 2.5652331044452477, + "grad_norm": 0.01695055328309536, + "learning_rate": 1.2623322979482355e-06, + "loss": 0.0049, + "step": 1775 + }, + { + "epoch": 2.5724611492591256, + "grad_norm": 0.03533555567264557, + "learning_rate": 1.2217450629548955e-06, + "loss": 0.0042, + "step": 1780 + }, + { + "epoch": 2.579689194073003, + "grad_norm": 0.012438401579856873, + "learning_rate": 1.181778542104075e-06, + "loss": 0.0025, + "step": 1785 + }, + { + "epoch": 2.586917238886881, + "grad_norm": 0.02840145118534565, + "learning_rate": 1.1424355614454718e-06, + "loss": 0.0047, + "step": 1790 + }, + { + "epoch": 2.594145283700759, + "grad_norm": 0.03050726279616356, + "learning_rate": 1.1037189029379925e-06, + "loss": 0.0078, + "step": 1795 + }, + { + "epoch": 2.6013733285146365, + "grad_norm": 0.025619490072131157, + "learning_rate": 1.0656313042530376e-06, + "loss": 0.0051, + "step": 1800 + }, + { + "epoch": 2.6086013733285145, + "grad_norm": 0.03565088286995888, + "learning_rate": 1.028175458580918e-06, + "loss": 0.0057, + "step": 1805 + }, + { + "epoch": 2.6158294181423924, + "grad_norm": 0.048903122544288635, + "learning_rate": 9.913540144404254e-07, + "loss": 0.0029, + "step": 1810 + }, + { + "epoch": 2.6230574629562704, + "grad_norm": 0.06714732199907303, + "learning_rate": 9.551695754915447e-07, + "loss": 0.0058, + "step": 1815 + }, + { + "epoch": 2.6302855077701484, + "grad_norm": 0.03730113059282303, + "learning_rate": 9.196247003513537e-07, + "loss": 0.0056, + "step": 1820 + }, + { + "epoch": 2.637513552584026, + "grad_norm": 0.008724790997803211, + "learning_rate": 8.84721902413097e-07, + "loss": 0.0042, + "step": 1825 + }, + { + "epoch": 2.644741597397904, + "grad_norm": 0.03197433799505234, + "learning_rate": 8.50463649668477e-07, + "loss": 0.0043, + "step": 1830 + }, + { + "epoch": 2.6519696422117818, + "grad_norm": 0.05495726689696312, + "learning_rate": 8.168523645331216e-07, + "loss": 0.0047, + "step": 1835 + }, + { + "epoch": 2.6591976870256593, + "grad_norm": 0.01701589673757553, + "learning_rate": 7.838904236753087e-07, + "loss": 0.0041, + "step": 1840 + }, + { + "epoch": 2.6664257318395372, + "grad_norm": 0.02677042968571186, + "learning_rate": 7.515801578479032e-07, + "loss": 0.0065, + "step": 1845 + }, + { + "epoch": 2.673653776653415, + "grad_norm": 0.014987285248935223, + "learning_rate": 7.199238517235541e-07, + "loss": 0.003, + "step": 1850 + }, + { + "epoch": 2.680881821467293, + "grad_norm": 0.011919076554477215, + "learning_rate": 6.889237437331398e-07, + "loss": 0.0036, + "step": 1855 + }, + { + "epoch": 2.688109866281171, + "grad_norm": 0.05691038444638252, + "learning_rate": 6.585820259074882e-07, + "loss": 0.005, + "step": 1860 + }, + { + "epoch": 2.6953379110950486, + "grad_norm": 0.021997489035129547, + "learning_rate": 6.289008437223798e-07, + "loss": 0.0061, + "step": 1865 + }, + { + "epoch": 2.7025659559089266, + "grad_norm": 0.04470158740878105, + "learning_rate": 5.998822959468409e-07, + "loss": 0.0044, + "step": 1870 + }, + { + "epoch": 2.7097940007228045, + "grad_norm": 0.023458922281861305, + "learning_rate": 5.715284344947358e-07, + "loss": 0.0052, + "step": 1875 + }, + { + "epoch": 2.7170220455366825, + "grad_norm": 0.007212420925498009, + "learning_rate": 5.438412642796686e-07, + "loss": 0.004, + "step": 1880 + }, + { + "epoch": 2.7242500903505604, + "grad_norm": 0.021170541644096375, + "learning_rate": 5.168227430732353e-07, + "loss": 0.0046, + "step": 1885 + }, + { + "epoch": 2.731478135164438, + "grad_norm": 0.04506688937544823, + "learning_rate": 4.904747813665656e-07, + "loss": 0.005, + "step": 1890 + }, + { + "epoch": 2.738706179978316, + "grad_norm": 0.03043074533343315, + "learning_rate": 4.6479924223524655e-07, + "loss": 0.0056, + "step": 1895 + }, + { + "epoch": 2.745934224792194, + "grad_norm": 0.06903711706399918, + "learning_rate": 4.39797941207577e-07, + "loss": 0.004, + "step": 1900 + }, + { + "epoch": 2.7531622696060714, + "grad_norm": 0.008001566864550114, + "learning_rate": 4.1547264613619243e-07, + "loss": 0.0052, + "step": 1905 + }, + { + "epoch": 2.7603903144199493, + "grad_norm": 0.016265859827399254, + "learning_rate": 3.9182507707305915e-07, + "loss": 0.0055, + "step": 1910 + }, + { + "epoch": 2.7676183592338273, + "grad_norm": 0.019273990765213966, + "learning_rate": 3.6885690614785197e-07, + "loss": 0.0043, + "step": 1915 + }, + { + "epoch": 2.774846404047705, + "grad_norm": 0.051555391401052475, + "learning_rate": 3.4656975744970846e-07, + "loss": 0.0046, + "step": 1920 + }, + { + "epoch": 2.782074448861583, + "grad_norm": 0.010696332901716232, + "learning_rate": 3.249652069124032e-07, + "loss": 0.0028, + "step": 1925 + }, + { + "epoch": 2.7893024936754607, + "grad_norm": 0.04233001545071602, + "learning_rate": 3.040447822028958e-07, + "loss": 0.0048, + "step": 1930 + }, + { + "epoch": 2.7965305384893386, + "grad_norm": 0.07739260792732239, + "learning_rate": 2.838099626133206e-07, + "loss": 0.004, + "step": 1935 + }, + { + "epoch": 2.8037585833032166, + "grad_norm": 0.05327356979250908, + "learning_rate": 2.642621789563848e-07, + "loss": 0.0069, + "step": 1940 + }, + { + "epoch": 2.810986628117094, + "grad_norm": 0.027605120092630386, + "learning_rate": 2.4540281346418946e-07, + "loss": 0.0037, + "step": 1945 + }, + { + "epoch": 2.818214672930972, + "grad_norm": 0.019115762785077095, + "learning_rate": 2.2723319969049307e-07, + "loss": 0.0036, + "step": 1950 + }, + { + "epoch": 2.82544271774485, + "grad_norm": 0.02181391790509224, + "learning_rate": 2.0975462241642042e-07, + "loss": 0.0024, + "step": 1955 + }, + { + "epoch": 2.832670762558728, + "grad_norm": 0.009364648722112179, + "learning_rate": 1.9296831755960753e-07, + "loss": 0.0052, + "step": 1960 + }, + { + "epoch": 2.839898807372606, + "grad_norm": 0.01776730641722679, + "learning_rate": 1.76875472086816e-07, + "loss": 0.0037, + "step": 1965 + }, + { + "epoch": 2.8471268521864834, + "grad_norm": 0.022552713751792908, + "learning_rate": 1.6147722392999887e-07, + "loss": 0.0038, + "step": 1970 + }, + { + "epoch": 2.8543548970003614, + "grad_norm": 0.05256934091448784, + "learning_rate": 1.467746619058341e-07, + "loss": 0.0049, + "step": 1975 + }, + { + "epoch": 2.8615829418142393, + "grad_norm": 0.01704435609281063, + "learning_rate": 1.327688256387416e-07, + "loss": 0.0042, + "step": 1980 + }, + { + "epoch": 2.868810986628117, + "grad_norm": 0.008595878258347511, + "learning_rate": 1.1946070548736532e-07, + "loss": 0.0035, + "step": 1985 + }, + { + "epoch": 2.876039031441995, + "grad_norm": 0.025025706738233566, + "learning_rate": 1.0685124247454159e-07, + "loss": 0.0038, + "step": 1990 + }, + { + "epoch": 2.8832670762558728, + "grad_norm": 0.018959928303956985, + "learning_rate": 9.494132822077007e-08, + "loss": 0.0043, + "step": 1995 + }, + { + "epoch": 2.8904951210697507, + "grad_norm": 0.01306887436658144, + "learning_rate": 8.373180488115529e-08, + "loss": 0.005, + "step": 2000 + }, + { + "epoch": 2.8904951210697507, + "eval_loss": 0.1443248987197876, + "eval_runtime": 1199.7992, + "eval_samples_per_second": 53.514, + "eval_steps_per_second": 1.673, + "step": 2000 + }, + { + "epoch": 2.8977231658836287, + "grad_norm": 0.07084480673074722, + "learning_rate": 7.322346508586209e-08, + "loss": 0.0043, + "step": 2005 + }, + { + "epoch": 2.904951210697506, + "grad_norm": 0.04223432019352913, + "learning_rate": 6.341705188407043e-08, + "loss": 0.0048, + "step": 2010 + }, + { + "epoch": 2.912179255511384, + "grad_norm": 0.019037162885069847, + "learning_rate": 5.431325869143189e-08, + "loss": 0.0054, + "step": 2015 + }, + { + "epoch": 2.919407300325262, + "grad_norm": 0.014710099436342716, + "learning_rate": 4.5912729241036624e-08, + "loss": 0.0038, + "step": 2020 + }, + { + "epoch": 2.9266353451391396, + "grad_norm": 0.012474890798330307, + "learning_rate": 3.821605753789648e-08, + "loss": 0.0038, + "step": 2025 + }, + { + "epoch": 2.933863389953018, + "grad_norm": 0.01217850111424923, + "learning_rate": 3.122378781694524e-08, + "loss": 0.0028, + "step": 2030 + }, + { + "epoch": 2.9410914347668955, + "grad_norm": 0.0497884601354599, + "learning_rate": 2.493641450454942e-08, + "loss": 0.0041, + "step": 2035 + }, + { + "epoch": 2.9483194795807735, + "grad_norm": 0.013583734631538391, + "learning_rate": 1.93543821835529e-08, + "loss": 0.0037, + "step": 2040 + }, + { + "epoch": 2.9555475243946514, + "grad_norm": 0.046894483268260956, + "learning_rate": 1.4478085561835387e-08, + "loss": 0.0041, + "step": 2045 + }, + { + "epoch": 2.962775569208529, + "grad_norm": 0.011021456681191921, + "learning_rate": 1.0307869444406981e-08, + "loss": 0.0043, + "step": 2050 + }, + { + "epoch": 2.970003614022407, + "grad_norm": 0.008113077841699123, + "learning_rate": 6.844028709024342e-09, + "loss": 0.0053, + "step": 2055 + }, + { + "epoch": 2.977231658836285, + "grad_norm": 0.02427099458873272, + "learning_rate": 4.086808285338472e-09, + "loss": 0.0043, + "step": 2060 + }, + { + "epoch": 2.9844597036501628, + "grad_norm": 0.0245045255869627, + "learning_rate": 2.0364031375819104e-09, + "loss": 0.0034, + "step": 2065 + }, + { + "epoch": 2.9916877484640407, + "grad_norm": 0.016123216599225998, + "learning_rate": 6.929582507719801e-10, + "loss": 0.0055, + "step": 2070 + }, + { + "epoch": 2.9989157932779182, + "grad_norm": 0.04469776526093483, + "learning_rate": 5.6568620471209035e-11, + "loss": 0.0047, + "step": 2075 + }, + { + "epoch": 3.0, + "step": 2076, + "total_flos": 6.274047864041636e+18, + "train_loss": 0.03664169063040653, + "train_runtime": 42810.6565, + "train_samples_per_second": 6.203, + "train_steps_per_second": 0.048 + } + ], + "logging_steps": 5, + "max_steps": 2076, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 6.274047864041636e+18, + "train_batch_size": 8, + "trial_name": null, + "trial_params": null +} diff --git a/training_args.bin b/training_args.bin new file mode 100644 index 0000000..4417c20 --- /dev/null +++ b/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:916e08054bd8e9396d6811d6bce8ada0998bfb7dd3fe3287853319e5c2d6147c +size 7352