初始化项目,由ModelHub XC社区提供模型

Model: kriteekathapa/gpt2-poems-finetuned-v1
Source: Original Platform
This commit is contained in:
ModelHub XC
2026-04-30 23:24:19 +08:00
commit ecd909d340
25 changed files with 603262 additions and 0 deletions

View File

@@ -0,0 +1,39 @@
{
"activation_function": "gelu_new",
"architectures": [
"GPT2LMHeadModel"
],
"attn_pdrop": 0.1,
"bos_token_id": 50256,
"dtype": "float32",
"embd_pdrop": 0.1,
"eos_token_id": 50256,
"initializer_range": 0.02,
"layer_norm_epsilon": 1e-05,
"model_type": "gpt2",
"n_ctx": 1024,
"n_embd": 768,
"n_head": 12,
"n_inner": null,
"n_layer": 12,
"n_positions": 1024,
"pad_token_id": 50256,
"reorder_and_upcast_attn": false,
"resid_pdrop": 0.1,
"scale_attn_by_inverse_layer_idx": false,
"scale_attn_weights": true,
"summary_activation": null,
"summary_first_dropout": 0.1,
"summary_proj_to_labels": true,
"summary_type": "cls_index",
"summary_use_proj": true,
"task_specific_params": {
"text-generation": {
"do_sample": true,
"max_length": 50
}
},
"transformers_version": "4.57.6",
"use_cache": true,
"vocab_size": 50257
}

View File

@@ -0,0 +1,9 @@
{
"_from_model_config": true,
"bos_token_id": 50256,
"eos_token_id": [
50256
],
"pad_token_id": 50256,
"transformers_version": "4.57.6"
}

50001
last-checkpoint/merges.txt Normal file

File diff suppressed because it is too large Load Diff

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:9e0b5cc2db9b4a001a5b70d226703c2df75808bce5011883b267a5aa6d546c0a
size 497774208

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:42b2400655ba492a0eb71784f181e8d293580b53874fd2ff4446ab4be546e618
size 995644811

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:4130c62e583f950a2825a4785ed841ac1c4e66c17ce685ea7e41123493b8529e
size 14645

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:7feda8b52286297d0c53e93081b5ac62b5655e4c0de0ec0d780d1f683f882ba5
size 1383

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:d37514585ac8675c1f965ca2928dda0e939445dd6f89448a9ff0607a5ff93b80
size 1465

View File

@@ -0,0 +1,6 @@
{
"bos_token": "<|endoftext|>",
"eos_token": "<|endoftext|>",
"pad_token": "<|endoftext|>",
"unk_token": "<|endoftext|>"
}

250306
last-checkpoint/tokenizer.json Normal file

File diff suppressed because it is too large Load Diff

View File

@@ -0,0 +1,21 @@
{
"add_prefix_space": false,
"added_tokens_decoder": {
"50256": {
"content": "<|endoftext|>",
"lstrip": false,
"normalized": true,
"rstrip": false,
"single_word": false,
"special": true
}
},
"bos_token": "<|endoftext|>",
"clean_up_tokenization_spaces": false,
"eos_token": "<|endoftext|>",
"extra_special_tokens": {},
"model_max_length": 1024,
"pad_token": "<|endoftext|>",
"tokenizer_class": "GPT2Tokenizer",
"unk_token": "<|endoftext|>"
}

File diff suppressed because it is too large Load Diff

View File

@@ -0,0 +1,3 @@
version https://git-lfs.github.com/spec/v1
oid sha256:886f9280896e3681e7713e4bb50ae7a6dca4915f5b88adb69939fff4bc0f0e8e
size 5841

File diff suppressed because one or more lines are too long