From c46294125e90eca27c0b8606dc9607aea7942f2e Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Sat, 2 May 2026 14:32:44 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: gauri0508/med-record-audit-qwen2.5-3b-grpo Source: Original Platform --- .gitattributes | 39 +++ README.md | 21 ++ adapter_config.json | 50 ++++ adapter_model.safetensors | 3 + baseline_vs_trained.png | Bin 0 -> 44793 bytes chat_template.jinja | 54 ++++ config.json | 70 +++++ loss_and_kl.png | 3 + model-00001-of-00002.safetensors | 3 + model-00002-of-00002.safetensors | 3 + model.safetensors.index.json | 441 +++++++++++++++++++++++++++++++ reward_components.png | 3 + tokenizer.json | 3 + tokenizer_config.json | 202 ++++++++++++++ total_reward_curve.png | 3 + training_args.bin | 3 + 16 files changed, 901 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 adapter_config.json create mode 100644 adapter_model.safetensors create mode 100644 baseline_vs_trained.png create mode 100644 chat_template.jinja create mode 100644 config.json create mode 100644 loss_and_kl.png create mode 100644 model-00001-of-00002.safetensors create mode 100644 model-00002-of-00002.safetensors create mode 100644 model.safetensors.index.json create mode 100644 reward_components.png create mode 100644 tokenizer.json create mode 100644 tokenizer_config.json create mode 100644 total_reward_curve.png create mode 100644 training_args.bin diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..5813199 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,39 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +tokenizer.json filter=lfs diff=lfs merge=lfs -text +total_reward_curve.png filter=lfs diff=lfs merge=lfs -text +reward_components.png filter=lfs diff=lfs merge=lfs -text +loss_and_kl.png filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..0a06d01 --- /dev/null +++ b/README.md @@ -0,0 +1,21 @@ +--- +base_model: unsloth/Qwen2.5-3B-Instruct-bnb-4bit +tags: +- text-generation-inference +- transformers +- unsloth +- qwen2 +license: apache-2.0 +language: +- en +--- + +# Uploaded finetuned model + +- **Developed by:** gauri0508 +- **License:** apache-2.0 +- **Finetuned from model :** unsloth/Qwen2.5-3B-Instruct-bnb-4bit + +This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. + +[](https://github.com/unslothai/unsloth) diff --git a/adapter_config.json b/adapter_config.json new file mode 100644 index 0000000..42fef3f --- /dev/null +++ b/adapter_config.json @@ -0,0 +1,50 @@ +{ + "alora_invocation_tokens": null, + "alpha_pattern": {}, + "arrow_config": null, + "auto_mapping": { + "base_model_class": "Qwen2ForCausalLM", + "parent_library": "transformers.models.qwen2.modeling_qwen2", + "unsloth_fixed": true + }, + "base_model_name_or_path": "unsloth/Qwen2.5-3B-Instruct-bnb-4bit", + "bias": "none", + "corda_config": null, + "ensure_weight_tying": false, + "eva_config": null, + "exclude_modules": null, + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 32, + "lora_bias": false, + "lora_dropout": 0, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "peft_version": "0.18.1", + "qalora_group_size": 16, + "r": 16, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "k_proj", + "v_proj", + "o_proj", + "gate_proj", + "up_proj", + "down_proj", + "q_proj" + ], + "target_parameters": null, + "task_type": "CAUSAL_LM", + "trainable_token_indices": null, + "use_dora": false, + "use_qalora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/adapter_model.safetensors b/adapter_model.safetensors new file mode 100644 index 0000000..3bd1163 --- /dev/null +++ b/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3c5f4c335333796dfd35428d1a7de42845151cb34536379771db0afcfc74fd7 +size 119801528 diff --git a/baseline_vs_trained.png b/baseline_vs_trained.png new file mode 100644 index 0000000000000000000000000000000000000000..1e6f9d887e0fab74c93d82ea53af461141b55fc7 GIT binary patch literal 44793 zcmeFZcT|(vyFZF$)EN~;RFG;J1x3IHNC{XdD$+%yMr4#GA{`O}c0h!vAgFYZ-b4sJ z!9r0wp+sp>=`E2G0))VQHqJTgd(U0#uKUM5|NUHRmdHTzX1{ws&!;_akiM?g=D&pg z;^E=heEyueArH?VJv=;XL^rI1@7xNbyTD(H?iyFzjht_|ds(|)=h3ltzkSo${igl3 z-JaK7-Rzxj9hXs*IdOQmoxA&OHziqFr~m!|8E02pS+S?RcJLy9-acpQ#=|494gI&u zHjWB!m2=~~`sqvF(c?YqUs@P*)t2%{-sp#I-SGFGf^)kQ*v;yv{x)2@_Q~VGy*hmA z8mrQ=hXZ$8z6os0J;x+p2-|SC`R`MAwqB6j^XD6hj{Wd$&tXvMgal+dgQ2M4FPP3THfDy641|te=?uvs`bY=}UWe6=%x) zjrP1_OSc?<<{KAaR2^5jUpnK(Ba|$1CPJ2G;ihw+J>N;A$kPdCU6GGmcVnlwB{avM zzE905W4vBs`!6fJxR4xi%OR@h%Ymp%PV``zQc2QvgI>t`=jn@s z3ayr2J?+$nHz8^PEsuSQZ?Dee=UB#kbOv2L$EYMB@Fgzwxh_g^2@ zrwUkkCnQH*P&Vd$Brx!^e*@2ESnRjwWXLmATxtb}Q6?=mbK-KUZcN6R_+s_XOS4_h zsZq`p#Q=@6@psy<LTG*MnDa`H(ze~26C@G) zRpt!4)N0L(7cb`jHjQ@gv`>wp-d!&?+;1dBGmHDQq0^bn>R4oxXjV~pas|5=Kff*E zP#aw9oG|B0uq@@JO!4=71&q?M6!^zUlGmgkKT~;f&QxWAHKstPm{1Ov^glPAuADaL zQgOMKrdW0U%UAKAKiBR*)%Cag`l|_*kwhypE*Z}xE0Cufl|350<_98&no|p@qS+0# zQn?m8R#y$cOP`Y^&loaB#BNn^XS=n-G>mw)6HRR5R0&jmJ(Xvl!aK&_e&VgOwjQq$ zJ>5V%wbp7<#i{tmBjctYK0`4ySo+PX0ZnSlz7(rlB?Hu6c)zbSy{;}VwZ&hbd=ojL z%8YYo19%(a_VX(CKfTO$EJ>{5=#(=nIa`8jTJ_X0=%Yk?beL{#oX(HqW|?-$-~Rr- zT&;5d(z^HwbYt}Q-ln@XVOq<5rbn||uQ2|k?{rMBX?hf}SvUe8SnXjx64N8p5Q|Z^ zbg9}g&lKklYmErg=QNkf_RW~xtSfaNd9J;>$Fs}H{@eA)hN2j^Mn&nK!=2yV1b_J3 zF{}~_WLxqT{1PUMD!AFBUzdc+qt29Cyb+yPmCd)I>x zZ}pTWPZW0LWLPvx#Q7SI`L;+gzuFLW#RF%ppzSe6Z^w;b)<;7nCY8M@|@-#=k<=gtz zw-|Y~>L}<`EVSoN&6HY=xny<1rgKy7o7*NkqZ4VsZx`-FZ8}aVrotwHcR=!^C}A87?E%!wr^@Sm#LUj zg2Cb^Nbc`ft?%a^Dm;6WCcgN6y*>5K`Sj1UGki?z25GCL>m<>luGX;<%H2zu%!)~s zoT^);>NX;Krd3Z*+<(y}N!+$TS*-4vmY6BimreKgkCm<60wu+KS|`}h z7|R_s$*g`9NW=SzEiBL~VjD5+?h|f<^=;7^YT|i4p=wKW1&vB${Y?q&HAft3T@v6L zgH>)+%ElMMS@G=-DXhwq@f-Qp>?q;4L#Sf4rDHZ}>gX{sqjE7u8I$s``!&zRl|Ole zekfGT>-b7ZZWdE1KH=P3PR+>7dA;mEoDi_{CHu}A-p{SNN+WNmYqqF7T@I~78{mC; zod1%KeG`VA^;$wkTc+fn^u)x((PbTlo`RXl8gbd%bh{chg-yx>I|%k!m=n$?nP>qO-9{^_2L$!#Vu-!x9^UyyWRBc;il%iPIaW^Btz*l=F+px zi>UOa*&?$@(Yh{&%8xHd-h^8*mQyBkyTUP-T-Zj&vom}aW=4+Fm!)#cT}P?} zuomC_+T()J&lHi$$@rOJgHGI7cIF_tFI0KZ{6@jS!JrBg*w@tkkrsw>*{aHnVw>2Z zHY2@kT;nx-jnnetkeYBeJ)nIEVlG);jrkdf5Uu)NXzr%dN07gDod;a^1}M zxBSt7F5~wrq^^(pTAxCh^FmP^F9xRu6&QGfkFBoMFs%T|)xIk5qK0?kW$JTKi zo$9ik&3u+IfJC&TZoWHjjL-D93ak-l;z>bV87sH;EqMn__Ojv4T()cG{<*Y;7i9ACO_;o*v zFV*#C?CAZlBpuEmQvMBlT#F$Zj)86fymzal|FhVmHkx5fUOgTr_N4a9b<QX`h>>T!;`i|*Xr6#q)DIDJY5&yY zx@9=mcR*@>dEpEj;7swT_SM6=@{3<$1mKiOE&|fb+7aH)_lQ_A(_g9{$;r zq~Pfr6cS(IyF5xhR#(_{t4t;P#t|#a5?j*-u`z${7&i0FTlj8v+ZE-B0wW4`EnoYZ zNAL1%YA!!)Xo<=y9c|KhaKVFhM2gM80?fU@tXN)J-MTMYa{VsWL+OEDm%Nw^4%2@* z%9$wJ<3TtIH(;qf*H4ZvhH0){W#Qq~>towac*yWrpE&tP>0*^n|L129$2lQATZ9dH zpZ(PKA7^VEWPeWT)cg)t+!=vw z(%Ib8rG->y4mm z2VYv+(M{nn%*SRjb4oKD{E3AxG@7kc4XDvV!lgByPEa@rH4m#FF@jy|1F?w_YJK5) zJKNj=&zgvOWzyAj{|={FcF9WpssYu+zbQxWLqa?>{5}jvR>|~M$R`q5EmLA42(^UeDsZ+k!5%tPc zouTcFe=v7O1nXBfWMnIfpq!Y(hE@DADt3mhyoQjs0-?AV{8vGvokaQ{1b7fut9xr|@g2RVK= z>`&DN3hNZCzS8tf%X@8p6`%C0if?jxlD_M^y1wXriJI@Odvtq!=j*VI3C?S!T?`Mc z4%=XH?N>=>zEj2AK>cu!6s$gV@4|PtvU75ZjCSl^ZG4#wPZ=w4c)$w$rq4O?VEjc^CHni6S+JJ@8@b*F_{KCP?z z!8}UvcB%Z)BX*{&(^3K~l77Ex^QEI{whedp+7KqicCPW&{48+)L9ZWYf)sn$iI_W7 z+D7ufMAvfQbujbGvEj6l=ntn(WwdvRdor42H**y3wFA;y07n=yBr6z!b2yoe4i2OmiMr;44-a z+$NFx_$ymDUp&@tVPxJg>ZIuoiOvl628)+8{Xl-{$+71fRBNw^hhRh1?`*BPGyKhj zcd2%daZMM1by>G2h2haO}uh^jhnu^X;P=72EONK}IUmG7HCb{LM>5&c1G}nRgGV%GLx#CW zyYck&&L93N>;e8Em+@eJua-t{-6+>HSFKODcOM)r){lFLxe?9|(D!rN&Yf58_5Vpb z6M`w^DKG&!a;GVq(c|`B3{(G8Tfd|T=cJ~Pty3b=IX@H^S8)Y%G(#~0l55Z$~R|&5Tc@y{WX?6qTnP-BKu@@`!*KqKrp00ay z_P%buyQG&7{bQ~1)G|NS-tzT3ros)iC$SzLPEk{(0xY@h5{W-w?pW1RI+hcK&j4Hz zG5v7o(jfC4sghOUMytjYbJQ?az*PA9Mju<06F!REJ0_G7QT zGI*Ynz;7#8bv}GU51fk+zS~MIZoIWHJFt4D(~4&B{8uKH>#$Ixw)Cuyt&RsKx3QvF zeaqaiu~@Q!*+OxvA(Zn=bYjvF!D-+zkL07*44xC`6|bsI*B_r22u@>@e>@VFAEgY$ zwaC;q<<^md6NhXFLa(jjB=2s+lq8A`@@a_}ICdpPHh-MeeJpD+U}e2=`fDaTEu+^l zBW0>!Zm$WquIFP;62-}cUI1kNSVOd8e#_Vs{iH!z(OT@8F>_tZRc?I5gc6ODW;qWn z`jZtM`e|+9&~bkX9Pq$UHppZYSzBYzwn!ZtB*y=J=GBH<4zt^CO5S>}Rl5 zEuYw%o7bduM?fNaujFI1#yz-&7wmz=x{C#aL$|HNvAHbK)(k!8(5VADJMLoF`VUx@ z8hDir<`tFPo6sOg7o-s4_q5}GY0a4rvZ7ye^DncoV(Ox^-o&8Zu1VYzs#;4@32w{W zL%xCYd+mr=^Gf+GvXr@Ns(gm#_OC{NvOJnkuOFbE1^M{0>K^`>{XhS{6t4J${Vdei z!YdeK5 z+%H~j8q)p;vz&Q~N4pYpsK*Rc4>-K)@Ca&F@!KXQA?MH$Kzyf>K-qR!Pj&vE`(5@1 z;%5o9bygN+s##RZI7cJsR91JUtF_hH0q()svQpgj2&`Jjjy>45ZUbCZ+=v>;sq;*4 zu5_&S@+3Su9cTXkz1OtRw$*+n;5%Dg5>#ELaKKgMh~Er^-sUu&265>vj==-v;_+5;%+FvGi&|{FP+x%BqBX_u z$3p?vpjQj(dgZQf#SxZWs!n#<9x&N8Z(uhzS9lU^cqnwi}nd5e8leb1E!&H zr6XzZ{S5;t`AWUcKB}O>niIJGhjXu7ok-XEQYQ`8C{#8v1O~Kkx7lqz`pd+LrJ5vR zHlv%XGoz26{brllAx`obE3MS4V1s_F;;i^xQp*2ZXlD}X;h2HZOk%K1hMV#PNcQv0 zw9QuLpSNP&>oCnbFV9PiwL5O``dg=m9c<{gS;Ib4@KWy4L|zxgbD7x>bROW5a_mTU zX7%YgUAtQSZd5L{uBN2Ayd%odoYXONB)q0BP#+Up*D2z^I61F0>OxZpbk{z}G#RkD zt>f~1GTcblzK|tb>zE)nv4$0-pSy7W+o%n{H%39%{ZU{TchYYuioTfA>tj0I72*E1LYVG-fy9LzzW4G#)>xUoqI)Jlk~Q_< ze6x*<`P4U;?$P#=JdzSMxY2u``eQ5eqk)&(46|kYQbb9%`}6}pRPI&J6}zqVbum?q zYk!whCGcw4NkA<*l!wy2MwZgA&a-t*tp+o3n6*QVY_+{c#O)$uLSP~&On6e`q`G-h zjlT;0=+-MvwLg4Rut>VPKXz#SqCu{LE+xDwDAe`*UCC*Fr$AOs-;;%0zMbvuuxF;l z_peV37^hKYNh)(zflmB#SQ~P1sk8c@$y?Q|i348hI^WpAQy|zKULW8R9muLY|8n&D zL<68qUv(Tsy9)|#Z}(oEq&!8w=XY0G!8!%#LdWC<1CW{0y*c@6Q6Clk%L01 ze_!~b{#mqU(ek=ncWNkk_{rA0o-B5kH}*-`Kps(@%b5zAc1?Zk-9+TZ zFUayM$LSN?=+|xrxSX%CCnNqXpQqdTrJhaS)ZASG1-EGm!f_5pk<7Ex< z`v!d!I_;9?Z~24Nsf~;qhqGCDzy?y%d6nCyNfW`YZO&||D4+aVUj>33~OTEvWH5D^EcAVXhay0tttcP5ioMDKad%!m=| zB{FD`kp1rcJIdV8;O?d3iL|89Lvq!?akRM{b`x6>gs{OV&!HHvh*Hl9Gg53U27ejT z2wEQ*iFGuNH_{8sb?T-MliCc8iT3}YaFGZLmQ}B(* z&gTz(Hn$a|ah5)g2W)H(sWs?d{z*v#dw{7RppCPEA|>#H21P6oJ)61^Q6< z{#a)xSU+46#XjVl{}`wIII>58q$*%JKNUl3ESBmoQ7Tmprpa#?2-hkyDACAQPyP}n zLHm~TXmP&DKh{|k84LG`c?;8)OWg*Po3IU%M_N-OUnlx@XH1ZKzYZm(1Eva9&=5I( zKlKT_A^cNqfuKOdncR@}=Hc_znFz#K=>|-&_$<|2_VijodM6sKB zjFdpDYX*l7Tu`f=-m0dE1sT&#Hh=D*Ynxy0L0s$$?VND@TZO+1jY=vn-V7L3nJ&nEO{!?hj!CZCN?vy0Csiid+Rlsz(`?KAk`!EK%0p>Zi+d$- z^&R6ouE?CMEWUg2jS<0=QR|G2*ydTa4$Pjz}AwzhJ5TSY+- z`;K>rmh9ZHq6NdC%dE=fQ0PXN@h2049AJ@?jyWf-G@K^~3nX(NcD&o7E<5cS^m1x{ z@5aQ)d6cEd0U1hVomRtX!~0NI)rrf)U&P{)krBVxUF{~q#OCEz=GKL(E(|Z4W?8JE zEUGRm!rCLk+3FBlgO3awY}n4pw-mm zx70LfLq8Tu6k;y@i$YtXTwV7#cs=t`+D@%oqHR4R)R9| zy8(%zKazpXqAOo0;fBSM2+OlY6*{17Yl(eR)%=H8m$=X$rbk7k1DUF**!dDF`Ly2X zsc{RMJy()7&3?uxisNt}q|hQbHVV?WIf*LKa9rLr$afuvm2o%TdfpmwRu*^?~C0ms_O2~C>fe!Axu9uGc|$Ec-F*994UTc&;!^;lHC3;P zhQZHfdEjn#A93m|IFK#j(9i_2jL3{tJg;S{t^U;6by1Y3a}6_yM?dScOH?BKmWHsJ zpD!8nee5+a&*A7Pyn|)Mu{C!>KjQcDY|l(*ri)Ti@XTKFG;1i%-)d}T*YO+uW2%IO z!yTS0+o5^ZeXf~oVZ(*bpYDO1pGA%NKmQ*AboP1N_}4s=t}I;@{HUrqF9+GdJ3;@%*Ppo3#PP)%zCZ zzWw-@J@Dg(eN-{m&wK0{-6h7Yet@A(H?i%GBe!))`)a+y>2lMIS_Jk6En z`@=Gqh|MZDOGi?Siy%7TGUABkm@VNUjWGf;$$4;=tgf1Zzy@g%l38E#0UUQY_DFlpf3`{+oe0uPsfwBu39K$?ul$ zYF#;-lA1Gh$ntU>=ACE&%Xw-E1GdkJtMBjJV9v{|oa{Tiw>k^#jql=@m$~ z#ptFs+&hq&m31+WFl)N-{r(neYMkF|9BcFin>h&4CiIzkWzKv-2q9%4!aOHAYR$tS z*JF8lO)<&}1Tfa-*BZG6Fg6RkUD+g^B#0-eVA;?&lcQEQB$hijy*#>4PT_Z+&ixmT zVQ?==<@?_ST0P6o^iTq3cK*NLCJV>&t$WSIGur7C=Z*iCppdQqGn)5QMnig~V%b$( z1Bih&n!%jd3RK!pKN4+*n0-OB+W~TYHR@}ZZJg0<obj;gC%cTkmfCUJewSRK{j1rMsZ=xB;IRkq4@BL3K1=ssUSvbuj&!BteAk}e zKXv8Yr+_U@v=uF=DDpz`igo*%xsu>%&os$dBaZpi1o+o%>}TBVkAId}`Oc(){W{c& zVh9EG$Ljhyw*VcMMC&*ub~gY#i+u#CF5JxUv&(>3XTstZ{|0f@mQ~1wV2EBBqttPf zmWv&Q2uIGm-{(C6Z@lOPue&};IxU5EeGoQ4`LmdX*@>tK3)e3*$blmRqBX8KHW#-% zAGi22NjfvT&`b%!B1T%jw-V2lb6{lG(E#VhF=tKlZzLljvIk;p_V4ekLV&>@dV>z? z0BeBjyGGul&F;84BYA3w8vkzP5m+yTPd4ze&-J2Uich zg@)STg%2+-0Y#09n$lGrkBeUAjecNL|)?uDxjeF zV89slTYzS4>DYxZu`m6KNfsO4wr;t zGyhJ3Y@!Yd(8s(hwUh8S?!Wt>~74 zdq>;Ets^s&4lF_iJw~5EE<0tZqliL!Pe`n+x57`n081fD1slziCUc>>P376v)4qXS zHaH!0rO*ugUAr>l1nrpr-fb4D2T{36_1fXaL8U_mnNSxLn)YALh{E7KDblrtZOI}O z8iWC*+Hv?zl?!l=Ot{M}TH*sN^`4oMlRODSjWJ`>1dH;Qc?4c-WqX6b>Wv^StC9mE>+NXbPqfGHt%6u-AA$K^T z0)6)IlknsCAo8oBB%cKX86)YCqkr!Eub2FTL5M-^z_3<%?gXs-0D_c!kp zRDbbs-@oN^twi+gGdcJD06o6}ekn_zosL^p*>Y&14!d*=#cb^%m8&i6RA2z-kxbR| zUugZ|r_T_0BOTw;fhdhY@cHwe1;|gum3lF)#t0NhRIYfh z6YuI|eDuE#-xc@xkuC!^ML<9LuakjIq{|(24yax^XXPBI ztYHkMA+N*P(A=ux}j(D6vN#+HxtixiVVkkkhOTzffNV$6K}BhHnW?w1N-{NgHw>Qr*83|`!oYmSTc%(%cY*a zzb@JO)1%}ZOLJyW$lw&DnKK)E87QXC+mE8Wi;j;v{{6`MMC{U3B$`Dy2Mx>V%NR5bi-0h-l{!ImAJNL)f^L3&-}eN4*JZ1NaJgNI+;)^bQx58m5P9 zzq^TAbbvP;8fqkECfNe7ifb*BN8+rR6RF)b%2BnWVwuaF$AOWMF8wbQY_kE1drobk zg-9*1Z>6I4quDbA2k`=mOH$$9#c4hfeC7#!K_r$AULneepF`N?e%&yzxrJI2%JVgg z>&3j|JA7&}pEB1|`OLJMCJ__?K{)3+kP+SEH|ozJr=}AKYpVd(_nBFgGhxge0mKnV z!bUUuKY%XtV%-+8=t5vIZ9D(v%tBjN;edy=f_vu;QFLzyS>T=Y7Sr|pHEk?W+MOQ&`&Nc2?0(b{Y&G3tLxAvb z{eAMI61Sm2NJ_+^AV_4v!QgEQ274uLaiDk_+J1Q*@eUNQWI)y@)jD25H4}JW^n!Zq z^uO2extGCrBM!i5$<2#r_x?>hb{rsS?)23xxLf`d6>y+@a5HC|&t*B&EWLC9&g&rl zc0c)gq5>C6Jr#nGv3lFvwv$E_zFxgQ1MKUB12?$q7Jy zW{%k}Iv>kO!%P$Wxl^Ti(meJN_(5j~Ga;SZ1mTw-OMHgM!0uol{P zpPQBw&RbTNhQBk)pUonA;C z(9)OMEp0Q<<7L~Fl4{tHCT~y)bZk>r`+jGQgS0e%;t*a?{^pfZ02_wpC?zja3Pgmf zcn53!aEGCJV#>Salj!@ z4N|Z66C|4MKFudDohc#mos#SaD63Ab5%-VoOoZgSv~w9m=Hz~L3H@HW_h$l+K-*k$ z5tC&743@Qm7xoN#F56;Mws18WwaDZfJNIv*gMieElJCe9aECI}Bv5HW1KG)9_P~XF zbuPOT*IpBJ4^@tI5}Iv5+`z429vyI!Ty5RTwg@p)^wM=`*lEj2ELw4(RNIPJZ% z3}q}F=pnhF_E6>0fg<;42OLVYELp>Uq76FmTo$nv zC03hJYjUJj>AA7P0Ss%vS@w{mATH+;>raf?FMgo@zJwm|F^ zWERJ?LW4`4Uy3_+{Q{+Jzu*35E8}n>Wpx>yV=%wvi}72Wi@5a{Pr=3kcH0#GFpvAC zUjDl&uN9oF7+ev=rIgl;4JU|B7kaiVfyq+je%vmJ|zyUAfc}d`#r<07f31Fc)D82jQ_uG z+=h4Sb5tY)8ne0ECJrC+RvUQ}MupxHiJ~0^ZHC&(5c%1UtYXwZKNq3k4Gd6NU2xJ_ z`+XySbkPH(WPov9+BgEeQo|v;w7`8uw7k!wsmiw*A zukig0$3;VE}djg&fSWRM!a#*d$N%fZ_m(<;;i8>R)=1!44R$#F1Buy zpuXeoi>+XnO48jv{2}lL5M(MqNm^$e%Ft&Yb2qVCp&9XykT8d3kI|eH7zz;0wkX{n z4O~SOlotTij|r3EF8Zz||Mvm{= zrT+pHV@_wY#vuNx3v)deK>)o7QzN%z4(5Y1FgOo>UxjhI22p!v+*ukaERMTgxqf>E zeOeBrpBw3N|03|Vw*bzD#!#CZf%|)9n6o^Ss-}Zf1R)jg+ntb}YlVM=`%MO8`>%4K zwreM8hFm~u@CtT77M?p0zf=@;9RZ{7rL&?!kGm0t{-4;R0Ca~GINy7m@~U=@z!*eQW~M4IkWE8lB7V%9P=&E!v63K2!}{+Ko+&W70OD z$QllS2^B6b(k!!sW4>~)?(9T}O)3VC@dyxnT_m_LOvUc9)(PKy`ic3-p;}9HE=^hN zgQqtDNVhaJZjuVpqzV_L$@&EEm%v*WMolVSGb)ys+#BR*Zc#z;v%HFbbAje%k6L2$)f>^${lO3>N-)HULvZu7vkf#fGrd21 zWo9%Qfkj-}_z%XO3)L`xqn1AU#j*z><5PBGiAA_JOdNQayaOPv^gEV0n=4K7<$!kY zi)^n(`AYj*C>I&`)vjIL!lY4NAZYuU(!fRl9t05 zh<883I&zz$ApgdV=VB4N#8qAx&~e4<;Y75 zW9ejbKnGKWGRv>yNs$qM;(qhoHhmN3bWe z=F*XgCGgwRSJST}Y%p*`6$~Q{GX@Z$m$_IN4sF{xUPSjpqnGG#+={?%->&1c3$585 zA(*g?McNO-52?s)0w>(8L~{Z3)38|IsV^sBK2zjz9B5Mw9gos1(PW02rF)A8V(caL zwf~I>Iu79OL^{^VKG}2j(Hp*fnr~3N+P*JDK^KX&9<66V>1S_N`R%flp=u{(T761AIBYueK08G{wxy3n~$^mi8ZYH<4ye^ zOp|=IVcuuhM=Y|m9-J`BG8X8u(vZ`ZHI&OH1`cT_`5p%5YgjMNO-lD{=dly0Wu_0X zukZ|~l}oJF*Q_hl94lU5z)t z!T2eDT$L7KdYHR}%uYDt64D?AuNZA;=3=@W3{jtj8&OUX(7m6QdW=O}hT*W87^bfT zJvC3RhYk{^Z0rM+UqD+v;&Os?z5NQIf_h(?v^`aed$Ko`Bwt0bBIuJ+J$j~@WH zlr_93cOwaZ1kj@MOsNs&896(QsFZE!=H zM2fY3`L`Kp4Oq<$h|Tf!)-0kyDAj{o59mg}L}|Lguysf%;T<{K_Hr2w#-lvPfm_+L z|5d&YAZJE~Y1qY+!}o}`sr&5OujV)NG;r|5g;ps3!sMI{4YD9eX1vGTgn~1&5)}r= z!;J^KGhrU)=;E|Ltqi^1#Ron=fhLvQj`+7|2pIRQsG~sOQ&2qr(CsaJALHP80VUm= zJpSzn@j;2c+63#FnUPjE9yE$HxG>t{r~*?miZpA~kIg`&5GLiM^-*9EWo~nX9J35a z^=L#$8OsDPMUs)aQ1c))7MC$vDCxzYs7lZ=P?f;dM2@xR@BWuU^Mn1J)@W~_bYS&3 zN`*H1Bv#{Ws&;3Qz*y9y#L78#_}-v;r*qcm{%b1U@nGh)QiTif}o!tW?)e?+weafOD`rIoqB1yB_1W(c(& zz02?%5wE-UEjASv60x4!OHmX7J^Nvv3DZytEJjb6LnrwDuaGb&h8s?pfi`raIBUab zQnkU%YdVB4hN23gr4O|D^recnYc2k<^2Sf5zN26%*H^A5sgecwlbf;1I7)lw7O{5} zsO)u_I!kgdK5+cF((an%aXg)4#dspZ4szIiy`>{f@R&{YlN(NM=ClJaySWNfl-HLn zJ=zPvWs*LE@yp=5cH;Y^%Enf~qlaYCyS@MT6~sa}=V`&PW2@%`)sD%aVQR6rCu(mf zq{>*9zJYoAqq(vUwKaEgTVe$-$tj6!5j?Wn31VU+u#0j^_3DLpXzcxm=!1s(k#I9O zi=+w`@8=7`n|tB_?)f6TL{lRBa*AF0ge*FsLPw)Op~}76GoiVR9|7XJXv7q`3z>d~ zD~7tLar(RP3vT9<$?qobKa%HR6Zo!<0eDe(kja*4jKcCkM6?w%h1I?Z_O~m?QZhn2cA|*&NgH9PWL@CIk33O!Im%LL? z)!jJTnhha@`vJVV;un{8lRUX5!*K^Fsz=qt(kO9Sn!=nZbh5_$taf6SZI*XIki~4f zrJ?W%807Zt$+vAXOJ$F?OJVh*> zeDMLXUvR*s3nqA)4N&rQz7sD6Gcs)T`1HN!mf>?x(Jjk7v-8*AzIeZ|JkI3OrG6py zWUMJxkDU@*U>_#e12K-6T$H;vOO4cOnt+maziy>=qYh;eI=CWbf=KJCH(yKvb$@CC zF&4(3ThPVxOvtzs`8?0IS?21jgm4s!vyj#`ucWyt$H#MDqn-N5tEu>mm7i^9)Yer> zmRSEsDra4!G4C*XU<$ms^D;!Yc9xJ&akj5=@et;3%_{EoC-^VhSdXp9PNZ6 zuNtNO{*u6T>yJ#7rgz1K3)YeqH0Yocefz+m{{7{~)6QC0^D^Escy8c+5;LFV{M6WuKTHGWieDB4Yz zqQ2`D*z4J&vw|utq-UXHzhckHl#Lt8O(5AH*GmiDSNHrGv!;PH7sV0-;^kX=~I9Kq!ob=dVNv`8~=aW2rbGjj+fpQOpE>_e_^n@qyg;P*%31A>KJZD4| z(%zcV67s*0RUiYf43V8=r26+jff|7v!$n{zZYypAO48r(S3&o>dD~v`;M#50pnOR8 zh}i8MASlnW26=m=AaIm!N;@V0ACjK_9Ywdn{`)ht6(t6G>>DR?U2m1LD|@K$s0$++ zctNF#1;_qV^Vvh+DiT%vs@IX;D$XX^v9y@DEkH=J@U|0>qN$1!|hD+{Zek zStG3y>lt~u3jbz?o__i)S;iNcb}svn+{9^+8S-hjbAdpdib_k+n-0c>1ou5rgHOLe zQrG_zQK&nYGf(6mGj>n^yk@R>aUu7s8ky5kb`RNMcB=tjWM5uV?CUl(8~aEy)Ud@I zer*I5)=Fi#N+qLg<;r%(pcj!{%^sfQMeEw>;V`nIYVn-H{~N{mf5}Mx^{)#5e|`kk z?mx1l|7Ts$a{NCtwf|>b{y(palal)~*g{l@)x>pNu?a!r7ArCjilJj=p!-3aGO4@4Pvz3M%0WeTAo{DC)$iVo)A%ofae}s-+QwJ0sK!s#SYuZ2+A#rFCbPIV%BP)4~hRZ{P zx+lK!B0LiYQJZ?iK|ep?G1^uznvpqZcuZ~rLSpj^5$3n-Z=0g$G{6Thqr&rVT0wU0 zK0L7kzT<7|!UnEG{;kAURF39oL?Lt+yKDyT)(j{SDmO$N(E}ug^eaFcM2}z?f&ADF zvm39j|K+!KWjzYCoQ}EsLsVgkbwMio5{f|c|6lEWc{J7gzxS5YsNtlj6q+P+3Pq*{ zg$x-sGL}?^qyd>}4uxbWG9=?Rm!V{6pi;?9kvW;?neeB+;xN5GOOCJdNQL`LoauloB~8Zi37_6BkFuSuhVev|R)cn8 zXjV=ACtT`|%NxSr6Sqc>OrPIgx58+hTmQRfoRfb*K$dERR#nd{&l}6>YwZF+602Z2$H+`u+Qk=YuPcIk^hJ~&AsaDnuSoLF_&K9P;_ zGrq@Aqry75ehUZvldhq*=N7Q)MMeea@+v>rv5CY(F)8Jd&Xr8jXc!5B-&LK9j{Jp- zqjKxhovr8p@vZgXB%UPL!>5JwZ4nMtdGZuCLwLIGd@P`$)Q#o87v=r=Rjy%orY8&?&KOX2}2_|?~??C zFXaSi56py6EfPs;BE_@$t5hQaF47@ah#;P){N*@cVa}2P=Fy@w-+Oa5lqoYT`SbnLJ>r?uR_ByL zB-YG4*)&u9{oAhcX%LV%XKn0g?dqjvMP}8CmyVeIy3$gP7IJ9HUhnANgzOYu?SPjb z6x4=ia#2bL-J%z6I4@g+!ZR2ii4vVWB>n7kk)pTMs1A`$l`q@<-G{^mr6EZ(24?tE zYI7@+lK5`(ibx5p#$h6h$pJ>%0Sm4ozSJ@YKvu`fG$W^hf(aMMyc`^mk;a-%7wuC@ zX#5vd#{eWqRa1>LFEqpkx}cpE`mL=DHyF0IQf#c>@QJ?e-Ne?RE}156JPQkprzY>VUb^pw(o?kgRxdfO5P_-Y04Z%bnE&$Fse`h4Fxf?t zm^;rbUHS`{x;~yF;6X2F!(kNdNNBi3CtWah8dOl}HVWCP4f3M<$J$@aAj5D??PmJ7 zwVO)S5nfva+nSsQH?kVY`XeF$IRk;Yx0YApOjyLonr=dJud(r`|L+p%Pjk8n$`C@L z$p4bnTT}-;s-2uEVr{(iuL?hTy_ZZ)gfG98M1LO?-DRz(=Kk{E*opoRt&ecIN{^6r9+6;?NW@{FI?>RnS?OV$ z#5-gjP98%p&>pGr#%mpL6(|C9R?I>vDaYpZG?SwL;Iz#ltRr6dGP}ulyL0bd8zC|r zDG0&Lka;L>;tD0o5Z8=Xtua9)0^06QNutUWNYCjbOg~B0J*jqpOz+}~QDw1jkzXbp zM!O#pfteF1I_EM-OF~p|tdVRO!A@9x@K3n&2tA*w3}&)c2gb1=VENC6{ga)z#(e4P z%5C)Yy#r$eJLiyc;2)kQ96%ov{oT@6oY7=iU>T6l5JLdBhZ@=lL5hl{noCE;C~3Pr zH)L_bb<+XZSe1}i64(^wfOGZQvf-amA!Kklsqch9kc^#M3HUBapH5)oVML+0A(|6V z`9v9V*$AQA`dZ__^+6J8Gui)k#n5b&I#2N!7>StYhv{#NVku?4f|zpX6>{B~NAy{F z80GB85oOZTVPP21al3P4}8n+DjPYT-x?F%mptKTjUo2M%L&XqVYKk+3ON9 zCrP8ky1Gxc06Ex5Ca7ehCT|oWG}%E(h`Yhi<#@q+ZhWmVZs*v%V-{yekPvZEWPz|K zhd<8u-QD$p*ve0w785Nskv$PeMDWNn01qscF7zj6%K6yQW@V)IigNm<1DZ~N7N%@oy>F9_QYJeHhUq8_1wDt{97&lUlRajMV{_QryqlbuuU1~FQKI@b0>Zl zrOuUH(>$p1(8re@DMN9H2TTJAP!d{9H1Rw<^(RL(|?1KT-3ul4yl@C zd0qNHhKfeJnq^n8M5olMK&uy&f8@*sD%!9ybNPt-`hk~jc+c4*qWa%g8 z{M#!@5#Tld{n|78$H<{g^adTshm5_POLQj~O5$QGe-~t?x-IgYc(&Ju01QPU%Prw= ze?|HX5r|$Ew0Sa%0Nv&L>xIp&dCd0_QF;~t({lWbdqh2>+hz}MVHpADM!_gWXwf%x z;(oXg3DbCF_mete!RREe^4*Lw9e9wDFGk0X9jo}O zw%_P&nD}xTrat=juv1Swoj_!CA(I%jx>4 z&ZWEx~Po#P<5w=R%y)(5#{!(t#19qFmvZ2+oFE|Hb~ z5b!I_UhuT9@o260Z-fCkYb!>+-Q*&Y*vPnn7`*L>`&PUvHZ2 zMy34(7H|j$)V~mltcfDZ^a?eLCR-E|ZxV@Roeg z2~9<5h3#DDKDD@x6YvqvhZrLAo*x|o!4r`}%98|ec%_3cCXZutMx(RYG3G{ys^FS) zb3an%G2t%Dpq#o75u;2)4?(TS)1}46%FhnT%-%ob%@INj#8r?E(zp32orBuQ;v5usZS zK*^8H{eK%U z*QPR&0Lh^f{+BT2hM%0Uvl9D=vTiR?hNR^)ANj#CMJnqiOZqAZpVs?P;!LD5gndlv zazXTKBmll$IDquT6LRrnhGc1?2IpVTpK!p^II9~!o2*nsM8;$$AXI1?(14#QGF3U? zxLe&^)uu_pKMJ_{#r5>v+^)l_^lr`m;hhaQL@p{k+%B-{@mOZ=NFt%LDkNNy(N|7# z9AMSo8x<|l>Zd`!mx8{&;8Kero`u?CFm>O%q#d$lFbFZ_e8*GU(;#$pjI^c7=eX(1 zg#GaS{bMP3ByC4d9!wtI{Q2*CfWJ`8$=ZgQfWOoLk%aR@Uj@LU96eh|E+SgV2G*WC z`d&5j2$lQN!~6d=j+r&NG5PMwoc<>&zfg6wb@b+b)jfc(W>7m?yRsa<0=&Wi_$1IC z^H&T}K_Ia|Pm#>I`vm45{uZ-~_74O8WX!mUrXjAXqqP)qymKkJZpUSyT*q0gK*HBGdnhF-|Y@-29Byb)|?`yM{`}Ea5fJy%q3p@5kc;sB!i7UIiax=ZAc?JT_e&dP^B6Pj=sN!PKe7+ z5&r((mcu{0P_(Luyn^V((z8jKH8Nk-GzINL+K1n}tcAeZ8DEV(>8gB)OMirRilS86 zOrY^89T7YODjN-1rGxuE>RU2X(`xBUI=t>iZ(Wwnkz4GxIx|yXt`liZ{|;+C3e^ku zvh`y8?3X?Z3zVIve{W5m*fGOx8-T)hK#i+R6dYyfa6*wF;!iYRL~vSxR@55S)X%$T ziQ|cY&)ZZ(zZ?ICZ=EgfOVu5KQ|{v82N+HTN1(GKVH4uk#G!LZc?^edhV`nswST98IsFVM5&kXHm9_Y7kNtmzuPlp-osZv=odmA@ zJEIQ8Z2__3xow;OznNbcDQOpp%-dz3f&ncP|*fR_R-*d1HV8cx^g12brp% z5-Vf=zEJ;jKfdVgUSjDj*$jUzRIC`XDN9amq>^9ypLt3DqZYy3#wkf58ff4|0p||g zDnSQWfpH?>8~~z{*JR_co@|sxijxRhm4I`Kw|2lY1xL=o*8p`=bcmyY^?rL3$CW8a z6aBNS2ZR&E(a7#5_bk{4lh78og2GHG{F{H?7|4nY>s@#rrVAFR(@)uduf_!R@{8e< zvfr^QNE?{e(zSHZQs7|w7h|UH3*oOt)LI1*6G6H80L5~<9r*$`){WL@Qg0Kzj|>cO zaYxp?G`RE-CvXjF;?O879ix8L%sp+A>hUTE92!1(^3k}RKJlOUkp3g0Ij;26IAC=W z&E`J}=#;JoBFZKlbCCdUw<43_m>KT-D4L{)SWrx5vLeOIk@i81{YOYhrY<}a^efm% zwv07V6qDJuK^oeh1%a1{DFIJdlmJB}qwQ7?9n~OHHhJ)bEVTyqK%_&00q^P0R<^eC zv}0D-#g9DVIt^!^6_G)Lo=G;0K`EL{0pcRw9DR37f!Ga`;Dd#uAe)uR4RLhW01*Wf zn3RmK*@EO*QA`6LPodwSo0(^&F^FjZO^lCbhZEB5H~ytqEIFif?UTgM2Peyjxp+g0 zb|zuaNciGJAKpS%|9kpt$xfByklsGM^LP6eji%P>N(b@4Y_sQ#!kiG z|0gxNLOhD-zUXDEqcs31l%qlLl6nTWswZgsih4-Gmdr$<>czgB7$3JLLOgyqw{Lu) zrK^Xb0clkgD`Af0&Vh6x?yxrpVVDK#9@*xF1nQj=ueBGp>|!X_=&X35tkr40>gN~1 z)*k&J4Mi3%sk&edvvWLmT~9a#OSteu`Doeu_uqSM!>TaHO^G&5x-Qd|pB4S1hTHdf zM`O~+DB2*C&8^Pa=kDn}mS1qvuvKzOq$nwlyS~<>o#^77rjg}RE#VBdeIg6#O-wo6 zcBO~J9}?1nWO9*+3@c*`28wB+P@afTt=#}Nb%ZC>TH&Jq9<6N{+nS5`B{TGDA4d9B z_B36$#-O*R1mu}+`NkDjSR6rCgcFe^l*_F#@LPCvuKX@9EtmQ)ccGu& zYmPxFDM8}+><`wEX`Y1frziD+;IOEAKZhyQR-ZmF27)OP@aIOl9}zf^A{3G*b%BGH z6DEjEsf7L!%}P_UPf65KG58&;llEzF=PANGz*Bv{4C7hA>7pQi$5cX3r!VLw<1j2J zOxN26T))?s(0>(5jm>XNG|`qX^3WT*v7&@j0VUxbpU*cqJACUUeAhbhCKa_s_CS#awQ1(C*LmnOmH)!AE)h z)lv3HU4#8v4&oyNK59Uj=fa!^NP*IbZZZR9iXin!5$t@@i8Gze%*Pe9y zxdQ^`fOGp_o>`AVXX)Ozqo`PHL5FoGM!%#R)3aN*$c?cm9oqO^`HntxX-{_8f8O}~o7uNF^w0p=!7 z$06k|pptA-4%Ra~RdD$lHg2J}gyB-=|4#Gx-PmA`G-B5a<`)`53)95~6)V94{Q7hyHrPl_jVPVQYA3pL4o8{n-JM*Pot+paJ-eeeOMJRP zT54eY9s4!2ps!Up!fb=ZRbU&Br`1{B7<*fQ&d>M<+Qa_P=Q{SPXI<)7ts(8uJ+wn_ z@mG~q^{YGwq3HQ>@drDvHAuO?1d5-wT!g8`GF32MI`!ncRh8&jE7at~HwK0EWZNl0 zYiG2g`=tpe+UaL^*MvoB1glp&Ubbl4lQaIc%JUXp{)3pIQ#Q9Ct5>TNcXPN|xh{UcDER%v{;{gFbf@nh$A!1g(h4j%<`k2^X5sd=%=+Ji!k(f-O(m)6 zg(c1i9GRH>7qPH^|J#2P3!BAH|A-_q{?;#jWYi?rTmMVdpvvAJKO$%)*(*f;!M|hP zFNh+27s?PfK9kd5I@6Gv&i2&50dp_^BVcYwz=DKGK|y8}L>4#hX)FUr6(3D!waKi- zw(GPKvVUid?a$KRY?=WJY!rSq;e@}TCy7G1Bff|8w{>JDkFXEKZHnZ;#GqM70OApd`TVU-BL!aTDZzuxW(oe92_`PXCBZtEK7~ z+-8$=_Fl<&F*75RJLR0m!*k%F?*g{f>h6j>yB_jtyj$R*C?BD@db4lpL9-oOnOT^3 zn%v&bQ_63${h-KNxJRojW~--u$#^;XXH|=jIY?JYbbWV>cC@lMl_xWm)8I&N0^LmR zpWp)qF!ztV*w%mkdCULBKW`|dUt3<92VKSD%!{ob56(hovh{|5UdGl*ZHmugAM1tZ z^VbO*U-uU>svqsGvmm~>Ziuqvhm-%H$nMAe@YRTB*MEr=gKi9P1MhEpQrxzb!hU1L zrRbM}D4SZ)JIvEvln;&aS&Vy(6AWHfL0)zSQZi2gos?HZ;C*mb5kEB_D2L>7t$3{> zWWDj#rkuE@{Af>U*i_hT6Eg$l{Ex5C&Uh2cr7eWzlM9By$ek_p<#L}%rd3P#+z!5@V!-XmSr#Z}2heTDODK2s=Okw4ym1Ud=AHCbDugKKh6#a3twHZf-rtv_^2aTCHa}m-^jR}E&tbLke< z8Tv0S9)k#`2dBt0h5OsS0d{Ljrm}$7es~P9N)PTN95YVfrP6eo-rVQ`9FPf2{L7gu zB;Uk+hK0f8=;3LF#Y}EWyc_Onr`Ln*=mtHUGH?Ep{7|G=#E=-^+~<&6Wd0JKzAx{~ zK=E1E{vexHbM0Ee9`&R>oJ#3f#o#`3D~qk=-xG3)IGE5V(^8 zAwn&U8czZ-UD6CId((m#C??i46Ji|9R_}vGVNB6BWZ2hqv()4W=zKG%qcebtR72?1 zvKZLcPKl8-v#{iKL!+L#goRfJ>uK!B(}f=HSt=FC0yiU9M@+Vu7e*WQQO<8CA3{IY z>tdPBc_-g5pEZ2{6N7B;wI&i7yE=Y&fAcg*(+r?&I`=lqokhGz?w?O+D=0o{aA_9n zrkl4;fAc@D?P^<)&CI2!^WKX+5y8E2x!UWCJv<#;2bv3dAc`6pTe*gTvaU2j&Qmg) zaQQBG$FZHJS`@=qu!(%|VBxcU<@m-XLdCNVbVaUvqQ#LAUn0<7)vrzY5UV4__@Tu5 zdot>oEiN{;&l|nMxKKAvJxWq7oR+JK=#YYg`(mA}Dq`9}fxLPXO?3U<=TJj3o$H_y(5OPglC{X(u ziigP@a=NNRz;b_lRQuKWOlN0)O=zVW%Yha1B5EAz?bHZim1_5zvs#W~27EMLLq@HU znUY+;8Y;zgp?@!BX97g0Bs1bPv|FiU564%^u`IQ$%@}b>JM}4GL~)$RY$|zg7I?0y zWzp6@0Zbzkl9Prn;%yuH7+gOKzu(%JJde^?a@}qe+_!za2=(U(ysV?y%@H?W!%1u@ zEkeDZnkpy?BKIl$x8Fed>3B!^^Jz;fxk9x%tFk9#1Z$xdZ0Q*!+}XgBuN`Z>(C37iZ}!b6w9TrXoa_u0%7+)i zo!aByi_MvaOp~t&4*exn>iq6#Ptr`E;=$PWFIi7TT%W=vww?ZilS&kdFZK2r#pomW z=%)=YD6L+jmvOekvbw@UnZ@-nm!6_X-@!shu5E_Ugney+BRdsD)nv71E+fTPZBOej zG~E-^=+(Q&9;Yik-hbm)+AvsjOB}(EUD=(D{-o7(RmQhf4#4A^0hsK4X!RIOB$7(5 zEN*vhNA7i7dgSf-H)-~zD`o4_WIrPFYlWLe{5pY~Z|3o&n+*X$OU^q?y(%FQD;LB& zMbl31bqDw3OZ|FOdLVz*$+rbuM*{e#xIXxw)!uI?a5}>Mnw`fj#hv;P@QB%TYBu<2 zSp=c@lww<*L>p$@zK620N!{XmisRV1yz2B5x9^rlX7Sbnr|veg*k%9i=_!|xwRK0A zCu{o}TAm)}u#~nKbzhQUUP(KJB7d9ZGJ~w7Z!g7HH@3`EFkye?YxFT~Yo<|0(y(UO zQ)JaYt)=3ojD3r5`b2uqN=hXcXzuZ7NHf#m(@uJ%U1*`RobeKcQYc$^&ft5OpLY0z zwPJn;2Ae>__fPpq9!;pYF$29+5+4d9ud{B9$gu6#1`5lKkVoN;}19E8tjp zW4H1T;M#>mr-?!77xXmWaJKznF(=1rGp#alc8{NgkTbEhIILmePpq=s1(c0ynyE25 zu?L@@_S>!&J{~(NRQge>KvIpEId{e7eLB!Ct zcz$|-+M9g?TGLj6LPotZ*X|vqxE)l}FE304v0CPtkhV$EY5aWN&HFyrt-hu_Zm<}w z#-{8W*qNW7axAg@%fw*oP;M{rMs*J{hOE`NQ(^OX!JV#W>EemT0?P!5wPpEK!-|b9 z8#|V7>}Zb-Taeu`&xmP5UHqi}?sS*MD6SV(C5b&JRZjg+R`nZ9Yerdx+3rYS2YeNo zjD?%Gh_4oR-8U89gj%HnYP$8w{hnIW=oy}qSp86)qx_PX@e;rhbJT{q&e(fvl)5}+ z6EsZJ74^19p{x_)xOs92EAE+AqVB{=@_8`$J0ln{M0>cS{TDv51l&wyc)}IO!8>DP z^TZkoMPAjM@yW7wK+O&bdz-GW`jtOKl9BMS46Q0L=h>2_k_~rAH;%761s_w?*pn8c z=yV3x?8FE=ZMa48hlQK@wFS#0w{@;gZ;o6Lz*Ox{S$-4SFfxL+S2<97Kgl0797uXE z!cVR2CBpjzkA0qzgxUy*t)90%-Fysfw&O)-sD`ElB1Pm?4!4T>Q5Bq>46Qz#JYtC4sFy6Juc(p()GwR62@B?6e|BORt zLCguO(j}74(Ye~K*XmczzKAH$#brSWb_2~bIjm~B*VPIaQslR?8)*B7Hs2GwVUuht zB+;9c+e>azA#KK)N3l18#Gcq(H>5fOLKs`tWy<9g9+~BppRAMm29d2T!aMS1s5xCo zC+&z_eILmoTd#@**|25F-`ng`c5J^=#kN%4-0h=rK8i7KD{gA;%1Sd!TpQ;|{qSC6 zLq%NMpO@c?hKmLJYgU^j@A!03Sgks6>8geOms1=T_^eKQUMjmg-CG=7kJCObno`M| zL(?Xrv4^M&XiB6L%x2yD1b%+q{h z&~WQI{wJ!it6S)zp9C3hQTz{=D{so0eWHZ$UjwM;|z`Scv*= z;U;O`!WjT14%r&>q!%e%_c^~JxMV8^9QSCkoxEARx@obl$)h_rZR^zSh3zs!JxXz+ z!<6NM`{@Tr75I9KMiU5DxpQNjHEEUTB5H7zVwpv>!t+B*sof`X1AY5FZsE0l86<== zY@IJ~F70;N1u6DEjVMF;fLl)n9P}0DONpItl-0{|NY6Xu-_@xaFXzi!M=cW6U45g# zx90tJh6|wP-npE1GB#*vU!H1GItqbZo%_#2Oi>E98gnpxN|9t98+nzz1E%kZ)QNEhzk4*}3J&m$v zcvQv?4@U{E#7)gOVh6;z*!FYe^lEsDQA@j~v$T{t*wjq97#ok?!+7nXo89{k#PbZJLmCTpB+1U*(~9pb_C;~Hbe@S{a;v8{(@{o``l~~}30}4z{(}^X z$sg!cb2Gh4`R}6x{l>|(_~wp3R4Sc&YSM?^i!72nYP^q8%|ESLay8dxYxM-+Jyp@l z>s6a~Th%|}wWY}347ZDU5X))`J6&{@Djhn@Ifi>`t51gWP6O*Xhf16Xlj2*xKfr}Kzq*4x|@CEP9_C>okiyTas^JKj> zA%rHdL8Gh@pw!66Tzv62Vk~LyN=r?+e$}C>bFH(VLvDP8YS`l+PQ$z2yn-CvH$==w zls{W1)i^gUA+%?}&s4jswDnVt$ns=`>wTQSX-U5@EKYGnpeW^XGdeV)_%?Z>5K%*>-aTzw1_ z`BBzgm>m00G@Hk3CFzU9u2^m?th##3h(sOE_z|}UF)ia+1=>qHw@Vk!r^r5Ks=g|b zDvHrj@A>)1Fhdl60V8I$Y?X4r8eYv0$lv2vecq+3rZCq3C~2Ljn|`M02Ci-M(HA&wJ4P3-08NZ{DL8`eDiAU_FYg(E(#P1 zE9NPUON4N&2D7(AisLo_x4dryN&Vq!_vDJ#+%)SWii&3@u14sw3ocjtIADa6pIJ-)hKd#^o?-Wj^mbYp)*N;|MsfHzP8I9OIcRqh-4DL?2TR_{opKGfejmRwpC@@U)l@fP>$SZ%%v;DxAFGI zwbP(O2cRLKmu7lz%+4tDoI!%nOKELtzqypgi;Jz+lNOZ5{pj`NSHo)V@WoCMF*Qk z%uAL?Sn;q@BA4XUn_E`Ws2L~TE(p7Ekz^Ns>q9f3yGW|~NT8tQ-SxXOAg>sA5)3&n zjn-vmlv4huDuK@$&eo?}1m-}7Z42+Ly;6l}8;H-g)ThvF+m*=9M9$N{TmlvjCc*h`|9 z{XmJFdY5``_U90q=!L=3&RGPckpE?(&N_U`CTN?D}wCSLxtBK^wAY8Gp`>ShZgxcQ^4^^az5AF(|n&r8|3?)us_ z%>lil_f6hhUtF)P#nLM=m?WJB|BpYUs3&K6SNia23(kt8gRT|Y=Btku#PiSl6|Bm^ zB?M^twpHSB)bprKu`o~!e;hj?5b2RS1HJK3AN3JSjtbfRTW99a>-WBK{iWC7e8rvD zIWjP#4LedrOq+&OS*3Uca^EnpGXwTIsG+3ZyfOdzpP3>blob{fZCR`^KjhcP0D)ja zRg0n1b``agl$K`3sp=k}zGZ&Ms5t-xqbS2z6O!)s%=^wG|*89Bb<+S`uNMU)Ti zY3>`6%J=KnX4%9`LDuRbVSx;-dWSSLL7~=gn?I$}Gz*~`Itd7Az~WHS5HKiATfYQ& zsLdUA;a;EObsyuP$S^J4Y{V}&4IFF~X5g>DFK5-F9Cf)mQL_0Ya=12$hp*d)5hTNl zeQym8c@q|1%@o+Ci|~V5i~H(c?q+2KD0B{o{fo??brROGGNYI8_vRl&317mUw7S;1 zjBhD3$%;zABpZ5Mdjpejn$p9a^%s>b%aHGNDy+$7``45OKHWf@#`OwEQ*Nts?bkVI z00JL!77>_5r<{EBu(c8|_tmG4(aNStOI%#?XT^b)%uJgU24Ito{MPUXcHBy-3}j2_ zuLl4Ph~vx2aEB@mNNG9NXea1270xf(bI-?g?Vj`D4e^}@$r=fJb@oebWkq8v{OiD7 z7 zVv}^!KThS(1_xMDd^qr%$W!->3ZgI%nqP;o{)^*a$9T4B>XR*lof`4UqRzPx6{ecV z0<>^?&B7|@N{w2>s{on~ zPUgaDyF;>EFsUcNp&@;n`oL=5#)p8*eRu^KFz7g<< zEgq7?{A{mfqkRnx_ZN4bas0>=Gu~_ptZb&{TKWTbql0mMVbU?^bVfnaj2AOd8qY7O z>`;>soCUP#u;TJEd1qWU^vgO>msW`2^^?lMVk1AY9jbWOt*p~w<@yP5+tk~}&#eF2VbtW^*UOoeRDp zMI+njS6obJP_HZWjXdeT?irWVify^puFgvnPu%+adeI54w?N1oUZN6Mf?KPL!4w^m zF{`6d`KBPfkrJCH`*dP<84troFNb$tu)uo1wqRJ@b7H+J6E_!=wbM}+SKU2vnU=Bj zr2#R_UD3a_>d|ae}~pyBnO!Yg6M!l(sjPx$mI9 zBbVFFWRVUco9Rul2Ge>=V-mxYyVx+-$=)h6TFVNp29Qm!4yppD((F98 zeYY95g2Xq}hun%c-R|5TwuZ*4$q;*^6gRZAdvbBDsBjdEnsL%S2mtmIA8VRj*?=sU zFc8o~Z_w8r^{|j&n+0HL&t(+8)v0|`e7>FOfmO|qd{@wVUqbbmW8Mea-$N={FLs_2 z8W_oY9L(!!PVoFjikg+>^EKIs7!^9Rzu<0KPT8Zsy4uDZ^hz%?%-xUlrW_JK+#wrxpEBrTW3*N_n;b!~KgXV!SNf&kaYmW~{nhqkW^d>n=Y z9739mH<$GxI-rnM$gn$g0B4j^;4I{fkz3gq{eM&~NVyrwtJN8#a!sAgiQJFgrav(b zXKtxETlFqA)bmSMSvBvQveW5esf?J33u@rjU~%Nv_xuQ2q)kp=Vt!qs-tNA2D=6#c zD_m||v??d1db8rS?-dqWOzWv>@k<^bbZ1Vz;lx;jW_qo;aGTop?x&h0lgT(9BHsf|sp*pdx-fi7}lyjVQ`)P9y zcWLdzKU&O^6f(|!lN5H{k$-?98^JRZw_Q{Dy(8hv(!M1arbtw>F!w&5iKNDBxg@VTenLwjO-#=6gr9Y_{wl65njLFm=igDe zN*es#v}^q?=}*+&7`WfKR{8SJ6%3RP9_a;9tM9MgORhG^`&qv2nQ=68FNXiB9mvob zu8v#c0z@ME?lZ;MTMT;!{2pn~zs9OZbE-Dx4|RARzvdSV>)p;Z;>?RsJ3?J1 z&G_`6r)VcFRAjsUVw2+5v)`W!+o~nIbkhvApUS>f=3<)_DM>@A&7DY4?!HLnD9Qo_S%p%B7$zaVIp$?2_iSCqE%N2o{q@q$)`#30VSY{~QQ-3&3` zNb_34E1q;h{PF%xs9)E(w^sjvSzZKJWPa$gS<(&H$jFDbTc)wi|NN_ zUMQwH^ip4v8ARic6pjKCp2B9S9yz%kx3;V+&@M0-aa!^A6GNkXex8f$wupS$_9ud5 z&O6zU@nMK}RF=i5>A@5r^X~V}qNfbswp3J4ql^z$$1xi+s-*9!g#rLB!nadd@j!tks%_fze};A%b>D9Vn4qc zKlM{7`ooDcX1DE3_bs6u-t5~;esi~TYz`o0x%%QS;{dIlzefK!>Esk`xLDHU%epj6 z!?kvS!5^LP=hYbB=a8iTBKzq+>@1%@mwBPE2NCMG{+ZU{rSI>oJBbG^W9qFta(6|c zEo}pZa_gG0uh;wdQ%bPxNEzlJ>>{DTbSiW5NedEb*ny`NO-dGY|am&EJa}z6XX)YPzU`3stz*s+i z=qYavRfJ``iwhzm3@7Tt_c3}~(fqV6)VT&()>qy<6kHpz)((JE4epH19$}+wDE#2Y z@bEh89>sI<9@B`R9m{`7W8-k@-^XW|{d|R&#?^fwI*pS7HiSx+q{>pKJA~D)8h;89 zFl0|Ym`(d8NA2ybjgjtUTe60k$+hxHfXQC;$48aaC6@>*1Q(m%ioS2;aZQ3*$iZ;2 zJvAeKrvYQ0+yzbd#RZ37VC;~5!Li_Wm-Pd~WM$K3)XFn_9EWb~YeqxNShQQMOuw>| zki22$Rc7tb!8z_#OP@z&d6~rSZ=yXVQXs8Xs%5Ii>w({}TDkCY!P{~3F*l7b6ff(; z(+_p+dEJ&C&N-E?qh7Yp*$jChEsL$VN<{k!!5`{Vz5htgwDp~%JtqP6t&4YA!=7G! z&nf4SKa(QAsj0cHK((5KKfgxajpWM5c5(I3O0P6v z-0Tlc_Xw7_igaCvU{Q+z?fy8nquj4*gsuW%f(V4;g|I0zO=dfI5XH%BSbREl>rcYFng@I6~ z@Uq23gCYuE;*4X5;f-SwJvM-;z9HbCaq(r{c!pOq=w5d&9toxvTy^M;(+*>1j4*sg z_z>LPVxKH6;A0O$YF;~PD#nONTj8166Xzv@s`^{goZH}q-^jdG*S6_!{{EV*_ z^01`(hS0tp@!Njl*i(~%PV01eCXr`95DhSTkmOpPM&d*UlqHjw*cQYTb-*X23x#Ai z2Io=3^f<0RhmV~gu~X~ZRovo=a1nvMvD~pz3n^zXTaj5}DK>-0Vxnz)l@iSmgmMOY zjWJe>rV)Mad9okJt5<&5s(Ckt#yE1)syzCxkeApVr%^iZ#+(zrXEY&#I0(>zU!A5O?}uPbCqdK;txg-SPGbheY2Wovkjgi`#t; z-g(M{Gj2_~M`?7LiZhRJ@=ThqZUF}&zYaxp?W4w8{0uZU9M)AGP&bNK4qgyRLRIPy z9SSZY*CQA}GNA?Z6%cvNu51p9&(4_}f9zbmEKkiyYMr#xa~2`P@3rr@4;a3*wM?#I z@;h!-lNJXe{d2=*_{khD8ouxG-rKu}RgD6AmdT}HE7uzmfN8W}a%;8dcItuGHqA%; z?Ka)KQO$cIyi<~^sIt}|xVZFjoZ`oLIlXr1vSd#qBZ6C~1bwp~Y<|_88sC`))TO?Q z4*dXBG{DddR<^cp8oDFVtvx1`#>Nk^3=cn`QDk55@#M!9-YU$y?2a>JiZ{$)^CJ1$ zdVouok1r)}%48*PjsJG-^$!4s>tH8Ky^B*{a4PPm0+ao$pJeac3ToXtlveJbskAOy;guKeTdFordaG7s8 zWdn621Q!kE`l#zY#IRWd&z%DH#LnwD+#QS!Dtp=5VT2DO;Vx!`_U+8Zuu{C+KMLb? zZfU|FS9L#J-_-P2^e(No(&f-IRk3s5P2utzExvu2%&iaOqS}G4bsLR=_5vK|XWqR9 z6X@(xf1K)3S7NQQ2X4X;_C9`I<>{Y*Men4CJvK%y{H%E)MklH%hRF1`I$+iqb1Xtx zCq$a`AJl94QE+M*zD`;jBW3|@OMlCt;%6Czh#7v1;T?x)O&4v}s^Mz9kexp-b@_uXy+E9U;~N9UMyUPF zpn1!L&{Gdw`m|_g3zUP7VMY-d=CnqQ;D?xg>~6eNag`%lx5G4_^V;#E=@%F&k?g!g z>FTpXqO3mZ$QsqGcv=pDa5V!F1fYY?EQd=8s)Zb>JIn*-Oyv=BO!#yR`A?*rDdtE% zt)1>zfukEh@miXqBOnoU<#!|Dx1}(bjdmkfCKa5GJy(`*8|gROsP~}p%yhh%4KbG( zfKpfQnz!HKA+36T;Uu8^uP$NY&Xd+*ry^dT=VAA+p0arYP72OJhZY(U>2@eqwyG&U zkKgGG!?SdwBN9n6Yyc;n==5LQ_0=Q=otlYu!o-4ZWx7Pgu~}N)XsjD zNd~nb>R;*vKeU^CNGNy$;)wKMn2^m3sHf)b5-C} z)1?>y_!%H_rNlNx1>0qcm48-%BCl`AyU=n-^-X-<6n+B#VXqCgfJW{5E|1u&@_d_n zO<6}0${H!Y?0~_u(|5HW!)a8*R~;;3)?8c^DpdBT*At4ibR<->?a3rJ8IS%0bP-*7z6JcVq$nzijO>LcPSf{d%nLYhZgjQe`idH zZ4d#@Z!Tx=EWRPO6RgjO*iPK1)1TWtfCwuli@L1z%xJZ)2LLo;9_Vv_0}w;m`0ZC- z*+%k*yBm-j7;4f0N{#3^Qw_V2YNVI2(?F`e^4w`~KBLHbd{@7|bDxNQIuwqE-)2B= z*vVz*bAKgO3nHUA;kVCEqL_w^b|9HR_hYp4+GW$w9H;!67^+1QU6c11Xm_*$Jx1|4 zcbNMWqXPx>svKW7z`<=OOn-Rm7FsqwW3&hFK+&JX=@2`lbf5l7a zbC2Cv!g3a8R|e|x!C49V&p0=?W!{o8L|T)D{J)*Mz&D~mCb-N_U*?RaCo?`=2~zp@ tKV^$x6ZrjS`Tv`8{r`s7Oi3GXb#Y0TVQH~kW)FBwEy_k{{>oSp?&}W literal 0 HcmV?d00001 diff --git a/chat_template.jinja b/chat_template.jinja new file mode 100644 index 0000000..bdf7919 --- /dev/null +++ b/chat_template.jinja @@ -0,0 +1,54 @@ +{%- if tools %} + {{- '<|im_start|>system\n' }} + {%- if messages[0]['role'] == 'system' %} + {{- messages[0]['content'] }} + {%- else %} + {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }} + {%- endif %} + {{- "\n\n# Tools\n\nYou may call one or more functions to assist with the user query.\n\nYou are provided with function signatures within XML tags:\n" }} + {%- for tool in tools %} + {{- "\n" }} + {{- tool | tojson }} + {%- endfor %} + {{- "\n\n\nFor each function call, return a json object with function name and arguments within XML tags:\n\n{\"name\": , \"arguments\": }\n<|im_end|>\n" }} +{%- else %} + {%- if messages[0]['role'] == 'system' %} + {{- '<|im_start|>system\n' + messages[0]['content'] + '<|im_end|>\n' }} + {%- else %} + {{- '<|im_start|>system\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\n' }} + {%- endif %} +{%- endif %} +{%- for message in messages %} + {%- if (message.role == "user") or (message.role == "system" and not loop.first) or (message.role == "assistant" and not message.tool_calls) %} + {{- '<|im_start|>' + message.role + '\n' + message.content + '<|im_end|>' + '\n' }} + {%- elif message.role == "assistant" %} + {{- '<|im_start|>' + message.role }} + {%- if message.content %} + {{- '\n' + message.content }} + {%- endif %} + {%- for tool_call in message.tool_calls %} + {%- if tool_call.function is defined %} + {%- set tool_call = tool_call.function %} + {%- endif %} + {{- '\n\n{"name": "' }} + {{- tool_call.name }} + {{- '", "arguments": ' }} + {{- tool_call.arguments | tojson }} + {{- '}\n' }} + {%- endfor %} + {{- '<|im_end|>\n' }} + {%- elif message.role == "tool" %} + {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %} + {{- '<|im_start|>user' }} + {%- endif %} + {{- '\n\n' }} + {{- message.content }} + {{- '\n' }} + {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %} + {{- '<|im_end|>\n' }} + {%- endif %} + {%- endif %} +{%- endfor %} +{%- if add_generation_prompt %} + {{- '<|im_start|>assistant\n' }} +{%- endif %} diff --git a/config.json b/config.json new file mode 100644 index 0000000..dc5f15b --- /dev/null +++ b/config.json @@ -0,0 +1,70 @@ +{ + "architectures": [ + "Qwen2ForCausalLM" + ], + "attention_dropout": 0.0, + "bos_token_id": null, + "torch_dtype": "float16", + "eos_token_id": 151645, + "hidden_act": "silu", + "hidden_size": 2048, + "initializer_range": 0.02, + "intermediate_size": 11008, + "layer_types": [ + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention", + "full_attention" + ], + "max_position_embeddings": 32768, + "max_window_layers": 70, + "model_type": "qwen2", + "num_attention_heads": 16, + "num_hidden_layers": 36, + "num_key_value_heads": 2, + "pad_token_id": 151654, + "rms_norm_eps": 1e-06, + "rope_parameters": { + "rope_theta": 1000000.0, + "rope_type": "default" + }, + "sliding_window": null, + "tie_word_embeddings": true, + "unsloth_fixed": true, + "unsloth_version": "2026.4.8", + "use_cache": false, + "use_sliding_window": false, + "vocab_size": 151936 +} \ No newline at end of file diff --git a/loss_and_kl.png b/loss_and_kl.png new file mode 100644 index 0000000..f17ca73 --- /dev/null +++ b/loss_and_kl.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e825d5e7b078c1f419c8702efdc748682e50623142b33f3a093569ce55d83b4a +size 195638 diff --git a/model-00001-of-00002.safetensors b/model-00001-of-00002.safetensors new file mode 100644 index 0000000..1981470 --- /dev/null +++ b/model-00001-of-00002.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:737c6aa5273de8b744ba022c202c7d9b2e7f98fc55b5cd2acabc184d3e72912b +size 3968658944 diff --git a/model-00002-of-00002.safetensors b/model-00002-of-00002.safetensors new file mode 100644 index 0000000..9256125 --- /dev/null +++ b/model-00002-of-00002.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:706e6579740f8dbd307aad99b72ef7a64e1ca55260bc9ae51a438fa9c2b55445 +size 2203268048 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000..38047c6 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,441 @@ +{ + "metadata": { + "total_size": 6171877376 + }, + "weight_map": { + "model.embed_tokens.weight": "model-00001-of-00002.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.0.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.0.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.1.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.10.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.10.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.10.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.10.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.11.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.11.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.11.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.11.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.12.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.12.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.12.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.12.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.13.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.13.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.14.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.14.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.14.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.14.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.15.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.15.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.15.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.15.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.16.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.16.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.16.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.16.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.17.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.17.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.17.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.17.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.18.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.18.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.18.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.18.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.19.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.19.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.19.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.19.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.2.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.2.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.20.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.20.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.20.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.20.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.21.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.21.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.21.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.22.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.22.self_attn.k_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.22.self_attn.q_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.22.self_attn.v_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.23.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.23.self_attn.k_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.23.self_attn.q_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.23.self_attn.v_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.24.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.24.self_attn.k_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.24.self_attn.q_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.24.self_attn.v_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.25.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.25.self_attn.k_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.25.self_attn.q_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.25.self_attn.v_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.26.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.26.self_attn.k_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.26.self_attn.q_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.26.self_attn.v_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.27.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.27.self_attn.k_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.27.self_attn.q_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.27.self_attn.v_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.28.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.28.self_attn.k_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.28.self_attn.q_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.28.self_attn.v_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.29.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.29.self_attn.k_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.29.self_attn.q_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.29.self_attn.v_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.3.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.3.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.3.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.30.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.30.self_attn.k_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.30.self_attn.q_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.30.self_attn.v_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.31.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.31.self_attn.k_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.31.self_attn.q_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.31.self_attn.v_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.32.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.32.self_attn.k_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.32.self_attn.q_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.32.self_attn.v_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.33.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.33.self_attn.k_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.33.self_attn.q_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.33.self_attn.v_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.34.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.34.self_attn.k_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.34.self_attn.q_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.34.self_attn.v_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.35.input_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00002-of-00002.safetensors", + "model.layers.35.self_attn.k_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.35.self_attn.q_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.35.self_attn.v_proj.bias": "model-00002-of-00002.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00002-of-00002.safetensors", + "model.layers.4.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.5.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.5.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.5.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.5.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.6.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.6.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.6.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.7.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.7.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.7.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.7.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.8.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.8.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.8.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.8.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.9.input_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00002.safetensors", + "model.layers.9.self_attn.k_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.9.self_attn.q_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors", + "model.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors", + "model.norm.weight": "model-00002-of-00002.safetensors" + } +} diff --git a/reward_components.png b/reward_components.png new file mode 100644 index 0000000..6356199 --- /dev/null +++ b/reward_components.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5f175c9c109a18ad8b26c195dc99a8d302170502b80254368873f4e3782688a8 +size 163775 diff --git a/tokenizer.json b/tokenizer.json new file mode 100644 index 0000000..5340d81 --- /dev/null +++ b/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd5948af71b4f56cf697f7580814c7ce8b80595ef985544efcacf716126a2e31 +size 11422356 diff --git a/tokenizer_config.json b/tokenizer_config.json new file mode 100644 index 0000000..544df20 --- /dev/null +++ b/tokenizer_config.json @@ -0,0 +1,202 @@ +{ + "add_prefix_space": false, + "backend": "tokenizers", + "bos_token": null, + "clean_up_tokenization_spaces": false, + "eos_token": "<|im_end|>", + "errors": "replace", + "is_local": false, + "model_max_length": 32768, + "pad_token": "<|PAD_TOKEN|>", + "padding_side": "left", + "split_special_tokens": false, + "tokenizer_class": "Qwen2Tokenizer", + "unk_token": null, + "added_tokens_decoder": { + "151643": { + "content": "<|endoftext|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + "151644": { + "content": "<|im_start|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + "151645": { + "content": "<|im_end|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + "151646": { + "content": "<|object_ref_start|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + "151647": { + "content": "<|object_ref_end|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + "151648": { + "content": "<|box_start|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + "151649": { + "content": "<|box_end|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + "151650": { + "content": "<|quad_start|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + "151651": { + "content": "<|quad_end|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + "151652": { + "content": "<|vision_start|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + "151653": { + "content": "<|vision_end|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + "151654": { + "content": "<|vision_pad|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + "151655": { + "content": "<|image_pad|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + "151656": { + "content": "<|video_pad|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + }, + "151657": { + "content": "", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": false + }, + "151658": { + "content": "", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": false + }, + "151659": { + "content": "<|fim_prefix|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": false + }, + "151660": { + "content": "<|fim_middle|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": false + }, + "151661": { + "content": "<|fim_suffix|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": false + }, + "151662": { + "content": "<|fim_pad|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": false + }, + "151663": { + "content": "<|repo_name|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": false + }, + "151664": { + "content": "<|file_sep|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": false + }, + "151665": { + "content": "<|PAD_TOKEN|>", + "single_word": false, + "lstrip": false, + "rstrip": false, + "normalized": false, + "special": true + } + }, + "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within XML tags:\\n\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n\\n\\nFor each function call, return a json object with function name and arguments within XML tags:\\n\\n{\\\"name\\\": , \\\"arguments\\\": }\\n<|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n\\n' }}\n {{- message.content }}\n {{- '\\n' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n" +} \ No newline at end of file diff --git a/total_reward_curve.png b/total_reward_curve.png new file mode 100644 index 0000000..7e7ea2b --- /dev/null +++ b/total_reward_curve.png @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0a8aafe82beb921a2dea0cb172c5b8b6e01e5432788fd6c75bd2bc7d3ee7b44c +size 174335 diff --git a/training_args.bin b/training_args.bin new file mode 100644 index 0000000..d275e04 --- /dev/null +++ b/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4fbd564a73c7fd8ca14e5c894248a8c9d7fe73340c76ec64a900b046822b31e9 +size 6865