From e426ac98d974ecf80d5269c2ed7c0331f8c40290 Mon Sep 17 00:00:00 2001 From: ModelHub XC Date: Thu, 30 Apr 2026 06:17:52 +0800 Subject: [PATCH] =?UTF-8?q?=E5=88=9D=E5=A7=8B=E5=8C=96=E9=A1=B9=E7=9B=AE?= =?UTF-8?q?=EF=BC=8C=E7=94=B1ModelHub=20XC=E7=A4=BE=E5=8C=BA=E6=8F=90?= =?UTF-8?q?=E4=BE=9B=E6=A8=A1=E5=9E=8B?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Model: DevQuasar/princeton-nlp.Llama-3-8B-ProLong-512k-Base-GGUF Source: Original Platform --- .gitattributes | 41 +++++++++++++++++++ README.md | 12 ++++++ configuration.json | 1 + ...nlp.Llama-3-8B-ProLong-512k-Base.Q2_K.gguf | 3 ++ ...p.Llama-3-8B-ProLong-512k-Base.Q3_K_M.gguf | 3 ++ ...p.Llama-3-8B-ProLong-512k-Base.Q4_K_M.gguf | 3 ++ ...p.Llama-3-8B-ProLong-512k-Base.Q5_K_M.gguf | 3 ++ ...nlp.Llama-3-8B-ProLong-512k-Base.Q6_K.gguf | 3 ++ ...nlp.Llama-3-8B-ProLong-512k-Base.Q8_0.gguf | 3 ++ 9 files changed, 72 insertions(+) create mode 100644 .gitattributes create mode 100644 README.md create mode 100644 configuration.json create mode 100644 princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q2_K.gguf create mode 100644 princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q3_K_M.gguf create mode 100644 princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q4_K_M.gguf create mode 100644 princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q5_K_M.gguf create mode 100644 princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q6_K.gguf create mode 100644 princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q8_0.gguf diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..ad047f0 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,41 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text +princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..590fded --- /dev/null +++ b/README.md @@ -0,0 +1,12 @@ +--- +base_model: +- princeton-nlp/Llama-3-8B-ProLong-512k-Base +pipeline_tag: text-generation +--- + +[](https://devquasar.com) + +'Make knowledge free for everyone' + +Quantized version of: [princeton-nlp/Llama-3-8B-ProLong-512k-Base](https://huggingface.co/princeton-nlp/Llama-3-8B-ProLong-512k-Base) +Buy Me a Coffee at ko-fi.com diff --git a/configuration.json b/configuration.json new file mode 100644 index 0000000..bbeeda1 --- /dev/null +++ b/configuration.json @@ -0,0 +1 @@ +{"framework": "pytorch", "task": "text-generation", "allow_remote": true} \ No newline at end of file diff --git a/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q2_K.gguf b/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q2_K.gguf new file mode 100644 index 0000000..619b58c --- /dev/null +++ b/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d23136af12f38f5ed8f27b227d04c56fd87fe039b26420b60bf06175654ec585 +size 3179131904 diff --git a/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q3_K_M.gguf b/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q3_K_M.gguf new file mode 100644 index 0000000..ae9b0b0 --- /dev/null +++ b/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eeed5689fe8752fb12cd15c1f4996c34d62541ec6c34991b9e17a943e33b5ca5 +size 4018918400 diff --git a/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q4_K_M.gguf b/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q4_K_M.gguf new file mode 100644 index 0000000..0013348 --- /dev/null +++ b/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:234871ffab168457f160c379638877b57739efd9f34a32620c94b4c0d3831ab9 +size 4920734720 diff --git a/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q5_K_M.gguf b/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q5_K_M.gguf new file mode 100644 index 0000000..8ca98dd --- /dev/null +++ b/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa0dbcc8596aac32b44ff82f43d998b5e7dc52b678e07df42c1627f0307c6a19 +size 5732987904 diff --git a/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q6_K.gguf b/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q6_K.gguf new file mode 100644 index 0000000..34081f3 --- /dev/null +++ b/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eedbdd9680d5116474346ec2fa8f9e85aa9c3aee344f6566d4bbd9c874cf65e7 +size 6596006912 diff --git a/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q8_0.gguf b/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q8_0.gguf new file mode 100644 index 0000000..080fbd4 --- /dev/null +++ b/princeton-nlp.Llama-3-8B-ProLong-512k-Base.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:87ac9a762e2960399ffdae5b9d619b70933e046ba4f2ca37545b23800e7236cc +size 8540771328