初始化项目,由ModelHub XC社区提供模型
Model: mradermacher/gemma-2B_Full_claude_1-i1-GGUF Source: Original Platform
This commit is contained in:
60
.gitattributes
vendored
Normal file
60
.gitattributes
vendored
Normal file
@@ -0,0 +1,60 @@
|
||||
*.7z filter=lfs diff=lfs merge=lfs -text
|
||||
*.arrow filter=lfs diff=lfs merge=lfs -text
|
||||
*.bin filter=lfs diff=lfs merge=lfs -text
|
||||
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
||||
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
||||
*.ftz filter=lfs diff=lfs merge=lfs -text
|
||||
*.gz filter=lfs diff=lfs merge=lfs -text
|
||||
*.h5 filter=lfs diff=lfs merge=lfs -text
|
||||
*.joblib filter=lfs diff=lfs merge=lfs -text
|
||||
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
||||
*.model filter=lfs diff=lfs merge=lfs -text
|
||||
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
||||
*.npy filter=lfs diff=lfs merge=lfs -text
|
||||
*.npz filter=lfs diff=lfs merge=lfs -text
|
||||
*.onnx filter=lfs diff=lfs merge=lfs -text
|
||||
*.ot filter=lfs diff=lfs merge=lfs -text
|
||||
*.parquet filter=lfs diff=lfs merge=lfs -text
|
||||
*.pb filter=lfs diff=lfs merge=lfs -text
|
||||
*.pickle filter=lfs diff=lfs merge=lfs -text
|
||||
*.pkl filter=lfs diff=lfs merge=lfs -text
|
||||
*.pt filter=lfs diff=lfs merge=lfs -text
|
||||
*.pth filter=lfs diff=lfs merge=lfs -text
|
||||
*.rar filter=lfs diff=lfs merge=lfs -text
|
||||
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
||||
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
||||
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
||||
*.tar filter=lfs diff=lfs merge=lfs -text
|
||||
*.tflite filter=lfs diff=lfs merge=lfs -text
|
||||
*.tgz filter=lfs diff=lfs merge=lfs -text
|
||||
*.wasm filter=lfs diff=lfs merge=lfs -text
|
||||
*.xz filter=lfs diff=lfs merge=lfs -text
|
||||
*.zip filter=lfs diff=lfs merge=lfs -text
|
||||
*.zst filter=lfs diff=lfs merge=lfs -text
|
||||
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
||||
imatrix.dat filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-IQ2_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-IQ2_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-IQ2_XXS.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-Q2_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
gemma-2B_Full_claude_1.i1-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
||||
76
README.md
Normal file
76
README.md
Normal file
@@ -0,0 +1,76 @@
|
||||
---
|
||||
base_model: omarabb315/gemma-2B_Full_claude_1
|
||||
language:
|
||||
- en
|
||||
library_name: transformers
|
||||
quantized_by: mradermacher
|
||||
tags: []
|
||||
---
|
||||
## About
|
||||
|
||||
<!-- ### quantize_version: 2 -->
|
||||
<!-- ### output_tensor_quantised: 1 -->
|
||||
<!-- ### convert_type: hf -->
|
||||
<!-- ### vocab_type: -->
|
||||
<!-- ### tags: nicoboss -->
|
||||
weighted/imatrix quants of https://huggingface.co/omarabb315/gemma-2B_Full_claude_1
|
||||
|
||||
<!-- provided-files -->
|
||||
static quants are available at https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-GGUF
|
||||
## Usage
|
||||
|
||||
If you are unsure how to use GGUF files, refer to one of [TheBloke's
|
||||
READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for
|
||||
more details, including on how to concatenate multi-part files.
|
||||
|
||||
## Provided Quants
|
||||
|
||||
(sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants)
|
||||
|
||||
| Link | Type | Size/GB | Notes |
|
||||
|:-----|:-----|--------:|:------|
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-IQ1_S.gguf) | i1-IQ1_S | 0.9 | for the desperate |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-IQ1_M.gguf) | i1-IQ1_M | 1.0 | mostly desperate |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 1.0 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-IQ2_XS.gguf) | i1-IQ2_XS | 1.1 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-IQ2_S.gguf) | i1-IQ2_S | 1.1 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-IQ2_M.gguf) | i1-IQ2_M | 1.2 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-Q2_K_S.gguf) | i1-Q2_K_S | 1.3 | very low quality |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 1.3 | lower quality |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-Q2_K.gguf) | i1-Q2_K | 1.3 | IQ3_XXS probably better |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-IQ3_XS.gguf) | i1-IQ3_XS | 1.4 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-IQ3_S.gguf) | i1-IQ3_S | 1.5 | beats Q3_K* |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-Q3_K_S.gguf) | i1-Q3_K_S | 1.5 | IQ3_XS probably better |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-IQ3_M.gguf) | i1-IQ3_M | 1.5 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-Q3_K_M.gguf) | i1-Q3_K_M | 1.6 | IQ3_S probably better |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-Q3_K_L.gguf) | i1-Q3_K_L | 1.7 | IQ3_M probably better |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-IQ4_XS.gguf) | i1-IQ4_XS | 1.7 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-IQ4_NL.gguf) | i1-IQ4_NL | 1.7 | prefer IQ4_XS |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-Q4_0.gguf) | i1-Q4_0 | 1.7 | fast, low quality |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-Q4_K_S.gguf) | i1-Q4_K_S | 1.7 | optimal size/speed/quality |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-Q4_K_M.gguf) | i1-Q4_K_M | 1.8 | fast, recommended |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-Q4_1.gguf) | i1-Q4_1 | 1.9 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-Q5_K_S.gguf) | i1-Q5_K_S | 2.0 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-Q5_K_M.gguf) | i1-Q5_K_M | 2.0 | |
|
||||
| [GGUF](https://huggingface.co/mradermacher/gemma-2B_Full_claude_1-i1-GGUF/resolve/main/gemma-2B_Full_claude_1.i1-Q6_K.gguf) | i1-Q6_K | 2.3 | practically like static Q6_K |
|
||||
|
||||
Here is a handy graph by ikawrakow comparing some lower-quality quant
|
||||
types (lower is better):
|
||||
|
||||

|
||||
|
||||
And here are Artefact2's thoughts on the matter:
|
||||
https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9
|
||||
|
||||
## FAQ / Model Request
|
||||
|
||||
See https://huggingface.co/mradermacher/model_requests for some answers to
|
||||
questions you might have and/or if you want some other model quantized.
|
||||
|
||||
## Thanks
|
||||
|
||||
I thank my company, [nethype GmbH](https://www.nethype.de/), for letting
|
||||
me use its servers and providing upgrades to my workstation to enable
|
||||
this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to.
|
||||
|
||||
<!-- end -->
|
||||
3
gemma-2B_Full_claude_1.i1-IQ1_M.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-IQ1_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:b66c962b1f87dd9295f3ec13a059a72a8ee45f52c2fa8f58413eaa667e09674b
|
||||
size 873797856
|
||||
3
gemma-2B_Full_claude_1.i1-IQ1_S.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-IQ1_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:c5616eabb0b0f704dbaaaf92d83616146e7fc252b5856b9f01ba72303c8fbc83
|
||||
size 832159968
|
||||
3
gemma-2B_Full_claude_1.i1-IQ2_M.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-IQ2_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:9f19dd74ce65c5b4588664128d2b34f61695098dc5c53a3342cae5b794a242a9
|
||||
size 1088014560
|
||||
3
gemma-2B_Full_claude_1.i1-IQ2_S.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-IQ2_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:8cbb16d4e1c3ab081c5bc7df5f93ae75a0c0c4cf1bedaeb356a4444c3cc3168b
|
||||
size 1032497376
|
||||
3
gemma-2B_Full_claude_1.i1-IQ2_XS.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-IQ2_XS.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:f4981cf86b5c2ca91e2ae9ff1fac3506dc0b680f5071b2720f5c500338290d80
|
||||
size 1002545376
|
||||
3
gemma-2B_Full_claude_1.i1-IQ2_XXS.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-IQ2_XXS.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:0994553f04b937c67080e5148365b9ce1cd313abb105208f3b8f85cde88dbc23
|
||||
size 943194336
|
||||
3
gemma-2B_Full_claude_1.i1-IQ3_M.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-IQ3_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:87f2e6e08e5bb1660b95edbf683e0a71a26a5c9e7d44f2a977a2abfe184a95b5
|
||||
size 1393561824
|
||||
3
gemma-2B_Full_claude_1.i1-IQ3_S.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-IQ3_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:bdf0af9534c23408d41f916baec483b8eed12cdc810151cf9b102ff880ea8d39
|
||||
size 1360660704
|
||||
3
gemma-2B_Full_claude_1.i1-IQ3_XS.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-IQ3_XS.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:9936432ff598ec1668ea3e04ca8e1de7f02e7b78ebde89d0ea07e42889fc5d24
|
||||
size 1314212064
|
||||
3
gemma-2B_Full_claude_1.i1-IQ3_XXS.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-IQ3_XXS.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:468096ad32a63027c3598346b044ea2949d0b4ed326c2eb3931ee9484adff058
|
||||
size 1181685984
|
||||
3
gemma-2B_Full_claude_1.i1-IQ4_NL.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-IQ4_NL.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:d6c67d4e7b690f52a713998ee386da14ec81eb9064bc13aabb03dd7363c3e6d6
|
||||
size 1629509856
|
||||
3
gemma-2B_Full_claude_1.i1-IQ4_XS.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-IQ4_XS.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:3267bf42a1aa896297474913127a77ee352c148d43e159e643511a4e4936f706
|
||||
size 1566251232
|
||||
3
gemma-2B_Full_claude_1.i1-Q2_K.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-Q2_K.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:1108ce53618cee7be0d9117ce13c0896426458f65336880acb75221c9d7df6a9
|
||||
size 1229830368
|
||||
3
gemma-2B_Full_claude_1.i1-Q2_K_S.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-Q2_K_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:166dff475282160284e296e3bf9b7af338ba219ca9ac870db27193301d135620
|
||||
size 1170000096
|
||||
3
gemma-2B_Full_claude_1.i1-Q3_K_L.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-Q3_K_L.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:5ff680ecb3ae799ae4d6af19684b4089ec8490f8b5690e6e35a1a71dfb07e355
|
||||
size 1550436576
|
||||
3
gemma-2B_Full_claude_1.i1-Q3_K_M.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-Q3_K_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:7eb6a04937a82129111d517309e83214672c97b017e8add66c2891d7b419d808
|
||||
size 1461668064
|
||||
3
gemma-2B_Full_claude_1.i1-Q3_K_S.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-Q3_K_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:4d0dd3c54445afe13b4aeb63e91c416353a49666ea2cc71087af8ebd5552feea
|
||||
size 1360660704
|
||||
3
gemma-2B_Full_claude_1.i1-Q4_0.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-Q4_0.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:053957b054cfc36521c2e727abdd78a8daafb9cca76a049ffdc2bb61a0f9817b
|
||||
size 1633491168
|
||||
3
gemma-2B_Full_claude_1.i1-Q4_1.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-Q4_1.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:39b01f216baf973fd705088d04b68478588ec0cabc890fa45fdcc36d07f6beca
|
||||
size 1756027104
|
||||
3
gemma-2B_Full_claude_1.i1-Q4_K_M.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-Q4_K_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:4068d6430efa8b616b2efde5d939bce43916408111c2a98b156e1436c9896a9d
|
||||
size 1708583136
|
||||
3
gemma-2B_Full_claude_1.i1-Q4_K_S.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-Q4_K_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:ac7ff3dcf5563ae7be0d18d5dd906094784d58da34ff16bbe3142959e9a477be
|
||||
size 1638652128
|
||||
3
gemma-2B_Full_claude_1.i1-Q5_K_M.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-Q5_K_M.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:a9740b493e1665220091cf015580ec4b88b5ac75fdd4838e00cb91071f1ee5ea
|
||||
size 1923279072
|
||||
3
gemma-2B_Full_claude_1.i1-Q5_K_S.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-Q5_K_S.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:14aa1c8970e20b3caa4a749af5bb31b6fd020a9fd2c0cb18608f7dd4a27631a6
|
||||
size 1882544352
|
||||
3
gemma-2B_Full_claude_1.i1-Q6_K.gguf
Normal file
3
gemma-2B_Full_claude_1.i1-Q6_K.gguf
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:a355a319469bd94f49c44bfaa984a73d04bd25521081d67d4a0536808390bb12
|
||||
size 2151393504
|
||||
3
imatrix.dat
Normal file
3
imatrix.dat
Normal file
@@ -0,0 +1,3 @@
|
||||
version https://git-lfs.github.com/spec/v1
|
||||
oid sha256:2b857599f0fb7b236b5f2e677c25d0863fe72e50fdf350c7d2058308120b90fa
|
||||
size 2375559
|
||||
Reference in New Issue
Block a user