commit d859d29f06e21f9375dd010d28a3ef242c7bfcf0 Author: ModelHub XC Date: Wed Apr 22 09:52:58 2026 +0800 初始化项目,由ModelHub XC社区提供模型 Model: RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf Source: Original Platform diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..1029ae7 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,57 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.IQ3_S.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.IQ3_M.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q3_K.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q4_0.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.IQ4_NL.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q4_K.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q4_1.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q5_0.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q5_K.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q5_1.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text +llama-3.2-3b-Content-Moderation.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..5d50834 --- /dev/null +++ b/README.md @@ -0,0 +1,243 @@ +Quantization made by Richard Erkhov. + +[Github](https://github.com/RichardErkhov) + +[Discord](https://discord.gg/pvy7H8DZMG) + +[Request more models](https://github.com/RichardErkhov/quant_request) + + +llama-3.2-3b-Content-Moderation - GGUF +- Model creator: https://huggingface.co/ZiweiLiu96/ +- Original model: https://huggingface.co/ZiweiLiu96/llama-3.2-3b-Content-Moderation/ + + +| Name | Quant method | Size | +| ---- | ---- | ---- | +| [llama-3.2-3b-Content-Moderation.Q2_K.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q2_K.gguf) | Q2_K | 1.27GB | +| [llama-3.2-3b-Content-Moderation.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.IQ3_XS.gguf) | IQ3_XS | 1.38GB | +| [llama-3.2-3b-Content-Moderation.IQ3_S.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.IQ3_S.gguf) | IQ3_S | 1.44GB | +| [llama-3.2-3b-Content-Moderation.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q3_K_S.gguf) | Q3_K_S | 1.44GB | +| [llama-3.2-3b-Content-Moderation.IQ3_M.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.IQ3_M.gguf) | IQ3_M | 1.49GB | +| [llama-3.2-3b-Content-Moderation.Q3_K.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q3_K.gguf) | Q3_K | 1.57GB | +| [llama-3.2-3b-Content-Moderation.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q3_K_M.gguf) | Q3_K_M | 1.57GB | +| [llama-3.2-3b-Content-Moderation.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q3_K_L.gguf) | Q3_K_L | 1.69GB | +| [llama-3.2-3b-Content-Moderation.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.IQ4_XS.gguf) | IQ4_XS | 1.71GB | +| [llama-3.2-3b-Content-Moderation.Q4_0.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q4_0.gguf) | Q4_0 | 1.79GB | +| [llama-3.2-3b-Content-Moderation.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.IQ4_NL.gguf) | IQ4_NL | 1.79GB | +| [llama-3.2-3b-Content-Moderation.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q4_K_S.gguf) | Q4_K_S | 1.8GB | +| [llama-3.2-3b-Content-Moderation.Q4_K.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q4_K.gguf) | Q4_K | 1.88GB | +| [llama-3.2-3b-Content-Moderation.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q4_K_M.gguf) | Q4_K_M | 1.88GB | +| [llama-3.2-3b-Content-Moderation.Q4_1.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q4_1.gguf) | Q4_1 | 1.95GB | +| [llama-3.2-3b-Content-Moderation.Q5_0.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q5_0.gguf) | Q5_0 | 2.11GB | +| [llama-3.2-3b-Content-Moderation.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q5_K_S.gguf) | Q5_K_S | 2.11GB | +| [llama-3.2-3b-Content-Moderation.Q5_K.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q5_K.gguf) | Q5_K | 2.16GB | +| [llama-3.2-3b-Content-Moderation.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q5_K_M.gguf) | Q5_K_M | 2.16GB | +| [llama-3.2-3b-Content-Moderation.Q5_1.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q5_1.gguf) | Q5_1 | 2.28GB | +| [llama-3.2-3b-Content-Moderation.Q6_K.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q6_K.gguf) | Q6_K | 2.46GB | +| [llama-3.2-3b-Content-Moderation.Q8_0.gguf](https://huggingface.co/RichardErkhov/ZiweiLiu96_-_llama-3.2-3b-Content-Moderation-gguf/blob/main/llama-3.2-3b-Content-Moderation.Q8_0.gguf) | Q8_0 | 3.19GB | + + + + +Original model description: +--- +library_name: transformers +tags: [] +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + +This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] + diff --git a/llama-3.2-3b-Content-Moderation.IQ3_M.gguf b/llama-3.2-3b-Content-Moderation.IQ3_M.gguf new file mode 100644 index 0000000..788e97f --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.IQ3_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66c1551071144aa483d03c2570e729ddaa3f450b9724352d5767334af52a02f5 +size 1599669760 diff --git a/llama-3.2-3b-Content-Moderation.IQ3_S.gguf b/llama-3.2-3b-Content-Moderation.IQ3_S.gguf new file mode 100644 index 0000000..49c313e --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.IQ3_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d6c62e1928f7a7d2a88676dcda6e9f6e455908d9bab96f3b285260e0dcb46109 +size 1542850048 diff --git a/llama-3.2-3b-Content-Moderation.IQ3_XS.gguf b/llama-3.2-3b-Content-Moderation.IQ3_XS.gguf new file mode 100644 index 0000000..8dbeefa --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.IQ3_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6cb3eab7a2f1b0f4814cdf84f618e6dfd4fbfbe917eaa25321ae6a6acaeea7e9 +size 1476789760 diff --git a/llama-3.2-3b-Content-Moderation.IQ4_NL.gguf b/llama-3.2-3b-Content-Moderation.IQ4_NL.gguf new file mode 100644 index 0000000..a97435f --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.IQ4_NL.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02ebfaf302c00b47add2d37424fd37ce332f99e25330381d304ce7ed098b40bb +size 1926628864 diff --git a/llama-3.2-3b-Content-Moderation.IQ4_XS.gguf b/llama-3.2-3b-Content-Moderation.IQ4_XS.gguf new file mode 100644 index 0000000..ff3ad1b --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.IQ4_XS.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d3cdcb805e4b0de167589227270a66f731ddd96ea4be33f9e027f92cbb27833 +size 1840907776 diff --git a/llama-3.2-3b-Content-Moderation.Q2_K.gguf b/llama-3.2-3b-Content-Moderation.Q2_K.gguf new file mode 100644 index 0000000..f95aee5 --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q2_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c2ae660f065e069371c76501c43afa4d9a1c09ac760a55913d115831d7dd507 +size 1363936768 diff --git a/llama-3.2-3b-Content-Moderation.Q3_K.gguf b/llama-3.2-3b-Content-Moderation.Q3_K.gguf new file mode 100644 index 0000000..3a8f3a2 --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q3_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:885bd110843721c0765d065a45fa051f3b908db67d70f489b492a3ab0508a989 +size 1687160320 diff --git a/llama-3.2-3b-Content-Moderation.Q3_K_L.gguf b/llama-3.2-3b-Content-Moderation.Q3_K_L.gguf new file mode 100644 index 0000000..37da386 --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q3_K_L.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:758818fb4bd0afebef0ad81ab0e1e7237bf0b4fbb8bc2766aba97ffcd91e7d41 +size 1815348736 diff --git a/llama-3.2-3b-Content-Moderation.Q3_K_M.gguf b/llama-3.2-3b-Content-Moderation.Q3_K_M.gguf new file mode 100644 index 0000000..3a8f3a2 --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q3_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:885bd110843721c0765d065a45fa051f3b908db67d70f489b492a3ab0508a989 +size 1687160320 diff --git a/llama-3.2-3b-Content-Moderation.Q3_K_S.gguf b/llama-3.2-3b-Content-Moderation.Q3_K_S.gguf new file mode 100644 index 0000000..4af424a --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q3_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db10fbbc1c6e699088a744d5bdb495c245a21effc06e7aa9e7abf72d25b00d6b +size 1542850048 diff --git a/llama-3.2-3b-Content-Moderation.Q4_0.gguf b/llama-3.2-3b-Content-Moderation.Q4_0.gguf new file mode 100644 index 0000000..1d60611 --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c2d2a0a460474d487ef023e644978a5e52a562790108bcb04d2c9f89e6798b78 +size 1917191680 diff --git a/llama-3.2-3b-Content-Moderation.Q4_1.gguf b/llama-3.2-3b-Content-Moderation.Q4_1.gguf new file mode 100644 index 0000000..e8165cd --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8aab453bb387311e339d4cbc130cd3214d797a8b78b075fd4f0e4486ec917e2 +size 2093352448 diff --git a/llama-3.2-3b-Content-Moderation.Q4_K.gguf b/llama-3.2-3b-Content-Moderation.Q4_K.gguf new file mode 100644 index 0000000..496ba84 --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q4_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f221ec5e28c9451b6fed4bb3eb7336ac92ae997885c93427e165933bf81dba8e +size 2019378688 diff --git a/llama-3.2-3b-Content-Moderation.Q4_K_M.gguf b/llama-3.2-3b-Content-Moderation.Q4_K_M.gguf new file mode 100644 index 0000000..496ba84 --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q4_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f221ec5e28c9451b6fed4bb3eb7336ac92ae997885c93427e165933bf81dba8e +size 2019378688 diff --git a/llama-3.2-3b-Content-Moderation.Q4_K_S.gguf b/llama-3.2-3b-Content-Moderation.Q4_K_S.gguf new file mode 100644 index 0000000..cdab09a --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q4_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f2d678fba2f94a150c62761f9c03223a22df96ddf6ee77fcaf7f0a8daf2bda28 +size 1928201728 diff --git a/llama-3.2-3b-Content-Moderation.Q5_0.gguf b/llama-3.2-3b-Content-Moderation.Q5_0.gguf new file mode 100644 index 0000000..c6649e2 --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q5_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e985ce0c8dab7a98acf6ead576fda6cc1993067f04f678ee2895dac340c7bb19 +size 2269513216 diff --git a/llama-3.2-3b-Content-Moderation.Q5_1.gguf b/llama-3.2-3b-Content-Moderation.Q5_1.gguf new file mode 100644 index 0000000..0910466 --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q5_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f2062f032fd073e7b65a6097f92279bbecbf5c2c8607d57400f8a2cabebd766 +size 2445673984 diff --git a/llama-3.2-3b-Content-Moderation.Q5_K.gguf b/llama-3.2-3b-Content-Moderation.Q5_K.gguf new file mode 100644 index 0000000..b174a0d --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q5_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a970e5d6e75973045a1e879beb267bb927472cd7864f156341f4d551935ad8e1 +size 2322155008 diff --git a/llama-3.2-3b-Content-Moderation.Q5_K_M.gguf b/llama-3.2-3b-Content-Moderation.Q5_K_M.gguf new file mode 100644 index 0000000..b174a0d --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q5_K_M.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a970e5d6e75973045a1e879beb267bb927472cd7864f156341f4d551935ad8e1 +size 2322155008 diff --git a/llama-3.2-3b-Content-Moderation.Q5_K_S.gguf b/llama-3.2-3b-Content-Moderation.Q5_K_S.gguf new file mode 100644 index 0000000..c1cee33 --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q5_K_S.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5560b6c750db72af3ccc336b2fd13b2f4a0b5ec97fcc24152f14ce034e93bff9 +size 2269513216 diff --git a/llama-3.2-3b-Content-Moderation.Q6_K.gguf b/llama-3.2-3b-Content-Moderation.Q6_K.gguf new file mode 100644 index 0000000..56b221d --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q6_K.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b6f3e1a6ce100f9c85b1902f36369ae3952846fcf35b23fa6a439de852b1918 +size 2643854848 diff --git a/llama-3.2-3b-Content-Moderation.Q8_0.gguf b/llama-3.2-3b-Content-Moderation.Q8_0.gguf new file mode 100644 index 0000000..ae62b9e --- /dev/null +++ b/llama-3.2-3b-Content-Moderation.Q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:21c2d4280cf4e4eba61ace6cc43c7bf93809a4e315026aba98b6be4197d21aeb +size 3421901760