base_model, inference, model_creator, model_name, pipeline_tag, quantized_by, tags
base_model inference model_creator model_name pipeline_tag quantized_by tags
obrmmk/tinycodellama-1.3b-5k false obrmmk tinycodellama-1.3b-5k text-generation afrideva
gguf
ggml
quantized
q2_k
q3_k_m
q4_k_m
q5_k_m
q6_k
q8_0

obrmmk/tinycodellama-1.3b-5k-GGUF

Quantized GGUF model files for tinycodellama-1.3b-5k from obrmmk

Name Quant method Size
tinycodellama-1.3b-5k.fp16.gguf fp16 2.58 GB
tinycodellama-1.3b-5k.q2_k.gguf q2_k 565.81 MB
tinycodellama-1.3b-5k.q3_k_m.gguf q3_k_m 644.08 MB
tinycodellama-1.3b-5k.q4_k_m.gguf q4_k_m 792.15 MB
tinycodellama-1.3b-5k.q5_k_m.gguf q5_k_m 921.00 MB
tinycodellama-1.3b-5k.q6_k.gguf q6_k 1.06 GB
tinycodellama-1.3b-5k.q8_0.gguf q8_0 1.37 GB

Original Model Card:

Description
Model synced from source: afrideva/tinycodellama-1.3b-5k-GGUF
Readme 25 KiB