language, license, library_name, pipeline_tag, tags, base_model
language license library_name pipeline_tag tags base_model
en apache-2.0 gguf text-generation
text-generation
gguf
llama.cpp
ollama
YoAbriel/KodaLite-1.3B

KodaLite-1.3B — GGUF quantizations

GGUF versions of YoAbriel/KodaLite-1.3B.

Files

File Quant Size Use case
kodalite-f16.gguf F16 ~2.5 GB Full precision reference
kodalite-Q8_0.gguf Q8_0 ~1.3 GB Near-lossless
kodalite-Q4_K_M.gguf Q4_K_M ~800 MB Best size/quality tradeoff

Usage

llama.cpp

llama-cli -m kodalite-Q4_K_M.gguf --reverse-prompt '<|end|>' -p '<|user|>\nHello\n<|assistant|>\n' -n 150

Ollama

cat > Modelfile << EOF
FROM ./kodalite-Q4_K_M.gguf
TEMPLATE """<|user|>
{{ .Prompt }}
<|assistant|>
"""
PARAMETER stop "<|end|>"
EOF
ollama create kodalite -f Modelfile
ollama run kodalite

LM Studio / Jan

Load the .gguf file directly. Stop sequence: <|end|>

License

Apache 2.0

Description
Model synced from source: YoAbriel/KodaLite-1.3B-GGUF
Readme 25 KiB