Model: YoAbriel/KodaLite-1.3B-GGUF Source: Original Platform
language, license, library_name, pipeline_tag, tags, base_model
| language | license | library_name | pipeline_tag | tags | base_model | ||||
|---|---|---|---|---|---|---|---|---|---|
| en | apache-2.0 | gguf | text-generation |
|
YoAbriel/KodaLite-1.3B |
KodaLite-1.3B — GGUF quantizations
GGUF versions of YoAbriel/KodaLite-1.3B.
Files
| File | Quant | Size | Use case |
|---|---|---|---|
| kodalite-f16.gguf | F16 | ~2.5 GB | Full precision reference |
| kodalite-Q8_0.gguf | Q8_0 | ~1.3 GB | Near-lossless |
| kodalite-Q4_K_M.gguf | Q4_K_M | ~800 MB | Best size/quality tradeoff |
Usage
llama.cpp
llama-cli -m kodalite-Q4_K_M.gguf --reverse-prompt '<|end|>' -p '<|user|>\nHello\n<|assistant|>\n' -n 150
Ollama
cat > Modelfile << EOF
FROM ./kodalite-Q4_K_M.gguf
TEMPLATE """<|user|>
{{ .Prompt }}
<|assistant|>
"""
PARAMETER stop "<|end|>"
EOF
ollama create kodalite -f Modelfile
ollama run kodalite
LM Studio / Jan
Load the .gguf file directly. Stop sequence: <|end|>
License
Apache 2.0
Description