2024-11-07 23:15:07 +00:00
---
base_model: elinas/Llama-3-13B-Instruct
library_name: transformers
tags:
- mergekit
- merge
- TensorBlock
- GGUF
license: llama3
---
< div style = "width: auto; margin-left: auto; margin-right: auto" >
< img src = "https://i.imgur.com/jC7kdl8.jpeg" alt = "TensorBlock" style = "width: 100%; min-width: 400px; display: block; margin: auto;" >
< / div >
2025-06-18 23:30:17 +00:00
[](https://tensorblock.co)
[](https://twitter.com/tensorblock_aoi)
[](https://discord.gg/Ej5NmeHFf2)
[](https://github.com/TensorBlock)
[](https://t.me/TensorBlock)
2024-11-07 23:15:07 +00:00
## elinas/Llama-3-13B-Instruct - GGUF
This repo contains GGUF format model files for [elinas/Llama-3-13B-Instruct ](https://huggingface.co/elinas/Llama-3-13B-Instruct ).
The files were quantized using machines provided by [TensorBlock ](https://tensorblock.co/ ), and they are compatible with llama.cpp as of [commit b4011 ](https://github.com/ggerganov/llama.cpp/commit/a6744e43e80f4be6398fc7733a01642c846dce1d ).
2024-11-16 00:40:12 +00:00
2025-04-20 22:29:32 +00:00
## Our projects
< table border = "1" cellspacing = "0" cellpadding = "10" >
< tr >
2025-07-08 22:27:11 +00:00
< th colspan = "2" style = "font-size: 25px;" > Forge< / th >
< / tr >
< tr >
< th colspan = "2" >
< img src = "https://imgur.com/faI5UKh.jpeg" alt = "Forge Project" width = "900" / >
< / th >
< / tr >
< tr >
< th colspan = "2" > An OpenAI-compatible multi-provider routing layer.< / th >
< / tr >
< tr >
< th colspan = "2" >
< a href = "https://github.com/TensorBlock/forge" target = "_blank" style = "
display: inline-block;
padding: 8px 16px;
background-color: #FF7F50 ;
color: white;
text-decoration: none;
border-radius: 6px;
font-weight: bold;
font-family: sans-serif;
">🚀 Try it now! 🚀< / a >
< / th >
< / tr >
< tr >
< th style = "font-size: 25px;" > Awesome MCP Servers< / th >
< th style = "font-size: 25px;" > TensorBlock Studio< / th >
< / tr >
< tr >
< th > < img src = "https://imgur.com/2Xov7B7.jpeg" alt = "MCP Servers" width = "450" / > < / th >
< th > < img src = "https://imgur.com/pJcmF5u.jpeg" alt = "Studio" width = "450" / > < / th >
2025-04-20 22:29:32 +00:00
< / tr >
< tr >
< th > A comprehensive collection of Model Context Protocol (MCP) servers.< / th >
< th > A lightweight, open, and extensible multi-LLM interaction studio.< / th >
< / tr >
2025-07-08 22:27:11 +00:00
< tr >
< th >
< a href = "https://github.com/TensorBlock/awesome-mcp-servers" target = "_blank" style = "
display: inline-block;
padding: 8px 16px;
background-color: #FF7F50 ;
color: white;
text-decoration: none;
border-radius: 6px;
font-weight: bold;
font-family: sans-serif;
">👀 See what we built 👀< / a >
< / th >
< th >
< a href = "https://github.com/TensorBlock/TensorBlock-Studio" target = "_blank" style = "
display: inline-block;
padding: 8px 16px;
background-color: #FF7F50 ;
color: white;
text-decoration: none;
border-radius: 6px;
font-weight: bold;
font-family: sans-serif;
">👀 See what we built 👀< / a >
< / th >
< / tr >
2025-04-20 22:29:32 +00:00
< / table >
2024-11-07 23:15:07 +00:00
## Prompt template
2024-11-16 00:40:12 +00:00
2024-11-07 23:15:07 +00:00
```
< |begin_of_text|>< |start_header_id|>system< |end_header_id|>
{system_prompt}< |eot_id|>< |start_header_id|>user< |end_header_id|>
{prompt}< |eot_id|>< |start_header_id|>assistant< |end_header_id|>
```
## Model file specification
| Filename | Quant type | File Size | Description |
| -------- | ---------- | --------- | ----------- |
2024-11-16 00:40:12 +00:00
| [Llama-3-13B-Instruct-Q2_K.gguf ](https://huggingface.co/tensorblock/Llama-3-13B-Instruct-GGUF/blob/main/Llama-3-13B-Instruct-Q2_K.gguf ) | Q2_K | 4.680 GB | smallest, significant quality loss - not recommended for most purposes |
| [Llama-3-13B-Instruct-Q3_K_S.gguf ](https://huggingface.co/tensorblock/Llama-3-13B-Instruct-GGUF/blob/main/Llama-3-13B-Instruct-Q3_K_S.gguf ) | Q3_K_S | 5.421 GB | very small, high quality loss |
| [Llama-3-13B-Instruct-Q3_K_M.gguf ](https://huggingface.co/tensorblock/Llama-3-13B-Instruct-GGUF/blob/main/Llama-3-13B-Instruct-Q3_K_M.gguf ) | Q3_K_M | 5.985 GB | very small, high quality loss |
| [Llama-3-13B-Instruct-Q3_K_L.gguf ](https://huggingface.co/tensorblock/Llama-3-13B-Instruct-GGUF/blob/main/Llama-3-13B-Instruct-Q3_K_L.gguf ) | Q3_K_L | 6.473 GB | small, substantial quality loss |
| [Llama-3-13B-Instruct-Q4_0.gguf ](https://huggingface.co/tensorblock/Llama-3-13B-Instruct-GGUF/blob/main/Llama-3-13B-Instruct-Q4_0.gguf ) | Q4_0 | 6.970 GB | legacy; small, very high quality loss - prefer using Q3_K_M |
| [Llama-3-13B-Instruct-Q4_K_S.gguf ](https://huggingface.co/tensorblock/Llama-3-13B-Instruct-GGUF/blob/main/Llama-3-13B-Instruct-Q4_K_S.gguf ) | Q4_K_S | 7.013 GB | small, greater quality loss |
| [Llama-3-13B-Instruct-Q4_K_M.gguf ](https://huggingface.co/tensorblock/Llama-3-13B-Instruct-GGUF/blob/main/Llama-3-13B-Instruct-Q4_K_M.gguf ) | Q4_K_M | 7.378 GB | medium, balanced quality - recommended |
| [Llama-3-13B-Instruct-Q5_0.gguf ](https://huggingface.co/tensorblock/Llama-3-13B-Instruct-GGUF/blob/main/Llama-3-13B-Instruct-Q5_0.gguf ) | Q5_0 | 8.427 GB | legacy; medium, balanced quality - prefer using Q4_K_M |
| [Llama-3-13B-Instruct-Q5_K_S.gguf ](https://huggingface.co/tensorblock/Llama-3-13B-Instruct-GGUF/blob/main/Llama-3-13B-Instruct-Q5_K_S.gguf ) | Q5_K_S | 8.427 GB | large, low quality loss - recommended |
| [Llama-3-13B-Instruct-Q5_K_M.gguf ](https://huggingface.co/tensorblock/Llama-3-13B-Instruct-GGUF/blob/main/Llama-3-13B-Instruct-Q5_K_M.gguf ) | Q5_K_M | 8.637 GB | large, very low quality loss - recommended |
| [Llama-3-13B-Instruct-Q6_K.gguf ](https://huggingface.co/tensorblock/Llama-3-13B-Instruct-GGUF/blob/main/Llama-3-13B-Instruct-Q6_K.gguf ) | Q6_K | 9.976 GB | very large, extremely low quality loss |
| [Llama-3-13B-Instruct-Q8_0.gguf ](https://huggingface.co/tensorblock/Llama-3-13B-Instruct-GGUF/blob/main/Llama-3-13B-Instruct-Q8_0.gguf ) | Q8_0 | 12.919 GB | very large, extremely low quality loss - not recommended |
2024-11-07 23:15:07 +00:00
## Downloading instruction
### Command line
Firstly, install Huggingface Client
```shell
pip install -U "huggingface_hub[cli]"
```
Then, downoad the individual model file the a local directory
```shell
huggingface-cli download tensorblock/Llama-3-13B-Instruct-GGUF --include "Llama-3-13B-Instruct-Q2_K.gguf" --local-dir MY_LOCAL_DIR
```
If you wanna download multiple model files with a pattern (e.g., `*Q4_K*gguf` ), you can try:
```shell
huggingface-cli download tensorblock/Llama-3-13B-Instruct-GGUF --local-dir MY_LOCAL_DIR --local-dir-use-symlinks False --include='*Q4_K*gguf'
```