Files
stable-code-3b-GGUF/README.md
ModelHub XC b353f8a6c0 初始化项目,由ModelHub XC社区提供模型
Model: brittlewis12/stable-code-3b-GGUF
Source: Original Platform
2026-04-14 01:57:05 +08:00

92 lines
3.6 KiB
Markdown
Raw Blame History

This file contains ambiguous Unicode characters

This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.

---
base_model: stabilityai/stable-code-3b
inference: false
language:
- en
license: other
license_link: https://huggingface.co/stabilityai/stable-code-3b/blob/main/LICENSE
model_creator: stabilityai
model_name: stable-code-3b
pipeline_tag: text-generation
datasets:
- tiiuae/falcon-refinedweb
- bigcode/the-stack-github-issues
- bigcode/commitpackft
- bigcode/starcoderdata
- EleutherAI/proof-pile-2
- meta-math/MetaMathQA
tags:
- causal-lm
- code
quantized_by: brittlewis12
---
# stable-code-3b GGUF
Original model: [stable-code-3b](https://huggingface.co/stabilityai/stable-code-3b)
Model creator: [StabilityAI](https://huggingface.co/stabilityai/stable-code-3b)
This repo contains GGUF format model files for StabilityAIs stable-code-3b with 16k context.
> stable-code-3b is a 2.7B billion parameter decoder-only language model pre-trained on 1.3 trillion tokens of diverse textual and code datasets. stable-code-3b is trained on 18 programming languages (selected based on the 2023 StackOverflow Developer Survey) and demonstrates state-of-the-art performance (compared to models of similar size) on the MultiPL-E metrics across multiple programming languages tested using BigCode's Evaluation Harness.
### What is GGUF?
GGUF is a file format for representing AI models. It is the third version of the format, introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp.
Converted using llama.cpp build 1897 (revision [2b3a665](https://github.com/ggerganov/llama.cpp/commit/2b3a665))
## Prompt template: Completion or Fill-in-Middle
### Completion
```
{{prompt}}
```
### Fill-in-Middle (FIM)
```
<fim_prefix>{{prefix code}}<fim_suffix>{{suffix code}}<fim_middle>
```
Example prompt with special prefix, suffix, and middle tokens in context:
```
<fim_prefix>def fib(n):
<fim_suffix>
else:
return fib(n - 2) + fib(n - 1)
<fim_middle>
```
---
## Download & run with [cnvrs](https://twitter.com/cnvrsai) on iPhone, iPad, and Mac!
![cnvrs.ai](https://pbs.twimg.com/profile_images/1744049151241797632/0mIP-P9e_400x400.jpg)
[cnvrs](https://testflight.apple.com/join/sFWReS7K) is the best app for private, local AI on your device:
- create & save **Characters** with custom system prompts & temperature settings
- download and experiment with any **GGUF model** you can [find on HuggingFace](https://huggingface.co/models?library=gguf)!
- make it your own with custom **Theme colors**
- powered by Metal ⚡️ & [Llama.cpp](https://github.com/ggerganov/llama.cpp), with **haptics** during response streaming!
- **try it out** yourself today, on [Testflight](https://testflight.apple.com/join/sFWReS7K)!
- follow [cnvrs on twitter](https://twitter.com/cnvrsai) to stay up to date
---
# Original Model Evaluation
![spiderchart](https://huggingface.co/stabilityai/stable-code-3b/raw/main/stable_code_3b_spiderchart.svg)
| Model | Size | Python | C++ | Javascript | Java | PHP | Rust |
|------------------|------|--------|------|------------|------|------|------|
| **Stable Code** | 3B | 32.4% | 30.9%| 32.1% | 32.1%| 24.2%| 23.0%|
| CodeLLama | 7B | 30.0% | 28.2%| 32.5% | 31.1%| 25.7%| 26.3%|
| Deepseek Coder | 1.3B | 28.6% | 29.2%| 28.7% | 29.0%| 23.6%| 18.5%|
| Wizard Coder | 3B | 31.6% | 25.6%| 26.2% | 25.8%| 25.3%| 20.4%|
| StarCoder | 3B | 21.6% | 19.8%| 21.5% | 20.5%| 19.0%| 16.9%|
| Replit Code V1.5 | 3B | 23.0% | 25.9%| 26.2% | 23.6%| 23.2%| 21.5%|
| Deci Coder | 1B | 19.1% | 6.8% | 18.4% | 16.7%| 2.1% | 1.7% |