Files
chipcraftx-rtlgen-7b/README.md
ModelHub XC 00ff14fc70 初始化项目,由ModelHub XC社区提供模型
Model: chipcraftx-io/chipcraftx-rtlgen-7b
Source: Original Platform
2026-04-10 23:14:07 +08:00

137 lines
4.6 KiB
Markdown

---
license: cc-by-nc-4.0
language:
- en
library_name: transformers
tags:
- verilog
- rtl
- hardware-design
- eda
- code-generation
- chip-design
base_model: Qwen/Qwen2.5-Coder-7B-Instruct
pipeline_tag: text-generation
model-index:
- name: chipcraftx-rtlgen-7b
results:
- task:
type: text-generation
name: Verilog Code Generation
dataset:
name: VerilogEval-Human
type: verilogeval
metrics:
- name: Functional Pass Rate (standalone)
type: pass@1
value: 36.5
verified: true
- name: Functional Pass Rate (ChipCraftX hybrid system)
type: pass@1
value: 98.7
verified: true
---
# chipcraftx-rtlgen-7b
**The local RTL generation engine powering [ChipCraftX](https://chipcraftx.io)** -- an AI platform that converts natural language specifications into synthesizable Verilog.
`chipcraftx-rtlgen-7b` handles first-pass Verilog generation at zero API cost. Within the full ChipCraftX hybrid pipeline, the system achieves **98.7% functional pass rate on VerilogEval-Human (154/156)**.
## Benchmark Results
### VerilogEval-Human (156 problems, functional simulation)
| Model | Parameters | Pass Rate |
|-------|-----------|-----------|
| VeriGen | 16B | 26.0% |
| **chipcraftx-rtlgen-7b (standalone)** | **7B** | **36.5%** |
| RTLCoder | 7B | 37.0% |
| CodeV | 7B | 53.2% |
| **ChipCraftX hybrid system** | **7B + Claude** | **98.7%** |
## Model Details
- **Base model**: [Qwen2.5-Coder-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-7B-Instruct)
- **Fine-tuning**: QLoRA (rank 64, alpha 128) on 76,811 Verilog training samples
- **Training**: 3 epochs, learning rate 2e-4, batch size 4
- **Architecture**: 28 layers, 3584 hidden size, 28 attention heads
- **Context window**: 4,096 tokens (generation), 32,768 (max position embeddings)
- **Precision**: bfloat16
## Usage
### Transformers
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model_name = "ChipCraftX/chipcraftx-rtlgen-7b"
tokenizer = AutoTokenizer.from_pretrained(model_name)
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype="auto", device_map="auto")
system_prompt = """You are ChipCraft-RTL, an expert Verilog design engineer.
Generate synthesizable, lint-clean RTL that exactly matches the specification.
Rules:
- Output ONLY Verilog code (no prose, no markdown fences).
- Use reg and wire types ONLY -- NEVER use logic.
- Use always @(posedge clk) and always @(*) -- NEVER use always_ff or always_comb.
- Module name MUST be TopModule."""
spec = """Implement a module named TopModule with the following interface.
- input clk
- input reset
- output [3:0] count
The module should implement a 4-bit up counter with synchronous reset."""
messages = [
{"role": "system", "content": system_prompt},
{"role": "user", "content": spec},
]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=4096, temperature=0.2, do_sample=True)
print(tokenizer.decode(outputs[0][inputs.input_ids.shape[1]:], skip_special_tokens=True))
```
### Ollama
See [chipcraftx-rtlgen-7b-GGUF](https://huggingface.co/ChipCraftX/chipcraftx-rtlgen-7b-GGUF) for quantized GGUF versions compatible with Ollama and llama.cpp.
## Training Data
The model was fine-tuned on a proprietary dataset of 76,811 Verilog samples.
## Intended Use
This model is designed for:
- First-pass RTL/Verilog code generation from natural language specs
- Integration into automated EDA pipelines with validation feedback loops
- Educational use in digital design courses
- Rapid prototyping of hardware modules
### Limitations
- Standalone pass rate (36.5%) means ~2 out of 3 complex problems need iteration or human review
- Strongest on combinational logic; weaker on FSMs and sequential designs
- Outputs target Verilog-2001 / Icarus Verilog compatibility (not full SystemVerilog)
- Should always be paired with EDA validation (iverilog, Yosys) before use in production
## About ChipCraftX
[ChipCraftX](https://chipcraftx.io) is an AI-powered platform that converts natural language specifications into verified, synthesizable hardware descriptions. The platform combines local and cloud models with automated EDA validation to achieve near-perfect scores on standard RTL benchmarks.
## Citation
```bibtex
@misc{chipcraftx-rtlgen-7b,
title={chipcraftx-rtlgen-7b: Local RTL Generation Engine for ChipCraftX},
author={Eryilmaz, Cagri},
year={2026},
publisher={HuggingFace},
url={https://huggingface.co/ChipCraftX/chipcraftx-rtlgen-7b}
}
```