49 lines
1.7 KiB
Markdown
49 lines
1.7 KiB
Markdown
---
|
|
base_model:
|
|
- Qwen/Qwen2.5-Coder-32B-Instruct
|
|
datasets:
|
|
- SWE-bench/SWE-smith
|
|
language:
|
|
- en
|
|
library_name: transformers
|
|
license: apache-2.0
|
|
pipeline_tag: text-generation
|
|
tags:
|
|
- agent
|
|
- software engineering
|
|
---
|
|
|
|
# <span style="color: #7FFF7F;">SWE-agent-LM-32B GGUF Models</span>
|
|
|
|
|
|
## <span style="color: #7F7FFF;">Model Generation Details</span>
|
|
|
|
This model was generated using [llama.cpp](https://github.com/ggerganov/llama.cpp) at commit [`064cc596`](https://github.com/ggerganov/llama.cpp/commit/064cc596ac44308dc326a17c9e3163c34a6f29d1).
|
|
|
|
|
|
|
|
<div align="center">
|
|
<img src="https://swe-agent.com/latest/assets/swe-agent.svg" alt="Logo" width="200">
|
|
<h1 align="center">SWE-agent LM</h1>
|
|
</div>
|
|
|
|
<p align="center">
|
|
<a href="https://github.com/SWE-bench/SWE-smith">Code</a>
|
|
•
|
|
<a href="https://arxiv.org/abs/2504.21798">Paper</a>
|
|
•
|
|
<a href="https://swesmith.com/">Site</a>
|
|
</p>
|
|
|
|
SWE-agent-LM-32B is a Language Model for Software Engineering trained using the [SWE-smith](https://github.com/SWE-bench/SWE-smith) toolkit.
|
|
We introduce this model as part of our work: [SWE-smith: Scaling Data for Software Engineering Agents](https://swesmith.com).
|
|
|
|
SWE-agent-LM-32B is 100% open source.
|
|
Training this model was simple - we fine-tuned Qwen 2.5 Coder Instruct on 5k trajectories generated by SWE-agent + Claude 3.7 Sonnet.
|
|
The dataset can be found [here](https://huggingface.co/datasets/SWE-bench/SWE-smith-trajs-250429).
|
|
|
|
SWE-agent-LM-32B is compatible with [SWE-agent](https://github.com/SWE-agent/SWE-agent).
|
|
Running this model locally only takes a few steps!
|
|
Check [here]() for more instructions on how to do so.
|
|
|
|
If you found this work exciting and want to push SWE-agents further, please feel free to connect with us (the [SWE-bench team](https://swe-bench.github.io/)) more! |