commit bd38b1a4e60b9b18a1c005b6d54170752d5b1a37 Author: Mungert Date: Wed Sep 24 15:44:04 2025 +0000 Super-squash history to reclaim storage diff --git a/.gitattributes b/.gitattributes new file mode 100644 index 0000000..89bb071 --- /dev/null +++ b/.gitattributes @@ -0,0 +1,81 @@ +*.7z filter=lfs diff=lfs merge=lfs -text +*.arrow filter=lfs diff=lfs merge=lfs -text +*.bin filter=lfs diff=lfs merge=lfs -text +*.bz2 filter=lfs diff=lfs merge=lfs -text +*.ckpt filter=lfs diff=lfs merge=lfs -text +*.ftz filter=lfs diff=lfs merge=lfs -text +*.gz filter=lfs diff=lfs merge=lfs -text +*.h5 filter=lfs diff=lfs merge=lfs -text +*.joblib filter=lfs diff=lfs merge=lfs -text +*.lfs.* filter=lfs diff=lfs merge=lfs -text +*.mlmodel filter=lfs diff=lfs merge=lfs -text +*.model filter=lfs diff=lfs merge=lfs -text +*.msgpack filter=lfs diff=lfs merge=lfs -text +*.npy filter=lfs diff=lfs merge=lfs -text +*.npz filter=lfs diff=lfs merge=lfs -text +*.onnx filter=lfs diff=lfs merge=lfs -text +*.ot filter=lfs diff=lfs merge=lfs -text +*.parquet filter=lfs diff=lfs merge=lfs -text +*.pb filter=lfs diff=lfs merge=lfs -text +*.pickle filter=lfs diff=lfs merge=lfs -text +*.pkl filter=lfs diff=lfs merge=lfs -text +*.pt filter=lfs diff=lfs merge=lfs -text +*.pth filter=lfs diff=lfs merge=lfs -text +*.rar filter=lfs diff=lfs merge=lfs -text +*.safetensors filter=lfs diff=lfs merge=lfs -text +saved_model/**/* filter=lfs diff=lfs merge=lfs -text +*.tar.* filter=lfs diff=lfs merge=lfs -text +*.tar filter=lfs diff=lfs merge=lfs -text +*.tflite filter=lfs diff=lfs merge=lfs -text +*.tgz filter=lfs diff=lfs merge=lfs -text +*.wasm filter=lfs diff=lfs merge=lfs -text +*.xz filter=lfs diff=lfs merge=lfs -text +*.zip filter=lfs diff=lfs merge=lfs -text +*.zst filter=lfs diff=lfs merge=lfs -text +*tfevents* filter=lfs diff=lfs merge=lfs -text +f16/SWE-Dev-32B-f16-00001-of-00002.gguf filter=lfs diff=lfs merge=lfs -text +f16/SWE-Dev-32B-f16-00002-of-00002.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-f16_q8_0.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-bf16_q8_0.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-f16_q6_k.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-bf16_q6_k.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-f16_q4_k.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-bf16_q4_k.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q2_k_l.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q3_k_l.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q4_k_l.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q5_k_l.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q6_k_l.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q2_k_m.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q2_k_s.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q3_k_m.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q3_k_s.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q4_k_m.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q4_k_s.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q5_k_m.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q5_k_s.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q6_k_m.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q8_0.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q4_0.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q4_1.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q4_0_l.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q4_1_l.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q5_0.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q5_1.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q5_0_l.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-q5_1_l.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-iq1_s.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-iq1_m.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-iq2_xs.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-iq2_xxs.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-iq2_s.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-iq2_m.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-iq3_xs.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-iq3_xxs.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-iq3_s.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-iq3_m.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-iq4_xs.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B-iq4_nl.gguf filter=lfs diff=lfs merge=lfs -text +SWE-Dev-32B.imatrix filter=lfs diff=lfs merge=lfs -text +bf16/SWE-Dev-32B-bf16-00001-of-00002.gguf filter=lfs diff=lfs merge=lfs -text +bf16/SWE-Dev-32B-bf16-00002-of-00002.gguf filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md new file mode 100644 index 0000000..3369605 --- /dev/null +++ b/README.md @@ -0,0 +1,121 @@ +--- +base_model: +- Qwen/Qwen2.5-Coder-32B-Instruct +library_name: transformers +license: mit +pipeline_tag: text-generation +--- + +# SWE-Dev-32B GGUF Models + + +## Model Generation Details + +This model was generated using [llama.cpp](https://github.com/ggerganov/llama.cpp) at commit [`0a5a3b5c`](https://github.com/ggerganov/llama.cpp/commit/0a5a3b5cdfd887cf0f8e09d9ff89dee130cfcdde). + + + + + +--- + +## Quantization Beyond the IMatrix + +I've been experimenting with a new quantization approach that selectively elevates the precision of key layers beyond what the default IMatrix configuration provides. + +In my testing, standard IMatrix quantization underperforms at lower bit depths, especially with Mixture of Experts (MoE) models. To address this, I'm using the `--tensor-type` option in `llama.cpp` to manually "bump" important layers to higher precision. You can see the implementation here: +πŸ‘‰ [Layer bumping with llama.cpp](https://github.com/Mungert69/GGUFModelBuilder/blob/main/model-converter/tensor_list_builder.py) + +While this does increase model file size, it significantly improves precision for a given quantization level. + +### **I'd love your feedbackβ€”have you tried this? How does it perform for you?** + + + + +--- + + + Click here to get info on choosing the right GGUF model format + + +--- + + + + + + +πŸ“ [Paper](https://arxiv.org/abs/2506.07636) | 🌐 [Github](https://github.com/THUDM/SWE-Dev/) + +- πŸ€— [SWE-Dev-7B (Qwen-2.5-Coder-7B-Instruct)](https://huggingface.co/THUDM/SWE-Dev-7B/) +- πŸ€— [SWE-Dev-9B (GLM-4-9B-Chat)](https://huggingface.co/THUDM/SWE-Dev-9B/) +- πŸ€— [SWE-Dev-32B (Qwen-2.5-Coder-32B-Instruct)](https://huggingface.co/THUDM/SWE-Dev-32B/) +- πŸ€— [SWE-Dev-train (Training Data)](https://huggingface.co/datasets/THUDM/SWE-Dev-train/) + +πŸš€ SWE-Dev, an open-source Agent for Software Engineering tasks! This repository contains the SWE-Dev-32B model as presented in the paper [SWE-Dev: Building Software Engineering Agents with Training and Inference Scaling](https://huggingface.co/papers/2506.07636). + +πŸ’‘ We develop a comprehensive pipeline for creating developer-oriented datasets from GitHub repositories, including issue tracking, code localization, test case generation, and evaluation. + +πŸ”§ Based on open-source frameworks (OpenHands) and models, SWE-Dev-7B and 32B achieved solve rates of 23.4% and 36.6% on SWE-bench-Verified, respectively, even approaching the performance of GPT-4o. + +πŸ“š We find that training data scaling and inference scaling can both effectively boost the performance of models on SWE-bench. Moreover, higher data quality further improves this trend when combined with reinforcement fine-tuning (RFT). For inference scaling specifically, the solve rate on SWE-Dev increased from 34.0% at 30 rounds to 36.6% at 75 rounds. + + + +--- + +# πŸš€ If you find these models useful + +Help me test my **AI-Powered Quantum Network Monitor Assistant** with **quantum-ready security checks**: + +πŸ‘‰ [Quantum Network Monitor](https://readyforquantum.com/?assistant=open&utm_source=huggingface&utm_medium=referral&utm_campaign=huggingface_repo_readme) + + +The full Open Source Code for the Quantum Network Monitor Service available at my github repos ( repos with NetworkMonitor in the name) : [Source Code Quantum Network Monitor](https://github.com/Mungert69). You will also find the code I use to quantize the models if you want to do it yourself [GGUFModelBuilder](https://github.com/Mungert69/GGUFModelBuilder) + +πŸ’¬ **How to test**: + Choose an **AI assistant type**: + - `TurboLLM` (GPT-4.1-mini) + - `HugLLM` (Hugginface Open-source models) + - `TestLLM` (Experimental CPU-only) + +### **What I’m Testing** +I’m pushing the limits of **small open-source models for AI network monitoring**, specifically: +- **Function calling** against live network services +- **How small can a model go** while still handling: + - Automated **Nmap security scans** + - **Quantum-readiness checks** + - **Network Monitoring tasks** + +🟑 **TestLLM** – Current experimental model (llama.cpp on 2 CPU threads on huggingface docker space): +- βœ… **Zero-configuration setup** +- ⏳ 30s load time (slow inference but **no API costs**) . No token limited as the cost is low. +- πŸ”§ **Help wanted!** If you’re into **edge-device AI**, let’s collaborate! + +### **Other Assistants** +🟒 **TurboLLM** – Uses **gpt-4.1-mini** : +- **It performs very well but unfortunatly OpenAI charges per token. For this reason tokens usage is limited. +- **Create custom cmd processors to run .net code on Quantum Network Monitor Agents** +- **Real-time network diagnostics and monitoring** +- **Security Audits** +- **Penetration testing** (Nmap/Metasploit) + +πŸ”΅ **HugLLM** – Latest Open-source models: +- 🌐 Runs on Hugging Face Inference API. Performs pretty well using the lastest models hosted on Novita. + +### πŸ’‘ **Example commands you could test**: +1. `"Give me info on my websites SSL certificate"` +2. `"Check if my server is using quantum safe encyption for communication"` +3. `"Run a comprehensive security audit on my server"` +4. '"Create a cmd processor to .. (what ever you want)" Note you need to install a [Quantum Network Monitor Agent](https://readyforquantum.com/Download/?utm_source=huggingface&utm_medium=referral&utm_campaign=huggingface_repo_readme) to run the .net code on. This is a very flexible and powerful feature. Use with caution! + +### Final Word + +I fund the servers used to create these model files, run the Quantum Network Monitor service, and pay for inference from Novita and OpenAIβ€”all out of my own pocket. All the code behind the model creation and the Quantum Network Monitor project is [open source](https://github.com/Mungert69). Feel free to use whatever you find helpful. + +If you appreciate the work, please consider [buying me a coffee](https://www.buymeacoffee.com/mahadeva) β˜•. Your support helps cover service costs and allows me to raise token limits for everyone. + +I'm also open to job opportunities or sponsorship. + +Thank you! 😊 diff --git a/SWE-Dev-32B-bf16_q8_0.gguf b/SWE-Dev-32B-bf16_q8_0.gguf new file mode 100644 index 0000000..56c4129 --- /dev/null +++ b/SWE-Dev-32B-bf16_q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32692e18ff299af2f678fd2ed18b64ff189b35903d22ea8afb3f8a9aa64a7117 +size 46661601888 diff --git a/SWE-Dev-32B-f16_q8_0.gguf b/SWE-Dev-32B-f16_q8_0.gguf new file mode 100644 index 0000000..614e232 --- /dev/null +++ b/SWE-Dev-32B-f16_q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c10806227c3a8f965707c10c5d919d01845c9f08055885689c84a9e2746ad05d +size 46661601888 diff --git a/SWE-Dev-32B-iq1_m.gguf b/SWE-Dev-32B-iq1_m.gguf new file mode 100644 index 0000000..a53d4af --- /dev/null +++ b/SWE-Dev-32B-iq1_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0fecf26aa92f4a1d361b3bc44a403dd9b29282b448e40fcaa9a2dd9f0b287b1e +size 9742306208 diff --git a/SWE-Dev-32B-iq1_s.gguf b/SWE-Dev-32B-iq1_s.gguf new file mode 100644 index 0000000..25c97e2 --- /dev/null +++ b/SWE-Dev-32B-iq1_s.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0266b3b05ef43db6e201924d9e1fd2c54ec5fb84dfde6902a03d0817939344f0 +size 8992492448 diff --git a/SWE-Dev-32B-iq2_m.gguf b/SWE-Dev-32B-iq2_m.gguf new file mode 100644 index 0000000..acbe83d --- /dev/null +++ b/SWE-Dev-32B-iq2_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df8287ce37ae93f398584651d6c16d52d324eab8142e6ea74e3b080cf06b4421 +size 11985849248 diff --git a/SWE-Dev-32B-iq2_s.gguf b/SWE-Dev-32B-iq2_s.gguf new file mode 100644 index 0000000..cc78866 --- /dev/null +++ b/SWE-Dev-32B-iq2_s.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b725e674efd3394cd6ff0828fd959702aab44c89a10b48afbe5e302e2cb65d2 +size 11382262688 diff --git a/SWE-Dev-32B-iq2_xs.gguf b/SWE-Dev-32B-iq2_xs.gguf new file mode 100644 index 0000000..0484de5 --- /dev/null +++ b/SWE-Dev-32B-iq2_xs.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99ec87b12a839f2ac5376950fc415f161e3fe9c67c4d5b91d5489dbb0c4ebeb9 +size 11016326048 diff --git a/SWE-Dev-32B-iq2_xxs.gguf b/SWE-Dev-32B-iq2_xxs.gguf new file mode 100644 index 0000000..c63f439 --- /dev/null +++ b/SWE-Dev-32B-iq2_xxs.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7d11619a2f5d5345dc11b4e18a10bbe73a7e4fc4c98caffb691cb703d1ee588 +size 10124954528 diff --git a/SWE-Dev-32B-iq3_m.gguf b/SWE-Dev-32B-iq3_m.gguf new file mode 100644 index 0000000..7ca5e76 --- /dev/null +++ b/SWE-Dev-32B-iq3_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f6c4681eb9f981fc6d388f049e71b55e92f23ea71e69eabd76d48837b8d53449 +size 15545355168 diff --git a/SWE-Dev-32B-iq3_s.gguf b/SWE-Dev-32B-iq3_s.gguf new file mode 100644 index 0000000..022212d --- /dev/null +++ b/SWE-Dev-32B-iq3_s.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34b3ddaa06b41383dd58e80811f612a790b4f436881f494e5d32c136f54b1c7a +size 15545355168 diff --git a/SWE-Dev-32B-iq3_xs.gguf b/SWE-Dev-32B-iq3_xs.gguf new file mode 100644 index 0000000..1ff7166 --- /dev/null +++ b/SWE-Dev-32B-iq3_xs.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3a7ecd04b88b4e93d2a8be7e0202f410f31ad699848f2167eb020d9f6d9be776 +size 13967002528 diff --git a/SWE-Dev-32B-iq3_xxs.gguf b/SWE-Dev-32B-iq3_xxs.gguf new file mode 100644 index 0000000..f3713be --- /dev/null +++ b/SWE-Dev-32B-iq3_xxs.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94b8298b3b3ffbb1471d86ed8242e37652aadad692d879f33f6743b67112c482 +size 13628836768 diff --git a/SWE-Dev-32B-iq4_nl.gguf b/SWE-Dev-32B-iq4_nl.gguf new file mode 100644 index 0000000..95f1856 --- /dev/null +++ b/SWE-Dev-32B-iq4_nl.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:32afed051904c16b07b7c9659449ab78080f4bac92262d18974e93253281aee9 +size 18682174368 diff --git a/SWE-Dev-32B-iq4_xs.gguf b/SWE-Dev-32B-iq4_xs.gguf new file mode 100644 index 0000000..4a4faac --- /dev/null +++ b/SWE-Dev-32B-iq4_xs.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:35ff85f3cc07305922b605328e4f4ee231147c972d1be9491c1bfc4938d47ff6 +size 17693154208 diff --git a/SWE-Dev-32B-q2_k_m.gguf b/SWE-Dev-32B-q2_k_m.gguf new file mode 100644 index 0000000..abe9bb9 --- /dev/null +++ b/SWE-Dev-32B-q2_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69178b9aa436c50d8acdd04367eae99b56ecc396d59dbede5eb6acacfe108982 +size 12590295968 diff --git a/SWE-Dev-32B-q2_k_s.gguf b/SWE-Dev-32B-q2_k_s.gguf new file mode 100644 index 0000000..37c7d12 --- /dev/null +++ b/SWE-Dev-32B-q2_k_s.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3bb375c9ca8e2de87ee8cd8cd0c6af1a22f4dee3d55f0395f3b1885ab7f6094 +size 12383488928 diff --git a/SWE-Dev-32B-q3_k_m.gguf b/SWE-Dev-32B-q3_k_m.gguf new file mode 100644 index 0000000..ea194e8 --- /dev/null +++ b/SWE-Dev-32B-q3_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b9c82f0ff9f5e5cf1ecb1fd3aba0facd45d018165ea6c14cda8d89fa9c0c2e2d +size 16126127008 diff --git a/SWE-Dev-32B-q3_k_s.gguf b/SWE-Dev-32B-q3_k_s.gguf new file mode 100644 index 0000000..9375aec --- /dev/null +++ b/SWE-Dev-32B-q3_k_s.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1d5ad6c739d75dd706c4da4a7c5c4aa8aa06b303ff5852ab48cf21936bebf021 +size 15919319968 diff --git a/SWE-Dev-32B-q4_0.gguf b/SWE-Dev-32B-q4_0.gguf new file mode 100644 index 0000000..2a2e617 --- /dev/null +++ b/SWE-Dev-32B-q4_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25a91b39798392e29216f85a47db74d3546c1dab1acb18a7c9c59fad949bced9 +size 18439506848 diff --git a/SWE-Dev-32B-q4_1.gguf b/SWE-Dev-32B-q4_1.gguf new file mode 100644 index 0000000..cdd32c0 --- /dev/null +++ b/SWE-Dev-32B-q4_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4beddef8347978ffa417ee7c4180a258ad3150a802f422edbb6c6e9bd48f9078 +size 20487179168 diff --git a/SWE-Dev-32B-q4_k_m.gguf b/SWE-Dev-32B-q4_k_m.gguf new file mode 100644 index 0000000..c0e821f --- /dev/null +++ b/SWE-Dev-32B-q4_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4ebf59bd3c46c1a03fe18340fd6c06d95e01d7671f5a209bce97832aec45fb65 +size 19824978848 diff --git a/SWE-Dev-32B-q4_k_s.gguf b/SWE-Dev-32B-q4_k_s.gguf new file mode 100644 index 0000000..015b688 --- /dev/null +++ b/SWE-Dev-32B-q4_k_s.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:db62de42a9f85359bfc0c6cf9f9909707f9e4e86ecb8c24cdff81e5e06ddbbf2 +size 19161426848 diff --git a/SWE-Dev-32B-q5_0.gguf b/SWE-Dev-32B-q5_0.gguf new file mode 100644 index 0000000..e493ebc --- /dev/null +++ b/SWE-Dev-32B-q5_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ca250ed769559f6773628961cfd955ad45113ba1b86878c9a1842c428a69e3de +size 22534851488 diff --git a/SWE-Dev-32B-q5_1.gguf b/SWE-Dev-32B-q5_1.gguf new file mode 100644 index 0000000..0a86a66 --- /dev/null +++ b/SWE-Dev-32B-q5_1.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f1349abd83682dbefac879e382378ad61e40fd65cd2d8c8738f8c9bf452bb36f +size 24582523808 diff --git a/SWE-Dev-32B-q5_k_m.gguf b/SWE-Dev-32B-q5_k_m.gguf new file mode 100644 index 0000000..9104ba8 --- /dev/null +++ b/SWE-Dev-32B-q5_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cfb334089481ec472986e0fbf9e28ca43e95503cea6839ec2c20a83df9894484 +size 23414303648 diff --git a/SWE-Dev-32B-q5_k_s.gguf b/SWE-Dev-32B-q5_k_s.gguf new file mode 100644 index 0000000..025b4df --- /dev/null +++ b/SWE-Dev-32B-q5_k_s.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2342aeae5042714ea12ffa0b607042e4b3a738f7f05fc1492b612a25e4a9437b +size 23063358368 diff --git a/SWE-Dev-32B-q6_k_m.gguf b/SWE-Dev-32B-q6_k_m.gguf new file mode 100644 index 0000000..3086bf0 --- /dev/null +++ b/SWE-Dev-32B-q6_k_m.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2623e42a6949ea1949be633c4732a9dd505bc69919cd28a4532d462f188aa80d +size 26886155168 diff --git a/SWE-Dev-32B-q8_0.gguf b/SWE-Dev-32B-q8_0.gguf new file mode 100644 index 0000000..2141e52 --- /dev/null +++ b/SWE-Dev-32B-q8_0.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dbefb397865aa86f29d7b0c352e4df1211abe5a8bc87c17a10d5f60570d1d55b +size 34820885088 diff --git a/SWE-Dev-32B.imatrix b/SWE-Dev-32B.imatrix new file mode 100644 index 0000000..e66a751 --- /dev/null +++ b/SWE-Dev-32B.imatrix @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b04c9a698920b9f32eab255d2c7e60a6cc1a0a2faa289fce29dbd22ec28da78a +size 14957132 diff --git a/bf16/SWE-Dev-32B-bf16-00001-of-00002.gguf b/bf16/SWE-Dev-32B-bf16-00001-of-00002.gguf new file mode 100644 index 0000000..754ac0b --- /dev/null +++ b/bf16/SWE-Dev-32B-bf16-00001-of-00002.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c4d21e098d32c4e91334bd0b525ec6a5dd262b92c1de21df6c97156fc838a4d4 +size 45902462976 diff --git a/bf16/SWE-Dev-32B-bf16-00002-of-00002.gguf b/bf16/SWE-Dev-32B-bf16-00002-of-00002.gguf new file mode 100644 index 0000000..daaedae --- /dev/null +++ b/bf16/SWE-Dev-32B-bf16-00002-of-00002.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e28cf065e3feda607abb4d70237e407bed95f9b5e86f1f6b3ca90f7486ed9884 +size 19633506912 diff --git a/f16/SWE-Dev-32B-f16-00001-of-00002.gguf b/f16/SWE-Dev-32B-f16-00001-of-00002.gguf new file mode 100644 index 0000000..24532a9 --- /dev/null +++ b/f16/SWE-Dev-32B-f16-00001-of-00002.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5e235d03a4780ff1b7dd759884857abd30a04ae44692e4fa2fd000ba0b81ed5 +size 45902462976 diff --git a/f16/SWE-Dev-32B-f16-00002-of-00002.gguf b/f16/SWE-Dev-32B-f16-00002-of-00002.gguf new file mode 100644 index 0000000..8ff8e2d --- /dev/null +++ b/f16/SWE-Dev-32B-f16-00002-of-00002.gguf @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38c2baf26f9d06c7e70083870d12b27052410296911d241c1f1ee994403c0748 +size 19633506912