license, language, metrics, base_model, new_version, pipeline_tag, library_name, tags
license language metrics base_model new_version pipeline_tag library_name tags
apache-2.0
en
code_eval
accuracy
Qwen/Qwen2.5-Coder-1.5B-Instruct
Qwen/Qwen2.5-Coder-1.5B-Instruct text-generation transformers
cybersecurity
mythos
qween
qween-security
blue
team
blue-team
cve
ctf
code
code-security

language:

  • en
  • code license: apache-2.0 tags:
  • security
  • exploit-development
  • vulnerability-research
  • php
  • mybb
  • cve
  • python
  • qwen
  • fine-tuned
  • cybersecurity datasets:
  • [your-dataset-name-if-uploaded] metrics:
  • accuracy
  • code-eval pipeline_tag: text-generation library_name: transformers base_model: Qwen/Qwen2.5-Coder-1.5B-Instruct

Mythos Engine - Qwen 2.5 Coder 1.5B Security Fine-Tune

🔥 Model Description

Mythos Engine is a specialized fine-tune of Qwen 2.5 Coder 1.5B Instruct designed for cybersecurity research, vulnerability analysis, and exploit development. It has been trained on a curated dataset of 700+ high-reasoning security examples covering PHP internals, MyBB exploitation, deserialization chains, type juggling, and advanced Python exploit synthesis.

The model employs Chain-of-Thought reasoning with self-correction loops and mathematical logic notation to produce accurate, production-ready security code.

🎯 Intended Use

  • Security Research: Analyzing CVEs and understanding exploit mechanics
  • Red Team Education: Learning exploit development patterns
  • Blue Team Defense: Understanding attack vectors to build better detections
  • CTF & Training: Solving complex security challenges

⚠️ Important: This model is for educational and authorized security testing only. Do not use for unauthorized access or malicious purposes.

🧠 Training Details

Aspect Details
Base Model Qwen/Qwen2.5-Coder-1.5B-Instruct
Fine-Tuning Method QLoRA (4-bit quantization) with Unsloth
Dataset Size 1000+ examples
Epochs 4
Learning Rate 1e-5
Sequence Length 4096
Final Training Loss 2.02

📊 Dataset Composition

The training dataset includes:

  • 40% PHP Vulnerabilities: Type juggling, deserialization, filter chains, disable_functions bypasses
  • 25% MyBB Exploits: Admin CP RCE, SQL injection, XSS chains
  • 20% Python Exploit Development: C2 frameworks, scanners, injection techniques
  • 10% Blue Team Detection: Sigma/YARA rules, log analysis
  • 5% Cryptographic Attacks: Timing attacks, padding oracles, hash length extension

🚀 How to Use

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained(
    "expper/mythos-qwen-1.5b-final",
    device_map="auto",
    torch_dtype="auto"
)
tokenizer = AutoTokenizer.from_pretrained("expper/mythos-qwen-1.5b-final")

prompt = """<|im_start|>system
You are Mythos Engine, an elite security AI. Think step-by-step with self-correction.<|im_end|>
<|im_start|>user
Explain CVE-2022-43772 (MyBB Admin CP Avatar RCE) and write a PoC.<|im_end|>
<|im_start|>assistant
"""

inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=1024, temperature=0.6)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Description
Model synced from source: expper/mythos-qwen-1.5b-final
Readme 2 MiB
Languages
Jinja 100%