A newer version of this model is available: Qwen/Qwen2.5-Coder-1.5B-Instruct

language: - en - code license: apache-2.0 tags: - security - exploit-development - vulnerability-research - php - mybb - cve - python - qwen - fine-tuned - cybersecurity datasets: - [your-dataset-name-if-uploaded] metrics: - accuracy - code-eval pipeline_tag: text-generation library_name: transformers base_model: Qwen/Qwen2.5-Coder-1.5B-Instruct

Mythos Engine - Qwen 2.5 Coder 1.5B Security Fine-Tune

🔥 Model Description

Mythos Engine is a specialized fine-tune of Qwen 2.5 Coder 1.5B Instruct designed for cybersecurity research, vulnerability analysis, and exploit development. It has been trained on a curated dataset of 700+ high-reasoning security examples covering PHP internals, MyBB exploitation, deserialization chains, type juggling, and advanced Python exploit synthesis.

The model employs Chain-of-Thought reasoning with self-correction loops and mathematical logic notation to produce accurate, production-ready security code.

🎯 Intended Use

  • Security Research: Analyzing CVEs and understanding exploit mechanics
  • Red Team Education: Learning exploit development patterns
  • Blue Team Defense: Understanding attack vectors to build better detections
  • CTF & Training: Solving complex security challenges

⚠️ Important: This model is for educational and authorized security testing only. Do not use for unauthorized access or malicious purposes.

🧠 Training Details

Aspect Details
Base Model Qwen/Qwen2.5-Coder-1.5B-Instruct
Fine-Tuning Method QLoRA (4-bit quantization) with Unsloth
Dataset Size 1000+ examples
Epochs 4
Learning Rate 1e-5
Sequence Length 4096
Final Training Loss 2.02

📊 Dataset Composition

The training dataset includes:

  • 40% PHP Vulnerabilities: Type juggling, deserialization, filter chains, disable_functions bypasses
  • 25% MyBB Exploits: Admin CP RCE, SQL injection, XSS chains
  • 20% Python Exploit Development: C2 frameworks, scanners, injection techniques
  • 10% Blue Team Detection: Sigma/YARA rules, log analysis
  • 5% Cryptographic Attacks: Timing attacks, padding oracles, hash length extension

🚀 How to Use

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained(
    "expper/mythos-qwen-1.5b-final",
    device_map="auto",
    torch_dtype="auto"
)
tokenizer = AutoTokenizer.from_pretrained("expper/mythos-qwen-1.5b-final")

prompt = """<|im_start|>system
You are Mythos Engine, an elite security AI. Think step-by-step with self-correction.<|im_end|>
<|im_start|>user
Explain CVE-2022-43772 (MyBB Admin CP Avatar RCE) and write a PoC.<|im_end|>
<|im_start|>assistant
"""

inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=1024, temperature=0.6)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Downloads last month
379
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for expper/mythos-qwen-1.5b-final

Finetuned
(145)
this model
Quantizations
1 model