ARTFEED — Contemporary Art Intelligence

CyberSecQwen-4B: A Small, Specialized AI Model for Defensive Cybersecurity

ai-technology · 2026-05-08

CyberSecQwen-4B is a 4-billion-parameter language model fine-tuned for narrow cybersecurity tasks like CWE classification and CTI Q&A. Developed by lablab-ai in the AMD Developer Hackathon, it runs on a single 12 GB consumer GPU, avoiding the cost and data exposure of hosted APIs. The model retains 97.3% of Cisco's 8B Foundation-Sec-Instruct's CTI-RCM accuracy while exceeding its CTI-MCQ score by 8.7 points. Training used MITRE/NVD CVE-to-CWE mappings and synthetic Q&A, deduplicated against CTI-Bench. The base model is Qwen3-4B-Instruct-2507, fine-tuned on a single AMD MI300X via ROCm 7. A companion 2B model, Gemma4Defense-2B, shows similar performance. The model is Apache 2.0 licensed and available on Hugging Face. It is designed for local, air-gapped environments and explicitly not for generating exploit code or making autonomous security decisions.

Key facts

  • CyberSecQwen-4B is a 4B-parameter model for defensive cybersecurity tasks.
  • It runs on a single 12 GB consumer GPU.
  • Retains 97.3% of Cisco's 8B model's CTI-RCM accuracy.
  • Exceeds Cisco's 8B model on CTI-MCQ by 8.7 points.
  • Trained on MITRE/NVD CVE-to-CWE mappings and synthetic Q&A.
  • Base model is Qwen3-4B-Instruct-2507.
  • Companion model Gemma4Defense-2B shows similar performance.
  • Model is Apache 2.0 licensed.

Entities

Institutions

  • Cisco
  • MITRE
  • NVD
  • AMD
  • Hugging Face
  • lablab-ai
  • AMD Developer Cloud

Sources