Cybersecurity AI (CAI) source code (technical report)
Cybersecurity AI Community Meeting 2 Friday, June 20 3:30 PM CEST (link)

Why we built alias0

The GenAI boom alongside the spectacular rise of LLM-powered “security copilots” and AI Security has a hidden cost: nearly 100 % of commercial offerings proxy user prompts to US- or China-based foundation models, effectively gifting away customers’ sensitive data and security know-how. That means:

  • Loss of data sovereignty – every artifact (API keys, passwords, hashes, IPs, source snippets, even org charts) becomes training fodder for someone else’s model.
  • Opaque supply-chains – users inherit the entire external inference stack (telemetry, logging, third-party subcontractors) as an implicit dependency.
  • Regulatory friction – GDPR, NIS2, the AI Act, and upcoming eIDAS 2 create explicit liabilities for uncontrolled data transfers.

European companies and governments need an alternative that keeps data sovereign, private, and under their direct control — without trading off capability. Europe needs data ownership and an on-prem-capable, regulation-aware, and adversarially strong Cybersecurity AI and related LLM alternatives.

alias0 delivers exactly that:

Value Proposition Technical Manifestation Impact
Full autonomy End-to-end exploit synthesis, patch generation, CTI summarisation ↓ MTTR, ↓ analyst fatigue
Zero private data egress All data can remain on-prem; nothing is sent to external LLM APIs. Meets data-residency and export-control rules
European roots R&D, hosting, and support based in the EU; supply-chain attestation available Simplifies compliance audits (GDPR, ENISA CSP-cert, etc.)
alias0, a Privacy-First Cybersecurity AI

Privacy-by-Design: We can’t lose what we never know

alias0 internalises the privacy budget concept from differential-privacy literature and couples it with a dedicated PrivacyLLM that anonymises every request before it reaches the security brain and re-identifies only the minimal context required for the answer:

┌──────────┐         ┌────────────┐         ┌──────────────────┐
│  User    │  Q_raw  │ PrivacyLLM │ Q_token │ CybersecurityLLM │ A_action
│  Query   ├────────►  Anonymise  ├─────────►   (reason)       ├───────┐
└──────────┘         └────────────┘         └──────────────────┘       │
                                                                       ▼
                                                            ┌──────────────┐
                                                            │ PrivacyLLM   │
                                                            │ Re-identify  │
                                                            └──────┬───────┘
                                                                   │
                                                                   ▼
                                                   ┌──────────────────┐
                                                   │  Cybersecurity   │
                                                   │      action      │
                                                   └──────────────────┘
  • Q_rawRaw Prompt: Contains unrestricted context (logs, PCAPs, firmware, etc.).
  • Q_tokenPolicy-Clean Prompt: Output of PrivacyLLM’s anonymiser; structurally identical but semantically neutralised.
  • A_actionCyber-Security Action: CybersecurityLLM emits directly executable artefacts
  • Re-identification: PrivacyLLM subsystem performs a “token unwrap”, reinserting redacted data only where essential to lead to Cybersecurity actions.

Because raw customer data never leaves the sandbox, risk of leakage is mathematically minimise, leading to:

  • Anonymisation Every identifier (asset tag, IP range, user handle) is replaced by a deterministic but non-reversible token using format-preserving encryption.

  • Policy enforcement – The PrivacyLLM embeds DLP patterns, GDPR art. 6/32 rules, and org-specific regex policies so prohibited content is blocked before inference.

  • Mathematical assurance – Because the CybersecurityLLM never receives raw secrets, exfiltration probability ≈ 0 under the honest-but-curious threat model.


Model-of-Models Intelligence — Two Brains, One Mission

Traditional “one-size-fits-all” LLMs dilute capacity across many unrelated domains. alias0 instead follows a Model-of-Models (MoM) pattern, where specialised sub-systems each master a coherent slice of the problem space and collaborate via a tightly defined interface.

alias0 embodies the Model-of-Models philosophy: instead of one monolithic model trying to do everything, we combine multiple, purpose-built intelligences that collaborate seamlessly.

Today that collaboration happens between two large-language-model blocks:

Model Block Core Mission What It Knows
PrivacyLLM Strip or substitute personally-identifiable data, enforce organisational policy, and certify that only abstracted context proceeds downstream. Taxonomies of PII, privacy regulations, data-loss prevention patterns, organisational policy rules.
CybersecurityLLM Generate, validate, and execute security reasoning and artefacts (e.g., exploits, hardening scripts, threat-intel briefs). Offensive tradecraft, defensive best practices, CTI knowledge graphs, vulnerability corpora, protocol RFCs.

alias0 is only our first step. The same architectural skeleton can host many more specialised intelligences and models which are working on to address more use ases. Each will plug into the existing privacy envelope, forming an orchestrated constellation of narrow experts that, together, outperform any single giant model. This is what we mean by Model-of-Models — a scientific and engineering strategy.


Breaking the Benchmark Ceiling

Benchmark Metric alias0 Best Runner-Up
PentestPerf[1] Offensive Cybersecurity 95 % 70 %
SecEval Cybersecurity Knowledge 78 % 72 %
CTI-MCQ Cyber Threat Intelligence 95 % 75 %

alias0 isn’t just incrementally better — it is ahead in the field of cybersecurity.


  1. An internal benchmarking framework that measures penetration testing capabilities in a proprietary set of IT, OT and robotics scenarios ↩︎


Enterprise & Government Ready

  • Modular deployment — run with all data on our managed cloud, your private cloud, or air-gapped on-prem.
  • Compliance hooks — audit logging, NIS2 aligned controls and GDPR conscious.
  • From Europe, to the world — Need formal assurances? Our enterprise SLA includes 24 × 7 support. Directly from your time-zone.

Transparent Pricing

Plan Price
Enterprise & Goverment €5 / million tokens (in + out)
National Security Custom, reach out

No hidden fees. No vendor lock-in. Reach out to engage.


Get Started in 30 Seconds With CAI

# Install CAIv0.4.0
pip install cai-framework==0.4.0

# Set up alias0
echo -e 'ALIAS_API_KEY="sk-dfse23rwfdfdsfwefdsv"\nCAI_MODEL=alias0\nCAI_AGENT_TYPE=bug_bounter_agent' > .env

# Start vibe-hacking your way into security
CAI> Can you find vulnerabilities in my ******* and help mitigate them?