Mohsin et al. 2024 – Human-AI Collaboration in SOCs

Citation

Mohsin, A., Janicke, H., Ibrahim, A., Sarker, I. H., & Camtepe, S. (2024). A Unified Framework for Human–AI Collaboration in Security Operations Centers with Trusted Autonomy. arXiv:2505.23397.

Core Contribution

A structured framework for human-AI collaboration in Security Operations Centers that integrates AI autonomy, trust calibration, and human-in-the-loop decision-making through a five-level autonomy taxonomy mapped to SOC analyst tiers.

Framing

The authors argue that existing SOC frameworks treat automation as binary rather than graduated, lacking systematic structures to manage human oversight, trust calibration, and scalable autonomy. Their key insight: effective human-AI collaboration requires dynamic calibration across three interconnected dimensions, autonomy, trust, and human involvement, that adapt to task complexity and risk.

Key Concepts

The Triadic Model

Three core pillars of human-AI collaboration:

  1. Autonomy (A) – AI’s ability to act independently
  2. Human-in-the-Loop (H) – Degree of human oversight
  3. Trust (T) – Confidence in AI’s reliability and transparency

Formal Definitions

Autonomy: A = 1 − (λ₁C + λ₂R)(1 − T)

  • C = task complexity
  • R = risk
  • T = trust
  • Higher complexity/risk reduces autonomy; higher trust increases it

HITL: H = 1 − A

  • Inverse relationship: as autonomy increases, human involvement decreases

Trust: T = α₁E + α₂P + α₃(1 − U)

  • E = explainability
  • P = performance history
  • U = uncertainty

Five Autonomy Levels (adapted from SAE J3016)

LevelDescriptionA Range
0Manual operations≈0
1AI-assisted (decision support)0.2–0.4
2Semi-autonomous (AI acts with approval)0.4–0.6
3Conditionally autonomous (HITL)0.7–0.8
4Fully autonomous (minimal oversight)0.9–1.0

Three HITL Configurations

  1. Human-in-Control – Full HITL; humans make all decisions
  2. Human-on-the-Loop (HOtL) – AI executes under conditional oversight
  3. Human-out-of-the-Loop (HOoTL) – AI autonomous; humans audit/govern

Transferable Insights

  1. Graduated autonomy beats binary automation – Most frameworks assume on/off; real collaboration requires a spectrum
  2. Trust mediates autonomy – You can’t scale autonomy without building trust through demonstrated performance and transparency
  3. Inverse relationship formalization – H = 1 − A elegantly captures the tradeoff between human control and AI independence
  4. Complexity and risk constrain autonomy – Even with high trust, high-complexity/high-risk tasks require human involvement
  5. Trust calibration is dynamic – Trust should evolve through interaction, not be assumed

Case Study Findings

The ACDC CyberAlly deployment (LLM-based SOC assistant) showed:

  • 50% reduction in false positives
  • 67% faster investigations
  • 80% reduction in mean time to respond
  • Graduated trust: analysts initially verified everything, gradually accepted AI autonomy