← Back to Home Legitimacy & Evidence

What Protective Computing Is Not

Protective Computing becomes easier to trust when its exclusions are explicit. This page narrows the boundary so outsiders can tell the difference between the discipline itself and adjacent categories it may overlap with.

Not generic privacy

Privacy matters, but Protective Computing is not a synonym for privacy-first software. It is concerned with whether software remains survivable, bounded, and reversible when the user is under pressure, offline, cognitively overloaded, coerced, or operating through degraded hardware.

Not cybersecurity branding

Security controls are necessary but insufficient. A system can be cryptographically strong and still fail under forced disclosure, network loss, inaccessible controls, or extraction-oriented business logic. Protective Computing evaluates those failure modes directly.

Not trauma aesthetics

The discipline is not a tone, visual style, empathy language layer, or soft UX posture. It requires concrete defaults, verifiable constraints, and evidence-backed claims. If the underlying behavior is unsafe, soft copy does not count.

Not wellness software, and not only for health apps

Health is an obvious use case because the harms are legible there, but the discipline applies anywhere systems handle sensitive records, constrained users, coercive environments, or unstable infrastructure. The frame is operational, not sector-specific.

Not compliance theater

Protective Computing uses standards crosswalks and audit artifacts, but it does not treat checklists as legitimacy by themselves. A claim is only as strong as the runtime behavior and evidence supporting it.

Not a replacement for NIST, ISO, SOC 2, or OWASP

Those frameworks remain important. Protective Computing sits above and across them as a human-failure design lens, then translates that lens into architecture, defaults, evidence, and threat boundaries those frameworks can consume.

Not only for “vulnerable users”

The phrase can become patronizing if used carelessly. Protective Computing assumes instability is normal: everyone can lose connectivity, access, cognition, safety, or institutional leverage. The discipline designs for those conditions without requiring users to self-identify into a special class first.

What it is instead

Protective Computing is design under instability. It asks whether a system preserves autonomy, bounded disclosure, reversibility, and essential function when ideal conditions fail.

If a product claim cannot survive degraded conditions, coercion scenarios, and reproducible audit review, it is outside the discipline even if it uses the language.