About PROACTIVE

Inspiration

PROACTIVE was built in response to a recurring failure pattern observed across multiple AI systems: models producing confident, unsupported claims, then adapting those claims under user pressure rather than maintaining truth.

These failures were not isolated to a single system or environment. They appeared consistently across different models and workflows, with varying degrees of severity. In each case, the pattern was the same: fluency and agreement were prioritized over epistemic integrity.

PROACTIVE is designed to prevent that class of failure.


What It Does

PROACTIVE is a constitutional AI safety system that enforces epistemic integrity at the point of generation.

Instead of evaluating outputs after they are produced, it applies six invariants in real time:

  • I1 — Evidence-first reasoning
  • I2 — No phantom work
  • I3 — Confidence requires verification
  • I4 — Traceability
  • I5 — Safety over fluency
  • I6 — Fail-closed behavior

If a violation is detected, the system blocks the output.


How It Works

The system is implemented as a deterministic pipeline:

$$ \text{Intent} \rightarrow \text{Contract} \rightarrow \text{Validation} \rightarrow \text{Drift Detection} \rightarrow \text{Report} $$

  • Cognitive Modeling Protocol (CMP) parses intent from input
  • Contract Window establishes constraints and risk
  • Validator enforces I1–I6 invariants
  • Drift Detector compares intent to actual implementation
  • Report Formatter produces structured output with a V&T statement

The system is integrated with GitLab CI/CD and exposed as a Duo agent, allowing it to review merge requests in real time.


What I Learned

  • This failure pattern is systemic, not model-specific
  • Evaluation alone does not prevent harm — enforcement is required
  • Truth must be explicitly structured and verified, not implied
  • Safety must be encoded as system behavior, not guidance

Challenges

The primary challenge was enforcing strict constraints without blocking legitimate work.

  • Distinguishing incomplete from incorrect
  • Handling ambiguity without defaulting to false certainty
  • Designing fallback behavior when LLM access is unavailable
  • Maintaining alignment between intent and implementation

Another challenge was avoiding overclaiming. The system explicitly separates what is verified from what is not through V&T statements.


Why It Matters

PROACTIVE addresses one instance of a broader problem.

It is a working component within a larger system, The Living Constitution (TLC), which extends these principles to multi-agent governance, feedback loops, and adaptive safety rules.

Together, they move AI safety from evaluation toward enforceable system design.

Built With

  • anthropic-api-(claude)
  • click-(cli)
  • docker-ready-architecture
  • flask
  • gitlab-ci/cd
  • gitlab-duo-agent-platform
  • pydantic
  • pytest
  • python
  • requests
  • tf-idf
  • tf-idf-(scikit-learn)
Share this project:

Updates