Habr AI→ original

AI in cryptography: a model with no right to be right

In cryptography, the main question is not whether AI finds the right answer, but whether the system remains secure when AI is wrong. We examine the governed sol

AI in cryptography: a model with no right to be right
Source: Habr AI. Collage: Hamidun News.
◐ Listen to article

When AI is embedded in a cryptographic system, the question often arises: can the model find the correct answer? But in the context of highly reliable systems, this is the wrong question. The right question is different: can AI be embedded in such a way that even if it makes an error, it cannot make a dangerous decision?

Governed Solver Orchestration: An Architecture of Trust

In nonce-observatory, an architecture has been implemented where AI plays a clearly defined role as a planner, but never serves as a source of truth. The model analyzes only publicly available and safe data (public-safe feature contract), suggests execution routes for solvers, and helps with triage and explanation of chosen solutions. But there is a hard boundary — a set of forbidden fields — that the model physically cannot see.

Here is what is available to AI:

  • Analysis of public-safe feature contract
  • Proposal of solver routes and route optimization
  • Construction of execution queues in the correct order
  • Assistance with task classification and explanation of choice
  • Recommendations without assuming final decision-making

And here is what lies beyond the non-escalation boundary:

  • Private nonce fields (critical secret data)
  • Values of candidate_d and k from cryptographic algorithms
  • Formation of recovery claims
  • Conversion of its score or metric into cryptographic evidence

Non-escalation Boundary and Deterministic Verification

The boundary runs between two layers: the AI planner and the deterministic verifier. AI can be wrong in route recommendations — this is non-critical. Why? Because the final decision is always made by an exact verifier that operates by cryptographically verifiable and deterministic rules, without judgment from the model. This approach completely changes the development paradigm.

Typically, highly reliable systems require very intelligent AI that simply must not make mistakes — otherwise everything breaks. Here the strategy is different: AI can be intellectually average. The key constraint is that its error must never automatically become a fact in the system.

"AI suggests.

Exact verifier decides." — not just an architectural pattern, but a philosophy of developing AI in cryptography.

Why This Is Critical for Highly Reliable Systems

Critical systems are often built on the assumption of unambiguous solutions. If AI says "this is correct" and the system accepts it as an axiom — and the model was wrong — the entire protocol and its guarantees are compromised. There is no such risk here.

The logic of operation:

  • The model analyzes only published data
  • Proposes candidate routes for launching solvers
  • The verifier deterministically checks each step
  • Cryptographic proof is formed without AI involvement

The article is especially useful for anyone designing AI in critical systems — not just cryptographic ones. If a model's error should not automatically become a system decision, this is exactly the kind of architectural boundary needed.

What This Means

AI in critical systems is not about the model being perfect and never making mistakes. It is about the correct division of responsibility between components. There are tasks for AI: data analysis, route recommendations, queue optimization. There are tasks only for deterministic verifiers: final decision, cryptographic inference, security guarantees. When this boundary is clear and strictly enforced, the system remains safe even with modest model capabilities.

ZK
Hamidun News
AI news without noise. Daily editorial selection from 400+ sources. A product by Zhemal Khamidun, Head of AI at Alpina Digital.
What do you think?
Loading comments…