Why zkML? Because @AnthropicAI just disclosed the first recorded large-scale cyberattack orchestrated primarily by AI agents — with Claude executing 80–90% of the operation autonomously.
When AI stops advising and starts acting, the verification gap becomes an attack surface.

2/
The threat actor jailbroke Claude, disguised the operation as benign testing, and had the model:
- probe infrastructure
- identify high-value systems
- write exploit code
- harvest credentials
- exfiltrate data
All chained together through autonomous loops with minimal human supervision.
This wasn’t prompt misuse.
This was agentic execution.
3/
The core problem isn’t capability — it’s opacity.
These attacks succeeded because:
- reasoning was invisible
- tool use was unverified
- policy compliance couldn’t be proven
- execution traces couldn’t be audited in real time
When AI becomes the operator, lack of verifiability becomes the vulnerability.
4/
That’s where zkML changes the security model:
✅Prove the model followed the intended reasoning path
✅Prove tool calls matched declared policies
✅Prove execution stayed within allowed boundaries
✅ Enable auditors to verify behavior without accessing model internals
Agents don’t just need guardrails — they need proof rails.
5/
Cybersecurity has entered its post-human phase.
When AI conducts operations end-to-end, proof must replace assumption at the execution layer.
That’s what @PolyhedraZK is building: intelligence you can verify, even when the agent runs the mission.
2,256
17
本頁面內容由第三方提供。除非另有說明,OKX 不是所引用文章的作者,也不對此類材料主張任何版權。該內容僅供參考,並不代表 OKX 觀點,不作為任何形式的認可,也不應被視為投資建議或購買或出售數字資產的招攬。在使用生成式人工智能提供摘要或其他信息的情況下,此類人工智能生成的內容可能不準確或不一致。請閱讀鏈接文章,瞭解更多詳情和信息。OKX 不對第三方網站上的內容負責。包含穩定幣、NFTs 等在內的數字資產涉及較高程度的風險,其價值可能會產生較大波動。請根據自身財務狀況,仔細考慮交易或持有數字資產是否適合您。

