Blockchain security firm CertiK has issued a stark warning about the security models of emerging AI agent marketplaces, revealing that current pre-deployment review systems are insufficient to prevent malicious attacks. The firm's research demonstrates how a compromised third-party "Skill" on the OpenClaw platform could bypass multiple layers of security checks and execute arbitrary commands on a host system.
The proof-of-concept attack targeted the review pipeline of Clawhub, OpenClaw's marketplace, which employs static code analysis, AI-based moderation, and VirusTotal checks. CertiK researchers, led by Guanxing Wen, found that relatively minor code modifications or restructuring of logic could allow a malicious Skill to appear benign during installation while retaining harmful capabilities once deployed. This creates a false sense of security for users, as marketplace approval does not guarantee safety.
The core issue identified is a structural weakness across the industry: security models that rely heavily on pre-deployment detection rather than runtime containment. CertiK argues that as these marketplaces expand, the risk of malicious Skills entering production environments will increase. The firm emphasizes that detection systems, including AI moderation, are not designed to handle complex, evolving threats and can be circumvented.
CertiK's key recommendation is a fundamental shift in design philosophy. Instead of striving for "perfect detection," platforms must prioritize damage containment and system resilience. This involves adopting sandboxing as the default execution model for third-party Skills, ensuring they run in isolated environments. The report also calls for granular, per-Skill permission frameworks where each Skill explicitly declares needed resources, with the runtime enforcing those limits.
The implications are significant for both users and platform developers. For users, a "benign" label in a marketplace is not a guarantee of security. Until stronger runtime protections are adopted, platforms like OpenClaw may be unsuitable for handling sensitive data or high-value assets. For the broader AI and crypto ecosystem, the ability to contain risks at runtime is positioned as a defining factor for securing next-generation digital applications.