terminal
howtonotcode.com
Project CodeGuard logo

Project CodeGuard

Term

A code name used to refer to a project or initiative.

article 2 storys calendar_today First seen: 2026-02-10 update Last seen: 2026-02-10 menu_book Wikipedia

Stories

Showing 1-2 of 2

Cisco donates CodeGuard to CoSAI as research exposes persistent LLM code vulnerabilities

Cisco donated its model-agnostic CodeGuard security ruleset to CoSAI while new research shows LLM code generators reliably repeat exploitable patterns, raising the bar for secure-by-default AI coding. OASIS Open details CodeGuard’s coverage and IDE-assistant integrations like Cursor, GitHub Copilot, Windsurf, and Claude Code ([Cisco Donates Project CodeGuard to Coalition for Secure AI](https://www.oasis-open.org/2026/02/09/cisco-donates-project-codeguard-to-coalition-for-secure-ai/)[^1]). Research on “vulnerability persistence” introduces FSTab to predict and exploit recurring flaws in LLM-generated software with high cross-domain success, and domain-focused safety stacks like Guardrails AI are emerging to catch dangerous outputs ([AI Code Generation Tools Repeat Security Flaws](https://quantumzeitgeist.com/ai-security-code-generation-tools-repeat-flaws/)[^2]; [Inside Guardrails AI](https://www.webpronews.com/inside-guardrails-ai-how-a-seattle-startup-is-deploying-clinical-expertise-to-neutralize-the-most-dangerous-failures-in-artificial-intelligence/)[^3]). [^1]: Official announcement of the CodeGuard donation, scope, and integrations with popular AI coding assistants. [^2]: Summarizes FSTab and evidence of predictable, repeatable vulnerabilities (e.g., high success versus Claude‑4.5 Opus). [^3]: Example of domain-specific guardrails and enterprise safety demand context.

calendar_today 2026-02-09
cisco project-codeguard coalition-for-secure-ai-cosai oasis-open cursor

Cisco open-sources CodeGuard as research flags predictable LLM code flaws

Cisco donated its CodeGuard security framework to OASIS’s Coalition for Secure AI as new research shows LLM code assistants repeat predictable vulnerabilities, raising the bar for secure-by-default AI coding workflows. Details of the open donation and integration targets (Cursor, Copilot, Windsurf, Claude Code) are in OASIS Open’s announcement [Cisco Donates Project CodeGuard to Coalition for Secure AI](https://www.oasis-open.org/2026/02/09/cisco-donates-project-codeguard-to-coalition-for-secure-ai/)[^1]. Complementary research findings show vulnerability persistence and a black-box FSTab method with up to 94% attack success on LLM-generated apps [AI Code Generation Tools Repeat Security Flaws, Creating Predictable Software Weaknesses](https://quantumzeitgeist.com/ai-security-code-generation-tools-repeat-flaws/)[^2], with broader context on latent backdoors in “clean” AI code [Backdoors With Manners](https://hackernoon.com/backdoors-with-manners-when-ai-writes-clean-code-that-turns-malicious-later?source=rss)[^3] and sector-specific safety layers emerging in healthcare [Inside Guardrails AI](https://www.webpronews.com/inside-guardrails-ai-how-a-seattle-startup-is-deploying-clinical-expertise-to-neutralize-the-most-dangerous-failures-in-artificial-intelligence/)[^4]. [^1]: Adds: Official details on CodeGuard scope, integrations, and governance via CoSAI. [^2]: Adds: Research summary explaining FSTab, vulnerability recurrence metrics, and attack success rates. [^3]: Adds: Perspective on behavioral trojans and delayed-malicious code patterns. [^4]: Adds: Example of domain-specific safety guardrails in production contexts.

calendar_today 2026-02-09
cisco oasis-open coalition-for-secure-ai-cosai project-codeguard cursor