Google’s Gemma model, which emphasizes alignment and safety protocols, shows a more balanced profile with a smaller security gap. This highlights the importance of prioritizing security and alignment in AI development, rather than solely focusing on capabilities.
The implications of this research extend beyond individual models to the broader AI ecosystem. As more enterprises adopt open-weight models for a range of applications, from customer service chatbots to autonomous vehicles, understanding and addressing vulnerabilities like those identified by Cisco becomes imperative.
Ultimately, the research underscores the need for a more holistic approach to AI security. It’s not enough to focus solely on single-turn attacks or individual vulnerabilities. Enterprises must consider the broader context of conversational persistence and the potential for multi-turn attacks to bypass safeguards.
As DJ Sampath of Cisco emphasizes, open-source models offer significant benefits in terms of flexibility and customization. However, these advantages come with security risks that must be carefully managed. By understanding the vulnerabilities inherent in open-weight models and implementing appropriate guardrails, enterprises can mitigate the risk of catastrophic failures under sustained adversarial pressure.
In a rapidly evolving threat landscape, staying ahead of potential attacks requires a proactive approach to AI security. By incorporating insights from research like Cisco’s study on open-weight model vulnerabilities, enterprises can better protect their AI systems and ensure the resilience of their operations in the face of sophisticated threats. In a recent statement, Google’s Gemma highlighted the importance of implementing “rigorous safety protocols” in AI systems to target a “low risk level” for misuse. The result of this focus on safety protocols led to the lowest gap at 10.53%, with a more balanced performance observed across both single- and multi-turn scenarios.
It is important to note that models optimized for capability and flexibility often come with less built-in safety measures. While this design choice may be suitable for many enterprise use cases, it is crucial for businesses to understand that prioritizing capability over security may leave them vulnerable to attacks. Therefore, it is essential for enterprises to allocate budget resources accordingly to address potential security risks.
A study conducted by Cisco tested 102 distinct subthreat categories, revealing that the top 15 categories exhibited high success rates across all models. This suggests that implementing targeted defensive measures could lead to significant improvements in security posture.
The research emphasizes that security should be viewed as a key enabler of AI adoption rather than an obstacle. By incorporating the right security measures, enterprises can unlock productivity for their users and facilitate the widespread adoption of AI tools. This approach ensures that organizations can harness the power of AI technologies while maintaining a strong security posture to prevent breaches or data leaks.
To enhance security in AI systems, enterprises should prioritize six critical capabilities, including context-aware guardrails, model-agnostic runtime protections, continuous red-teaming exercises, hardened system prompts, comprehensive logging for forensic visibility, and threat-specific mitigations for the top 15 subthreat categories identified in the research.
Sampath, a security expert, warns against waiting for AI to stabilize, urging businesses to take immediate action to strengthen their security defenses. The research findings highlight the importance of addressing vulnerabilities in multi-turn attacks, model-specific weaknesses, and high-risk threat patterns promptly to mitigate potential risks.
In conclusion, enterprises must shift their focus from testing single-turn defenses to securing entire conversations to effectively safeguard AI systems from potential threats. By implementing robust security measures and staying proactive in addressing vulnerabilities, businesses can ensure the safe and efficient adoption of AI technologies in their operations.

