AI models block 87% of single attacks, but just 8% when attackers persist
Google’s Gemma-3-1B-IT, which focuses on alignment, shows a more balanced profile between single- and multi-turn attack success rates. This indicates a stronger emphasis on rigorous safety protocols and a lower risk level for misuse.
The philosophy of the lab behind the AI model directly impacts the security outcomes of the model itself. Labs that prioritize capabilities may overlook security vulnerabilities, leading to larger security gaps when faced with multi-turn attacks. On the other hand, labs that prioritize alignment and safety protocols tend to have a more balanced profile and lower risk of being exploited.
The implications of this research are significant for enterprises relying on open-weight AI models for various applications. Understanding the vulnerabilities associated with multi-turn attacks is crucial for deploying these models securely. Adding appropriate guardrails and security measures is essential to mitigate the risks posed by sustained adversarial pressure.
In conclusion, the research conducted by Cisco’s AI Threat Research and Security team sheds light on a critical security gap in open-weight AI models. By exposing the vulnerabilities of these models to multi-turn attacks, the study underscores the importance of implementing robust security measures to protect against real-world threats. Enterprises must be aware of the risks associated with deploying open-weight models and take proactive steps to safeguard their AI systems from malicious actors. Google’s Gemma is making strides in emphasizing “rigorous safety protocols” to target a “low risk level” for misuse. With the outcome showing the lowest gap at 10.53% and a more balanced performance across single- and multi-turn scenarios, it’s clear that security is a top priority for the tech giant.
When it comes to AI models optimized for capability and flexibility, there is often a trade-off with built-in safety measures. While this design choice may be suitable for many enterprise use cases, it’s essential for organizations to understand that prioritizing capability first can sometimes mean sacrificing security. Therefore, it’s crucial for enterprises to allocate budget resources accordingly.
A recent study conducted by Cisco tested 102 subthreat categories, revealing that the top 15 categories had high success rates across all models. This highlights the importance of implementing targeted defensive measures to enhance security and mitigate risks effectively.
Security is not just a barrier to AI adoption but a key enabler, according to Sampath from Google. By implementing the right security measures, enterprises can unlock productivity and unleash the full potential of AI tools without compromising data integrity or privacy.
The research suggests that enterprises should focus on six critical capabilities to enhance their security posture, including context-aware guardrails, model-agnostic runtime protections, continuous red-teaming, hardened system prompts, comprehensive logging, and threat-specific mitigations for top subthreat categories.
It’s crucial for organizations to act swiftly and not wait for the AI landscape to settle down. The rapid evolution of AI technologies and the increasing sophistication of cyber threats require proactive measures to safeguard data and systems effectively.
In conclusion, the report underscores the urgency for enterprises to prioritize multi-turn defenses over single-turn strategies and implement robust security measures to protect against evolving threats. By taking proactive steps to secure AI-powered systems and conversations, organizations can ensure a safer and more resilient digital ecosystem.



