AIBase
Home
AI NEWS
AI Tools
AI Models
MCP
AI Services
AI Compute
AI Tutorial
EN

AI News

View More

Embarrassing! Meta's AI Security System Easily Bypassed by 'Spaces' Attack

The Prompt-Guard-86M model released by Meta is designed to defend against prompt injection attacks by restricting large language models from processing inappropriate inputs, thereby protecting system security. However, the model itself also exposes risks of being attacked. Research conducted by Aman Priyanshu found that by adding simple character spacing such as spaces or removing punctuation in the input, the model disregards prior security instructions, achieving an almost 100% success rate for attacks. This finding highlights the importance of AI security, despite Prompt

12.8k 2 days ago
Embarrassing! Meta's AI Security System Easily Bypassed by 'Spaces' Attack

Models

View More

Prompt Guard 86M

meta-llama

P

PromptGuard is a text classification model designed to detect and protect against LLM prompt attacks, capable of identifying malicious prompt injections and jailbreak attempts.

Natural Language ProcessingTransformersTransformersEnglish
meta-llama
33.9k
250
AIBase
Empowering the future, your artificial intelligence solution think tank
English简体中文繁體中文にほんご
FirendLinks:
AI Newsletters AI ToolsMCP ServersAI NewsAIBaseLLM LeaderboardAI Ranking
© 2025AIBase
Business CooperationSite Map