AI models could be hacked by a whole new type of Skeleton Key attacks, Microsoft warns

Estimated read time 2 min read



Microsoft has shared details on a new hacking method which bypasses the security systems baked into AI models and makes them return malicious, dangerous, and harmful content.

The researchers call the technique Skeleton Key, and it applies across well-known models including Meta Llama3-70b-instruct (base), Google Gemini Pro (base), OpenAI GPT 3.5 Turbo (hosted), OpenAI GPT 4o (hosted), Mistral Large (hosted), Anthropic Claude 3 Opus (hosted), and Cohere Commander R Plus (hosted).



Source link

You May Also Like

More From Author

+ There are no comments

Add yours