AI safety guardrails easily thwarted, security study finds
The Insider News
2
Posts
2
Posters
0
Views
1
Watching
-
The "guardrails" created to prevent large language models (LLMs) such as OpenAI's GPT-3.5 Turbo from spewing toxic content have been shown to be very fragile.
Unsafe at any speed
-
The "guardrails" created to prevent large language models (LLMs) such as OpenAI's GPT-3.5 Turbo from spewing toxic content have been shown to be very fragile.
Unsafe at any speed
If even the 3 laws got thwarted... do you really think this is going to go anywhere?
M.D.V. ;) If something has a solution... Why do we have to worry about?. If it has no solution... For what reason do we have to worry about? Help me to understand what I'm saying, and I'll explain it better to you Rating helpful answers is nice, but saying thanks can be even nicer.