473
Jailbroken AI Chatbots Can Jailbreak Other Chatbots
(www.scientificamerican.com)
This is a most excellent place for technology news and articles.
Info hazards are going to be more common place with this kind of technology. At the core of the problem is the ease of access of dangerous information. For example a lot of chat bots will confidently get things wrong. Combine that easy directions to make something like napalm or meth then we get dangerous things that could be incorrectly made. (Granted napalm or meth isn’t that hard to make)
As to what makes it dangerous information, it’s unearned. A chemistry student can make drugs, bombs, etc. but they learn/earn that information (and ideally the discipline) to use it. Kind of like in the US we are having more and more mass shootings due to ease of access of firearms. Restrictions on information or firearms aren’t going to solve the problems that cause them but it does make it (a little) harder.
At least that’s my understanding of it.
I don't exactly agree with the "earned" part but guess you have a point with the missing 'how to safely handle'.
By earned I mean it takes some efforts to gain that knowledge. For example some kind of training, studying, practice, etc. it’s typically during that process you learn how to safely and correctly do things
Anyone who wants to make even slightly complex organic compounds will also need to study five different types of isomerism and how they determine major / minor product. That should be enough of a deterrent.