282
Someone got Gab's AI chatbot to show its instructions
(mbin.grits.dev)
A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.
Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.
Subcommunities on Beehaw:
This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.
All of these AI prompts sound like begging. We're begging computers to do things for us now.
Basically all the instruction dumps I've seen
We always have been, it's just that the begging started out looking like math and has gradually gotten more abstract over time. We've just reached the point where we've explained to it in mathematical terms how to let us beg in natural language in certain narrow contexts.
If somebody told me five years ago about Adversarial Prompt Attacks I'd tell them they're horribly misled and don't understand how computers work, but yet here we are, and folks are using social engineering to get AI models to do things they aren't supposed to