I asked the AI if its been affecting me and it told me that was a really observant question that shows my great emotional intelligence, so I think I'm smart enough to notice if it ever becomes a sycophant, don't worry I got this 😎
Technology
Which posts fit here?
Any news that are at least tangentially connected to the technology, social media platforms, informational technologies or tech policy.
Post guidelines
[Opinion] prefix
Opinion (op-ed) articles must use [Opinion] prefix before the title.
Rules
1. English only
Title and associated content has to be in English.
2. Use original link
Post URL should be the original link to the article (even if paywalled) and archived copies left in the body. It allows avoiding duplicate posts when cross-posting.
3. Respectful communication
All communication has to be respectful of differing opinions, viewpoints, and experiences.
4. Inclusivity
Everyone is welcome here regardless of age, body size, visible or invisible disability, ethnicity, sex characteristics, gender identity and expression, education, socio-economic status, nationality, personal appearance, race, caste, color, religion, or sexual identity and orientation.
5. Ad hominem attacks
Any kind of personal attacks are expressly forbidden. If you can't argue your position without attacking a person's character, you already lost the argument.
6. Off-topic tangents
Stay on topic. Keep it relevant.
7. Instance rules may apply
If something is not covered by community rules, but are against lemmy.zip instance rules, they will be enforced.
Companion communities
!globalnews@lemmy.zip
!interestingshare@lemmy.zip
Icon attribution | Banner attribution
If someone is interested in moderating this community, message @brikox@lemmy.zip.
You've got to remember that these are just simple farmers. These are people of the land. The common clay of the new West. You know... morons.
This is what it must feel like to be a billionaire, surrounded by yes-men. That's why they love AI, not because they understand it, but because they don't see how its not normal.
Damn, we're so easy to manipulate.
Do you and yours a big favor and stay away from that shit like it's heroin.
I use it, but have established a realistic mindset that it's alwqys confidentially incorrect and in many cases I'm better off walking away and just doing the thing myself.
In saying that, I've also established a mindset that people who actively rely on genAI must be low on intelligence. Not only lacking in knowledge or pursuing knowledge of whatever they're using it for, but genuinely of a mental calibre that is unable to discern or realise its low performance.
Someone here pointed out the error of the old "even a broken clock is right twice a day" cliche. If you have to independently check if it's correct, then it's not giving you any useful information.
Yes, but only 22 times out of 24 🤣
I gave mine rules to always question me and provide critical feedback. It's quite annoying sometimes but much better than when I was a genius for just about anything.
I watched an interview with Hannah Fry a few weeks ago and she said that is how she prompts the LLMs she uses.
heroin
Not harmful and psychosis inducing enough.
They're more like PCP.
Why not a mix of both?
Flattery gets you everywhere… handsome ;)
Sycophantic or highly unreasonable up-talking instantly makes me think you are a sleazeball.
I would like AIs a whole lot more if they would: 1) respond in as few words as possible, and 2) be right way more often then they currently are. As it is, I only use them if all other research methods have failed (very rarely). And even then, I don't actually read their output, I skim for keywords to do research on.
A completely made up example on a topic I already know things about: If I'm looking for a stronger drill but I'm just finding more drills, maybe it will say something about an impact driver and I can go research what that is and figure out if it is what I need.
Yeah their excessive use of lists and tables is also something common to LLMs. Sometimes you ask an LLM a basic question and then it responds with all these unnecessary tables and lists, and then clarifications of the previous tables and lists with more tables and lists, then a summary of all these tables and lists with another list… It’s a lot. If a person were using that many tables and lists in their day to day texting then I’d assume that they were suffering from a psychotic episode
The first you can control to some extent. Both local and public llms have ways to edit or add to the system prompt, which is what guides the overall behavior. I actually had a local llm do the opposite of what you are looking for - somehow the prompt had been changed to a very simple "You will answer short and concise" without me realizing it, and I couldn't figure out why it had changed from a flowing, dynamic output to a few sentences.
But it's not perfect either. Sometimes you want a bit more than a simple sentence, or it might need more information and a short reply will cut off the important things.
As for fixing the second one - to be right more often would mean they understand what they're outputting, which is what we don't have yet. I'd just rather have it admit when it doesn't have enough to satisfactorily be sure on the answer. Which doesn't happen because they are trained first and foremost to always have an answer, because that's more marketable than a model that says it doesn't know.
This is 100% my experience. Ai simply can not solve problems. It isnt capable of thinking objectively at all, no sense of any kind of permanence beyond the immediate task. I have found it educational in the sense that un-fucking something that ai has put together, can teach me a lot about a system I was previously unfamiliar with.
It is a machine that outputs huge amounts of useless garbage with little practical value.
Interested to know if this is affecting certain cultures more than others. Here (UK) we seem to find a lot of Americans "false" in the way they communicate because it's too big, too obvious. "You're trying too hard to be nice". We'll understate both positive and negative comments.
It would suggest Brits wouldn't trust a sycophantic LLM as much, but I wonder if that's true.
Even that’s cultural. We aren’t “nice” in New England and that really bothers the southerners.
~~Folk~~ Idiots
Idiots
Do you have the slightest idea how little that narrows it down?
Yeah, it has always seemed creepy to me how positive it is about anything you ask it.
I hardly ever use it, and when I do I imagine I am talking to a beautiful saleswoman with a large name tag with the logo of the company.
The AI may be pretty, but it always represents someone else's interests
Use it enough and you’ll see it’s not like that. There’s plenty it will push back on. Depending on the ai… you can see the narrative they are pushing through what it pushes back on.
There are some topics it absolutely denies.
*absolutely right
Last time it said I had a realy galaxy brain idea. I wish we could tone down the sycophant mode.
In the future you will only be able to prove you are a human by simply being a contrarian.
pfft. that will never happen.
Verfied Human. Comment approved.