28
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
this post was submitted on 16 Nov 2023
28 points (74.1% liked)
Privacy
32165 readers
216 users here now
A place to discuss privacy and freedom in the digital world.
Privacy has become a very important issue in modern society, with companies and governments constantly abusing their power, more and more people are waking up to the importance of digital privacy.
In this community everyone is welcome to post links and discuss topics related to privacy.
Some Rules
- Posting a link to a website containing tracking isn't great, if contents of the website are behind a paywall maybe copy them into the post
- Don't promote proprietary software
- Try to keep things on topic
- If you have a question, please try searching for previous discussions, maybe it has already been answered
- Reposts are fine, but should have at least a couple of weeks in between so that the post can reach a new audience
- Be nice :)
Related communities
much thanks to @gary_host_laptop for the logo design :)
founded 5 years ago
MODERATORS
It's illegal if you copy-paste someone's work verbatim. It's not illegal to, for example, summarize someone's work and write a short version of it.
As long as overfitting doesn't happen and the machine learning model actually learns general patterns, instead of memorizing training data, it should be perfectly capable of generating data that's not copied verbatim from humans. Whom, exactly, a model is plagiarizing if it generates a summarized version of some work you give it, particularly if that work is novel and was created or published after the model was trained?
All these AI do is algorithmically copy-paste. They don't have original thoughts and or original conclusions or original ideas, all if it is just copy-paste with extra steps.
Learning is, essentially, "algorithmically copy-paste". The vast majority of things you know, you've learned from other people or other people's works. What makes you more than a copy-pasting machine is the ability to extrapolate from that acquired knowledge to create new knowledge.
And currently existing models can often do the same! Sometimes they make pretty stupid mistakes, but they often do, in fact, manage to end up with brand new information derived from old stuff.
I've tortured various LLMs with short stories, questions and riddles, which I've written specifically for the task and which I've asked the models to explain or rewrite. Surprisingly, they often get things either mostly or absolutely right, despite the fact it's novel data they've never seen before. So, there's definitely some actual learning going on. Or, at least, something incredibly close to it, to the point it's nigh impossible to differentiate it from actual learning.
deleted
Not once did I claim that LLMs are sapient, sentient or even have any kind of personality. I didn't even use the overused term "AI".
LLMs, for example, are something like... a calculator. But for text.
A calculator for pure numbers is a pretty simple device all the logic of which can be designed by a human directly.
When we want to create a solver for systems that aren't as easily defined, we have to resort to other methods. E.g. "machine learning".
Basically, instead of designing all the logic entirely by hand, we create a system which can end up in a number of finite, yet still near infinite states, each of which defines behavior different from the other. By slowly tuning the model using existing data and checking its performance we (ideally) end up with a solver for something a human mind can't even break up into the building blocks, due to the shear complexity of the given system (such as a natural language).
And like a calculator that can derive that 2 + 3 is 5, despite the fact that number 5 is never mentioned in the input, or that particular formula was not a part of the suit of tests that were used to verify that the calculator works correctly, a machine learning model can figure out that "apple slices + batter = apple pie", assuming it has been tuned (aka trained) right.
Chat bots do not learn, stop anthropomorphizing them.
Not once did I claim that LLMs are sapient, sentient or even have any kind of personality. I didn't even use the overused term "AI".
LLMs, for example, are something like... a calculator. But for text.
A calculator for pure numbers is a pretty simple device all the logic of which can be designed by a human directly.
When we want to create a solver for systems that aren't as easily defined, we have to resort to other methods. E.g. "machine learning".
Basically, instead of designing all the logic entirely by hand, we create a system which can end up in a number of finite, yet still near infinite states, each of which defines behavior different from the other. By slowly tuning the model using existing data and checking its performance we (ideally) end up with a solver for something a human mind can't even break up into the building blocks, due to the shear complexity of the given system (such as a natural language).
And like a calculator that can derive that 2 + 3 is 5, despite the fact that number 5 is never mentioned in the input, or that particular formula was not a part of the suit of tests that were used to verify that the calculator works correctly, a machine learning model can figure out that "apple slices + batter = apple pie", assuming it has been tuned (aka trained) right.