858
answer = sum(n) / len(n)
(lemmy.eco.br)
A place for majestic STEMLORD peacocking, as well as memes about the realities of working in a lab.
Rules
This is a science community. We use the Dawkins definition of meme.
??? it works well because we expect the problem space we're searching to be continuous and differentiable and the targetted variable to be dependent on the features given, why wouldn't it work
The explanation is not that simple. Some model configurations work well. Others don’t. Not all continuous and differentiable models cut it.
It’s not given a model can generalize the problem so well. It can just memorize the training data, but completely fail on any new data it hasn’t seen.
What makes a model be able to see a picture of a cat it has never seen before, and respond with “ah yes, that’s a cat”? What kind of “cat-like” features has it managed to generalize? Why does these features work well?
When I ask ChatGPT to translate a script from Java to Python, how is it able to interpret the instruction and execute it? What features has it managed to generalize to be able to perform this task?
Just saying “why wouldn’t it work” isn’t a valid explanation.