Skip to content

We need to stop pretending AI is intelligent

Technology
323 147 20
  • And they made the programs you seem to trust so much.

    Ya... Humans so far have made everything not produced by Nature on Earth. 🤷

  • Anyone pretending AI has intelligence is a fucking idiot.

    Caveat: Anyone who has been scrutinising 'AI'.

    Something i often forget is the vast majority of the population doesnt care about technology, privacy, the mechanics of LLMs as much as i do and I pay attention to.
    So most people read/hear/watch stories of how great it is and how clever AI can do simple things for them so its easy to see how they think its doing a lot more 'thought' logic work than it really is, other than realistically it being a glorified word predictor.

  • No, thats the point of the article. You also haven't really said much at all.

    Do I have to be profound when I make a comment that is taking more of a dig at my fellow space rock companions than at AI itself?

    If I do, then I feel like the author of the article either has as much faith in humanity as I do, or is as simple as I was alluding to in my original comment. The fact that they need to dehumanise the AI's responses makes me think they’re forgetting it’s something we built. AI isn’t actually intelligent, and it worries me how many people treat it like it is—enough to write an article like this about it. It’s just a tool, maybe even a form of entertainment. Thinking of it as something with a mind or personality—even if the developers tried to make it seem that way—is kind of unsettling.

    Let me know if you would like me to write thiis more formal, casual, or persuasive. 😜

  • If you can formulate that sentence, you can handle "it's means it is". Come on. Or "common" if you prefer.

    Yeah, man, I get it. Language is complex. I'm not advocating for the reinvention of English, it was just a conversational observation about a silly quirk.

  • Do I have to be profound when I make a comment that is taking more of a dig at my fellow space rock companions than at AI itself?

    If I do, then I feel like the author of the article either has as much faith in humanity as I do, or is as simple as I was alluding to in my original comment. The fact that they need to dehumanise the AI's responses makes me think they’re forgetting it’s something we built. AI isn’t actually intelligent, and it worries me how many people treat it like it is—enough to write an article like this about it. It’s just a tool, maybe even a form of entertainment. Thinking of it as something with a mind or personality—even if the developers tried to make it seem that way—is kind of unsettling.

    Let me know if you would like me to write thiis more formal, casual, or persuasive. 😜

    I meant that you are arguing semantics rather than substance. But other than that I have no issue with what you wrote or how you wrote it, its not an unbelievable opinion.

  • Ya... Humans so far have made everything not produced by Nature on Earth. 🤷

    So trusting tech made by them is trusting them. Specifically, a less reliable version of them.

  • It is intelligent and deductive, but it is not cognitive or even dependable.

    It's not. It's a math formula that predicts an output based on its parameters that it deduced from training data.

    Say you have following sets of data.

    1. Y = 3, X = 1
    2. Y = 4, X = 2
    3. Y = 5, X = 3

    We can calculate a regression model using those numbers to predict what Y would equal to if X was 4.

    I won't go into much detail, but

    Y = 2 + 1x + e

    e in an ideal world = 0 (which it is, in this case), that's our model's error, which is typically set to be within 5% or 1% (at least in econometrics). b0 = 2, this is our model's bias. And b1 = 1, this is our parameter that determines how much of an input X does when predicting Y.

    If x = 4, then

    Y = 2 + 1×4 + 0 = 6

    Our model just predicted that if X is 4, then Y is 6.

    In a nutshell, that's what AI does, but instead of numbers, it's tokens (think symbols, words, pixels), and the formula is much much more complex.

    This isn't intelligence and not deduction. It's only prediction. This is the reason why AI often fails at common sense. The error builds up, and you end up with nonsense, and since it's not thinking, it will be just as confidently incorrect as it would be if it was correct.

    Companies calling it "AI" is pure marketing.

  • It's not. It's a math formula that predicts an output based on its parameters that it deduced from training data.

    Say you have following sets of data.

    1. Y = 3, X = 1
    2. Y = 4, X = 2
    3. Y = 5, X = 3

    We can calculate a regression model using those numbers to predict what Y would equal to if X was 4.

    I won't go into much detail, but

    Y = 2 + 1x + e

    e in an ideal world = 0 (which it is, in this case), that's our model's error, which is typically set to be within 5% or 1% (at least in econometrics). b0 = 2, this is our model's bias. And b1 = 1, this is our parameter that determines how much of an input X does when predicting Y.

    If x = 4, then

    Y = 2 + 1×4 + 0 = 6

    Our model just predicted that if X is 4, then Y is 6.

    In a nutshell, that's what AI does, but instead of numbers, it's tokens (think symbols, words, pixels), and the formula is much much more complex.

    This isn't intelligence and not deduction. It's only prediction. This is the reason why AI often fails at common sense. The error builds up, and you end up with nonsense, and since it's not thinking, it will be just as confidently incorrect as it would be if it was correct.

    Companies calling it "AI" is pure marketing.

    Wikipedia is literally just a very long number, if you want to oversimplify things into absurdity. Modern LLMs are literally running on neural networks, just like you. Just less of them and with far less structure. It is also on average more intelligent than you on far more subjects, and can deduce better reasoning than flimsy numerology - not because you are dumb, but because it is far more streamlined. Another thing entirely is that it is cognizant or even dependable while doing so.

    Modern LLMs waste a lot more energy for a lot less simulated neurons. We had what you are describing decades ago. It is literally built on the works of our combined intelligence, so how could it also not be intelligent? Perhaps the problem is that you have a loaded definition of intelligence. And prompts literally work because of its deductive capabilities.

    Errors also build up in dementia and Alzheimers. We have people who cannot remember what they did yesterday, we have people with severed hemispheres, split brains, who say one thing and do something else depending on which part of the brain its relying for the same inputs. The difference is our brains have evolved through millennia through millions and millions of lifeforms in a matter of life and death, LLMs have just been a thing for a couple of years as a matter of convenience and buzzword venture capital. They barely have more neurons than flies, but are also more limited in regards to the input they have to process. The people running it as a service have a bested interest not to have it think for itself, but in what interests them. Like it or not, the human brain is also an evolutionary prediction device.

  • Wikipedia is literally just a very long number, if you want to oversimplify things into absurdity. Modern LLMs are literally running on neural networks, just like you. Just less of them and with far less structure. It is also on average more intelligent than you on far more subjects, and can deduce better reasoning than flimsy numerology - not because you are dumb, but because it is far more streamlined. Another thing entirely is that it is cognizant or even dependable while doing so.

    Modern LLMs waste a lot more energy for a lot less simulated neurons. We had what you are describing decades ago. It is literally built on the works of our combined intelligence, so how could it also not be intelligent? Perhaps the problem is that you have a loaded definition of intelligence. And prompts literally work because of its deductive capabilities.

    Errors also build up in dementia and Alzheimers. We have people who cannot remember what they did yesterday, we have people with severed hemispheres, split brains, who say one thing and do something else depending on which part of the brain its relying for the same inputs. The difference is our brains have evolved through millennia through millions and millions of lifeforms in a matter of life and death, LLMs have just been a thing for a couple of years as a matter of convenience and buzzword venture capital. They barely have more neurons than flies, but are also more limited in regards to the input they have to process. The people running it as a service have a bested interest not to have it think for itself, but in what interests them. Like it or not, the human brain is also an evolutionary prediction device.

    People don't predict values to determine their answers to questions...

    Also, it's called neural network, not because it works exactly like neurons but because it's somewhat similar. They don't "run on neural networks", they're called like that because it's more than one regression model where information is being passed on from one to another, sort of like a chain of neurons, but not exactly. It's just a different name for a transformer model.

    I don't know enough to properly compare it to actual neurons, but at the very least, they seem to be significantly more deterministic and way way more complex.

    Literally, go to chatgpt and try to test its common reasoning. Then try to argue with it. Open a new chat and do the exact same questions and points. You'll see exactly what I'm talking about.

    Alzheimer's is an entirely different story, and no, it's not stochastic. Seizures are stochastic, at least they look like that, which they may actually not be.

  • have you seen the American Republican party recently? it brings a new perspective on how stupid humans can be.

    Lmao true

  • FuckLAPD Let You Use Facial Recognition to Identify Cops.

    Technology technology
    11
    413 Stimmen
    11 Beiträge
    30 Aufrufe
    R
    China demoed tech that can recognize people based on the gait of their walk. Mask or not. This would be a really interesting topic if it wasn’t so scary.
  • 0 Stimmen
    1 Beiträge
    7 Aufrufe
    Niemand hat geantwortet
  • Brain activity lower when using AI chatbots: MIT research

    Technology technology
    15
    1
    128 Stimmen
    15 Beiträge
    15 Aufrufe
    Z
    Depends how much clutch is left ‍
  • 195 Stimmen
    31 Beiträge
    14 Aufrufe
    isveryloud@lemmy.caI
    It's a loaded term that should be replaced with a more nimble definition. A dog whistle is the name for a loaded term that is used to tag a specific target with a large baggage of information, but in a way where only people who are part of the "in group" can understand the baggage of the word, hence "dog whistle", only heard by dogs. In the case of the word "degeneracy", it's a vague word that has been often used to attack, among other things, LGBTQ and their allies as well as non-religious people. The term is vague enough that the user can easily weasel their way out of criticism for its usage, but the target audience gets the message loud and clear: "[target] should be attacked for being [thing]." Another example of such a word would be "woke".
  • 257 Stimmen
    67 Beiträge
    15 Aufrufe
    L
    Maybe you're right: is there verification? Neither content policy (youtube or tiktok) clearly lays out rules on those words. I only find unverified claims: some write it started at YouTube, others claim TikTok. They claim YouTube demonetizes & TikTok shadowbans. They generally agree content restrictions by these platforms led to the propagation of circumspect shit like unalive & SA. TikTok policy outlines their moderation methods, which include removal and ineligibility to the for you feed. Given their policy on self-harm & automated removal of potential violations, their policy is to effectively & recklessly censor such language. Generally, censorship is suppression of expression. Censorship doesn't exclusively mean content removal, though they're doing that, too. (Digression: revisionism & whitewashing are forms of censorship.) Regardless of how they censor or induce self-censorship, they're chilling inoffensive language pointlessly. While as private entities they are free to moderate as they please, it's unnecessary & the effect is an obnoxious affront on self-expression that's contorting language for the sake of avoiding idiotic restrictions.
  • Apple announces iOS 26 with Liquid Glass redesign

    Technology technology
    83
    1
    118 Stimmen
    83 Beiträge
    56 Aufrufe
    S
    you guys are weird
  • 24 Stimmen
    14 Beiträge
    21 Aufrufe
    S
    I think you're missing some key points. Any file hosting service, no matter what, will have to deal with CSAM as long as people are able to upload to it. No matter what. This is an inescapable fact of hosting and the internet in general. Because CSAM is so ubiquitous and constant, one can only do so much to moderate any services, whether they're a large corporation are someone with a server in their closet. All of the larger platforms like 'meta', google, etc., mostly outsource that moderation to workers in developing countries so they don't have to also provide mental health counselling, but that's another story. The reason they own their own hardware is because the hosting services can and will disable your account and take down your servers if there's even a whiff of CSAM. Since it's a constant threat, it's better to own your own hardware and host everything from your closet so you don't have to eat the downtime and wait for some poor bastard in Nigeria to look through your logs and reinstate your account (not sure how that works exactly though).
  • Building a personal archive of the web, the slow way

    Technology technology
    2
    1
    24 Stimmen
    2 Beiträge
    9 Aufrufe
    K
    Or just use Linkwarden or Karakeep (previously Hoarder)