Skip to content

It's rude to show AI output to people

Technology
30 28 0
  • This is a good post.

    Thinking about it some more, I don't necessarily mind if someone said "I googled it and..." then provides some self generated summary of what they found which is relevant to the discussion.

    I wouldn't mind if someone did the same with an LLM response. But just like I don't want to read a copy and paste of chatgpt results I don't want to read someone copy/pasting search results with no human analysis.

  • This is exactly something that has annoyed me in a sports community I follow back on Reddit. Posts with titles along the lines of “I asked ChatGPT what it thinks will happen in the game this weekend and here is what it said”.

    Why? What does ChatGPT add to the conversation here? Asking the question directly in the subreddit would have encouraged the same discussion.

    We’ve also learned nothing about the OPs opinion on the matter, other than maybe that they don’t have one. And even more to the point, it’s so intellectually lazy that it just feels like karma farming. “Ya I have nothing to add but I do love me them updoots”.

    I would rather someone posted saying they knew shit all about the sport but they were interested, than someone feigning knowledge by using ChatGPT as some sort of novel point of view, which it never is. It’s ways the most milquetoast response possible, ironically adding less to the conversation than the question it’s responding to.

    But that argument always just feels overly combative for what is otherwise a pretty relaxed sports community. It’s just not worth having that fight there.

    I would rather someone posted saying they knew shit all about the sport but they were interested, than someone feigning knowledge by using ChatGPT as some sort of novel point of view, which it never is. It’s ways the most milquetoast response possible, ironically adding less to the conversation than the question it’s responding to.

    That's literally the point of them. They're supposed to generate what the most likely result would be. They aren't supposed to be creative or anything like that. They're supposed to be generic.

  • Dude, the problem is you have no fucking idea if it’s wrong yourself, have nothing to back it up

    That's not true. For starters you can evaluate it on its own merits to see if it makes logical sense - the AI can help solve a maths equation for you and you can see that it checks out without needing something else to back it up.

    Second, agentic or multiple-step AI:s will dig out the sources for you so you can check them. It's just a smarter search engine with no ads and better focus on the question asked.

    If you have evaluated the statement for its correctness and relevance, then you can just own up to the statement yourself. There is no need to defer responsibility by prefacing it with “I asked [some AI service] and here’s what it said”. That is the point of the article that is being discussed, if you'd like to give it a read sometime.

  • I think sometimes when we ask people something we're not just seeking information. We're also engaging with other humans. We're connecting, signaling something, communicating something with the question, and so on. I use LLMs when I literally just want to know something, but I also try to remember the value of talking to other human beings as well.

  • Dude, the problem is you have no fucking idea if it’s wrong yourself, have nothing to back it up

    That's not true. For starters you can evaluate it on its own merits to see if it makes logical sense - the AI can help solve a maths equation for you and you can see that it checks out without needing something else to back it up.

    Second, agentic or multiple-step AI:s will dig out the sources for you so you can check them. It's just a smarter search engine with no ads and better focus on the question asked.

    Ok, I didn't need you to act as a middle man to tell me what the LLM just hallucinated, I can do this myself.

    The point is that raw AI output provides absolutely no value to a conversation, and is thus noisy and rude.

    When we ask questions on a public forum, we're looking to talk to people about their own experience and research through the lens of their own being and expertise. We're all capable of prompting an AI agent. If we wanted AI answers, we'd prompt an AI agent.

  • This is a good post.

    Thinking about it some more, I don't necessarily mind if someone said "I googled it and..." then provides some self generated summary of what they found which is relevant to the discussion.

    I wouldn't mind if someone did the same with an LLM response. But just like I don't want to read a copy and paste of chatgpt results I don't want to read someone copy/pasting search results with no human analysis.

    If you're going to use an LLM, at least follow the links it provides to the source of what they output. You really need to check their work.

  • Dude, the problem is you have no fucking idea if it’s wrong yourself, have nothing to back it up

    That's not true. For starters you can evaluate it on its own merits to see if it makes logical sense - the AI can help solve a maths equation for you and you can see that it checks out without needing something else to back it up.

    Second, agentic or multiple-step AI:s will dig out the sources for you so you can check them. It's just a smarter search engine with no ads and better focus on the question asked.

    I am speaking from experience.

    The latest example of that I encountered had a blatant logical inconsistency in its summary, a CVE that wasn't relevant to what was discussed, because it was corrected years before the technology existed. Someone pointed at it.

    The poster hadn't done the slightest to check what they posted, they just regurgitated it. It's not the reader's job to check the crap you've posted without the slightest effort.

  • This is a good post.

    Thinking about it some more, I don't necessarily mind if someone said "I googled it and..." then provides some self generated summary of what they found which is relevant to the discussion.

    I wouldn't mind if someone did the same with an LLM response. But just like I don't want to read a copy and paste of chatgpt results I don't want to read someone copy/pasting search results with no human analysis.

    I have a few colleagues that are very skilled and likeable people, but have horrible digital etiquette (40-50 year olds).

    Expecting people to read regurgitated gpt-summaries are the most obvious.

    But another one that bugs me just as much, are sharing links with no annotation. Could be a small article or a long ass report or white paper with 140 pages. Like, you expect me to bother read it, but you can't bother to say what's relevant about it?

    I genuinely think it's well intentioned for the most part. They're just clueless about what makes for good digital etiquette.

  • The worst is being in a technical role, and having project managers and marketing people telling me how it is based on some chathpt output

    Like shut the fuck up please, you literally don’t know what you are talking about

  • I would rather someone posted saying they knew shit all about the sport but they were interested, than someone feigning knowledge by using ChatGPT as some sort of novel point of view, which it never is. It’s ways the most milquetoast response possible, ironically adding less to the conversation than the question it’s responding to.

    That's literally the point of them. They're supposed to generate what the most likely result would be. They aren't supposed to be creative or anything like that. They're supposed to be generic.

    Hey! ChatGPT can be creative if you ask it to roast fictional characters .. somewhat!

  • The worst is being in a technical role, and having project managers and marketing people telling me how it is based on some chathpt output

    Like shut the fuck up please, you literally don’t know what you are talking about

    Sadly we had that problem before AI too... "Some dude I know told me this is super easy to do"

  • You're damn right, if somebody puts slop in my face I get visibly aggressive.

  • I think sometimes when we ask people something we're not just seeking information. We're also engaging with other humans. We're connecting, signaling something, communicating something with the question, and so on. I use LLMs when I literally just want to know something, but I also try to remember the value of talking to other human beings as well.

    You should pretty much assume everything that a chatbot says could be false to a much higher degree than human written content, making it effectively useless for your stated purpose.

  • Every now and then I see a guy barging in a topic bringing nothing else than "I asked [some AI service] and here's what it said", followed by 3 paragraphs of AI-gened gibberish. And then when it's not well received they just don't seem to understand.

    It's baffling to me. Anyone can ask an AI. A lot of people specifically don't, because they don't want to battle with its output for an hour trying to sort out from where it got its information, whether it represented it well, or even whether it just hallucinated half of it.

    And those guys come posting a wall of text they may or may not have read themselves, and then they have the gall to go "What's the problem, is any of that wrong?"... Dude, the problem is you have no fucking idea if it's wrong yourself, have nothing to back it up, and have only brought automated noise to the conversation.

    I was trying to help onboard a new lead engineer and I was working through debugging his caddy config on Slack. I'm clearly putting in effort to help him diagnose his issue and he posts "I asked chatgpt and it said these two lines need to be reversed", which was completely false (caddy has a system for reordering directives) and honestly just straight up insulting. Fucking pissed me off. People need to stop brining AI slop into conversations. It isn't welcome and can fuck right off.

    The actual issue? He forgot to restart his development server. 😡

  • 314 Stimmen
    141 Beiträge
    930 Aufrufe
    zacryon@feddit.orgZ
    I see. If moving to another country, where you don't have to suffer such conditions, is also not an option then I hope you're looking for something else while you're at your current job. These are no conditions anyone should suffer.
  • 300 Stimmen
    47 Beiträge
    307 Aufrufe
    T
    I worked in a bank for a bit. Literally any transaction that's large and unusual for the account will be flagged. Also people do bonkers things with their money for the stupidest reasons all the time so all that one has to do if they're making large transactions is be prepared to talk to the bank and explain what's going on. Unless of course you are handling money in relation to organized crime, in which case you were fucked the moment the money touched the banking system
  • How Social Media Brings Out the Worst in Us

    Technology technology
    14
    1
    120 Stimmen
    14 Beiträge
    74 Aufrufe
    sturgist@lemmy.caS
    Suffering from asthma? 9/10 Doctors recommend menthol cigarettes! Peppermint fresh puts the pep in your step!
  • My AI Skeptic Friends Are All Nuts

    Technology technology
    31
    1
    13 Stimmen
    31 Beiträge
    162 Aufrufe
    J
    I did read it, and my comment is exactly referencing the attitude of the author which is "It's good enough, so you should use it". I disagree, and say it's another dumbass shortcut to cash grab on a less than stellar ecosystem and product. It's training wheels for failure.
  • Hands-On: EufyMake E1 UV Printer

    Technology technology
    18
    1
    38 Stimmen
    18 Beiträge
    91 Aufrufe
    S
    I watched a bit of Michael Alm's video on this, but noped out when I saw all of the little boxes of consumables appearing. If regular printer ink is already exorbitant, I can only imagine what these proprietary cartridges will cost.
  • 0 Stimmen
    6 Beiträge
    39 Aufrufe
    L
    Divide and conquer. Non state-actors and special interest have a far easier time attacking a hundred small entities than one big one. Because people have much less bandwidth to track all this shit than it is to spread it around. See ALEC and the strategy behind state rights. In the end this is about economic power. The only way to curb it is through a democratic government. Lemmy servers too can be bought and sold and the communities captured that grew on them.
  • 531 Stimmen
    31 Beiträge
    148 Aufrufe
    ulrich@feddit.orgU
    If you want a narrative, look at all the full-price $250k Roadster pre-orders they've been holding onto for like 8 years now with zero signs of production and complete silence for the last...5 years?
  • MDM Thoughts?

    Technology technology
    2
    0 Stimmen
    2 Beiträge
    26 Aufrufe
    R
    Hello folks! Interested in learning new skills? Check out the best courses in graphic design- https://www.admecindia.co.in/courses/graphic-design-courses/ https://www.admecindia.co.in/course/advanced-graphic-design-master-course/ https://www.admecindia.co.in/course/most-advanced-graphic-design-course-master-plus/