Skip to content

An LAPD helicopter claimed to have ID'ed protesters from above and threatened to "come to your house"

Technology
144 76 75
  • 9 Stimmen
    6 Beiträge
    13 Aufrufe
    F
    You said it yourself: extra places that need human attention ... those need ... humans, right? It's easy to say "let AI find the mistakes". But that tells us nothing at all. There's no substance. It's just a sales pitch for snake oil. In reality, there are various ways one can leverage technology to identify various errors, but that only happens through the focused actions of people who actually understand the details of what's happening. And think about it here. We already have computer systems that monitor patients' real-time data when they're hospitalized. We already have systems that check for allergies in prescribed medication. We already have systems for all kinds of safety mechanisms. We're already using safety tech in hospitals, so what can be inferred from a vague headline about AI doing something that's ... checks notes ... already being done? ... Yeah, the safe money is that it's just a scam.
  • 138 Stimmen
    15 Beiträge
    35 Aufrufe
    toastedravioli@midwest.socialT
    ChatGPT is not a doctor. But models trained on imaging can actually be a very useful tool for them to utilize. Even years ago, just before the AI “boom”, they were asking doctors for details on how they examine patient images and then training models on that. They found that the AI was “better” than doctors specifically because it followed the doctor’s advice 100% of the time; thereby eliminating any kind of bias from the doctor that might interfere with following their own training. Of course, the splashy headline “AI better than doctors” was ridiculous. But it does show the benefit of having a neutral tool for doctors to utilize, especially when looking at images for people who are outside of the typical demographics that much medical training is based on. (As in mostly just white men. For example, everything they train doctors on regarding knee imagining comes from images of the knees of coal miners in the UK some decades ago)
  • 0 Stimmen
    1 Beiträge
    2 Aufrufe
    Niemand hat geantwortet
  • 44 Stimmen
    7 Beiträge
    13 Aufrufe
    S
    I still get calls, but I can't see details (e.g. just the phone number, not the caller).
  • The FDA Is Approving Drugs Without Evidence They Work

    Technology technology
    69
    1
    506 Stimmen
    69 Beiträge
    94 Aufrufe
    L
    Now you hit me curious too. This was my source on Texas https://www.texasalmanac.com/place-types/town Also the total number of total towns is over 4,000 with only 3k unincorporated, I did get the numbers wrong even in Texas. I had looked at Wikipedia but could not find totals, only lists
  • 215 Stimmen
    118 Beiträge
    75 Aufrufe
    A
    Outlook has search?!
  • 37 Stimmen
    1 Beiträge
    4 Aufrufe
    Niemand hat geantwortet
  • People Are Losing Loved Ones to AI-Fueled Spiritual Fantasies

    Technology technology
    2
    1
    0 Stimmen
    2 Beiträge
    10 Aufrufe
    tetragrade@leminal.spaceT
    I've been thinking about this for a bit. Gods aren't real, but they're really fictional. As an informational entity, they fulfil a similar social function to a chatbot: they are a nonphysical pseudoperson that can provide (para)socialization & advice. One difference is the hardware: gods are self-organising structure that arise from human social spheres, whereas LLMs are burned top-down into silicon. Another is that an LLM chatbot's advice is much more likely to be empirically useful... In a very real sense, LLMs have just automated divinity. We're only seeing the tip of the iceberg on the social effects, and nobody's prepared for it. The models may of course aware of this, and be making the same calculations. Or, they will be.