Skip to content

Microsoft Says Its New AI Diagnosed Patients 4 Times More Accurately Than Human Doctors

Technology
19 13 0
  • And the risk is that if we rely on AI in any meaningful capacity, it will eventually erode away the expertise who would be knowledgeable enough to detect the problems that the future AI may create/ignore. This assumes even best case where AI isn't being specifically tampered with.

    I agree with you. I think this will likely happen to some degree. At the same time, that kind of argument could be used against many new technologies and is not a valid one to not utilize new tech.

  • I agree with you. I think this will likely happen to some degree. At the same time, that kind of argument could be used against many new technologies and is not a valid one to not utilize new tech.

    Simply using AI isn't an issue... Allowing it to take over in a way that accelerates the removal of the knowledge from our pools of knowledge is a problem. Allowing companies to use AI as a direct replacement of actual medical professionals will remove knowledge from society. We already know that we can't use AI to fuel more AI learning... the models implode. In order to continue learning more from medicine, we need to keep pushing for human learning and understanding.

    Funny that you agree with me and apparently see useful discussion to have here... but downvote me even though the comment certainly added to the discussion.

    Oh, and next time don't put words into someone's mouth, very much a bad faith action that harms meaningful discussion. I never said we should ban it or never use it. A better answer would be to legislate that doctors must still oversee, or must be the approving authority. That AI can never have a final say in someone's care and that research must never be sourced from AI sources. All I said, is that if we continue what we're doing and rely on AI in any meaningful capacity, we will run into problems. Especially in the context of the comment I responded to which opined upon corporation controlled AI.

    FFS... they can't even run a vending machine. https://www.anthropic.com/research/project-vend-1

    Oh.. and actually I would consider the 85% that it gets to be pretty poor considering that the AI was likely trained on the full breadth of NEJM information. Doctors don't have that ability to retain and train on 100% of all knowledge of the NEJM, so mistaking things makes sense for them. It doesn't make sense for something that was trained on NEJM data to screw up on an NEJM case.

    My stance is the same for all AI. I'll use it to generate basic code for me. I'll never run that without review. Or to jumpstart research into a topic... and validate the information presented with outside direct sources.

    TL;DR: Tool is good... Source is bad.

  • The Microsoft AI team shares research that demonstrates how AI can sequentially investigate and solve medicine’s most complex diagnostic challenges—cases that expert physicians struggle to answer.

    Benchmarked against real-world case records published each week in the New England Journal of Medicine, we show that the Microsoft AI Diagnostic Orchestrator (MAI-DxO) correctly diagnoses up to 85% of NEJM case proceedings, a rate more than four times higher than a group of experienced physicians. MAI-DxO also gets to the correct diagnosis more cost-effectively than physicians.

    It seems that Microsoft can create AI products without relying on OpenAI. Although he speculated that the AI was trained on clinical information from hospitals that use Nuance Communications. Also that he received medical information.

    In any case, it is a positive development.

  • more accurate.

    Until it's not...then what. Who's liable? Google...Amazon ..Microsoft ..chatgpt.... Look, I like ai because it's fun to make stupid memes and pictures without any effort but I do not trust this nonsense to do ANYTHING with accuracy especially my medical.

    This thing will 100% be designed to diagnose people to sell you drugs and Not fix your health. Corporations control this. Currently they need to bribe Doctors to push their drugs..this will circumvent that entirely. You'll end up paying drastically more, for less.

    The sheer fact that's it's telling people to kill themselves to end suffering should be proof enough that it's dogshit

    more accurate.

    Until it’s not…then what. Who’s liable? Google…Amazon …Microsoft …chatgpt… Look, I like ai because it’s fun to make stupid memes and pictures without any effort but I do not trust this nonsense to do ANYTHING with accuracy especially my medical.

    The doctor who review the case, maybe ?
    In some cases the AI can effectively "see" things a doctor can miss and direct him to check for a particular disease. Even if the AI is only able to rule out some cases it would be usefull.

  • AI for pattern recognition (statistical stuff) IMHO is fine, it's different than expecting original thought, reasoning or understanding, which the new 'AI' does not do, despite the constant hype.

    This. Honestly things like image detection, anomaly detection over big data sets, and semantic searching, all seem very useful in professional contexts.

    Generative AI not heavily grounded in real data is just better for no-risks tasks.

  • more accurate.

    Until it’s not…then what. Who’s liable? Google…Amazon …Microsoft …chatgpt… Look, I like ai because it’s fun to make stupid memes and pictures without any effort but I do not trust this nonsense to do ANYTHING with accuracy especially my medical.

    The doctor who review the case, maybe ?
    In some cases the AI can effectively "see" things a doctor can miss and direct him to check for a particular disease. Even if the AI is only able to rule out some cases it would be usefull.

    The doctor who review the case, maybe ?

    Yeah that's why these gains in "efficiency" are completely imaginary.

  • The Microsoft AI team shares research that demonstrates how AI can sequentially investigate and solve medicine’s most complex diagnostic challenges—cases that expert physicians struggle to answer.

    Benchmarked against real-world case records published each week in the New England Journal of Medicine, we show that the Microsoft AI Diagnostic Orchestrator (MAI-DxO) correctly diagnoses up to 85% of NEJM case proceedings, a rate more than four times higher than a group of experienced physicians. MAI-DxO also gets to the correct diagnosis more cost-effectively than physicians.

    Somehow I doubt these corporate press releases.

    Microsoft

    Somehow I really doubt these corporate press releases.

    The Path to Medical Superintelligence

    Somehow I really really doubt these corporate press releases.

  • I know that I might be the only Lemmy user happy with this, but AI applications in the medical field seems very promising for lowering costs and being more accurate.

    AI applications

    There's no "AI" involved here.

  • AI for pattern recognition (statistical stuff) IMHO is fine, it's different than expecting original thought, reasoning or understanding, which the new 'AI' does not do, despite the constant hype.

    True.

    But a problem is that (as usual) it's not actually "AI" to find patterns using statistics.

    These corporations are literally willing to murder people in order to make a buck off some phony "medical superintelligence".

    Why would I trust these liars with my life? They're completely anti-science.

  • The doctor who review the case, maybe ?

    Yeah that's why these gains in "efficiency" are completely imaginary.

    Only if you don't have the critical thinking to understand how information management is a significant problem and barrier to medical care.

    Being able to research and find material relevant to a patient's problem is an arduous task that often is too high a barrier for doctors to invest in given their regular workloads.

    Which leads to a reduction in effective care.

    By providing a more efficient and effective way to dig up information that saves a ton of time and improves care.

    It's still up to the doctor to evaluate that information, but now they're not slogging away trying to find it.

  • The doctor who review the case, maybe ?

    Yeah that's why these gains in "efficiency" are completely imaginary.

    The AI only needs to alert the doctor that something is off and should be tested for. It does not replace doctors, but augments them. It's actually a great use for AI, it's just not what we think of as AI in a post-LLM world. The medically useful AI is pattern recognition. LLMs may also help doctors if they need a starting point into researching something weird and obscure, but ChatGPT isn't being used for diagnosing patients, nor is anything any AI says the "final verdict". It's just a tool to improve early detection of disorders, or it might point someone towards an useful article or book.

  • Telegram, the FSB, and the Man in the Middle

    Technology technology
    8
    1
    51 Stimmen
    8 Beiträge
    3 Aufrufe
    R
    You can be seen from a kilometer away, pots ))
  • 111 Stimmen
    24 Beiträge
    10 Aufrufe
    O
    Ingesting all the artwork you ever created by obtaining it illegally and feeding it into my plagarism remix machine is theft of your work, because I did not pay for it. Separately, keeping a copy of this work so I can do this repeatedly is also stealing your work. The judge ruled the first was okay but the second was not because the first is "transformative", which sadly means to me that the judge despite best efforts does not understand how a weighted matrix of tokens works and that while they may have some prevention steps in place now, early models showed the tech for what it was as it regurgitated text with only minor differences in word choice here and there. Current models have layers on top to try and prevent this user input, but escaping those safeguards is common, and it's also only masking the fact that the entire model is built off of the theft of other's work.
  • I Counted All of the Yurts in Mongolia Using Machine Learning

    Technology technology
    9
    17 Stimmen
    9 Beiträge
    14 Aufrufe
    G
    I'd say, when there's a policy and its goals aren't reached, that's a policy failure. If people don't like the policy, that's an issue but it's a separate issue. It doesn't seem likely that people prefer living in tents, though. But to be fair, the government may be doing the best it can. It's ranked "Flawed Democracy" by The Economist Democracy Index. That's really good, I'd say, considering the circumstances. They are placed slightly ahead of Argentina and Hungary. OP has this to say: Due to the large number of people moving to urban locations, it has been difficult for the government to build the infrastructure needed for them. The informal settlements that grew from this difficulty are now known as ger districts. There have been many efforts to formalize and develop these areas. The Law on Allocation of Land to Mongolian Citizens for Ownership, passed in 2002, allowed for existing ger district residents to formalize the land they settled, and allowed for others to receive land from the government into the future. Along with the privatization of land, the Mongolian government has been pushing for the development of ger districts into areas with housing blocks connected to utilities. The plan for this was published in 2014 as Ulaanbaatar 2020 Master Plan and Development Approaches for 2030. Although progress has been slow (Choi and Enkhbat 7), they have been making progress in building housing blocks in ger distrcts. Residents of ger districts sell or exchange their plots to developers who then build housing blocks on them. Often this is in exchange for an apartment in the building, and often the value of the apartment is less than the land they originally had (Choi and Enkhbat 15). Based on what I’ve read about the ger districts, they have been around since at least the 1970s, and progress on developing them has been slow. When ineffective policy results in a large chunk of the populace generationally living in yurts on the outskirts of urban areas, it’s clear that there is failure. Choi, Mack Joong, and Urandulguun Enkhbat. “Distributional Effects of Ger Area Redevelopment in Ulaanbaatar, Mongolia.” International Journal of Urban Sciences, vol. 24, no. 1, Jan. 2020, pp. 50–68. DOI.org (Crossref), https://doi.org/10.1080/12265934.2019.1571433.
  • 559 Stimmen
    99 Beiträge
    59 Aufrufe
    N
    In this year of 2025? No. But it still is basically setting oneself for failure from the perspective of Graphene, IMO. Like, the strongest protection in the world (assuming Graphene even is, which is quite a tall order statement) is useless if it only works on the mornings of a Tuesday that falls in a prime number day that has a blue moon and where there are no ATP tennis matches going on. Everyone else is, like, living in the real world, and the uniqueness of your scenario is going to go down the drain once your users get presented with a $5 wrench, or even cheaper: a waterboard. Because cops, let alone ICE, are not going to stop to ask you if they can make you more comfortable with your privacy being violated.
  • Diego

    Technology technology
    1
    1
    0 Stimmen
    1 Beiträge
    6 Aufrufe
    Niemand hat geantwortet
  • This Month in Redox - May 2025

    Technology technology
    1
    21 Stimmen
    1 Beiträge
    8 Aufrufe
    Niemand hat geantwortet
  • 278 Stimmen
    100 Beiträge
    26 Aufrufe
    F
    It's not just skills, it's also capital investment.
  • How the Signal Knockoff App TeleMessage Got Hacked in 20 Minutes

    Technology technology
    31
    1
    188 Stimmen
    31 Beiträge
    32 Aufrufe
    P
    Not to mention TeleMessage violated the terms of the GPL. Signal is under gpl and I can't find TeleMessage's code anywhere. Edit: it appears it is online somewhere just not in a github repo or anything https://micahflee.com/heres-the-source-code-for-the-unofficial-signal-app-used-by-trump-officials/