Skip to content

With a Trump-driven reduction of nearly 2,000 employees, F.D.A. will Use A.I. in Drug Approvals to ‘Radically Increase Efficiency’

Technology
91 69 429
  • Or maybe that is part of the allure of automation: the eschewing of human responsibility, such that any bias in decision making appears benign (the computer deemed it so, no one’s at fault) and any errors - if at all recognized as such - become simply a matter of bug-fixing or model fine-tuning. The more inscrutable the model the better in that sense. The computer becomes an oracle and no one’s to blame for its divinations.

  • Or maybe that is part of the allure of automation: the eschewing of human responsibility, such that any bias in decision making appears benign (the computer deemed it so, no one’s at fault) and any errors - if at all recognized as such - become simply a matter of bug-fixing or model fine-tuning. The more inscrutable the model the better in that sense. The computer becomes an oracle and no one’s to blame for its divinations.

    I am convinced that law enforcement wants intentionally biased AI decision makers so that they can justify doing what they’ve always done with the cover of “it’s not racist because a computer said so!”

    The scary part is most people are ignorant enough to buy it.

  • Your argument becomes idiotic once you understand the actual technology. The AI bullshit machine's agenda is "give nice answer" ("factual" is not an idea that has neural center in the AI brain), and "make reader happy". The human "bullshit" machine, has many agendas, but it would have not got so far if it was spouting just happy bullshit (but I guess America is a becoming a very special case).

    It doesn't. I understand the actual technology. There are applications of human decision making where it's possibly better.

  • Text to avoid paywall

    The Food and Drug Administration is planning to use artificial intelligence to “radically increase efficiency” in deciding whether to approve new drugs and devices, one of several top priorities laid out in an article published Tuesday in JAMA.

    Another initiative involves a review of chemicals and other “concerning ingredients” that appear in U.S. food but not in the food of other developed nations. And officials want to speed up the final stages of making a drug or medical device approval decision to mere weeks, citing the success of Operation Warp Speed during the Covid pandemic when workers raced to curb a spiraling death count.

    “The F.D.A. will be focused on delivering faster cures and meaningful treatments for patients, especially those with neglected and rare diseases, healthier food for children and common-sense approaches to rebuild the public trust,” Dr. Marty Makary, the agency commissioner, and Dr. Vinay Prasad, who leads the division that oversees vaccines and gene therapy, wrote in the JAMA article.

    The agency plays a central role in pursuing the agenda of the U.S. health secretary, Robert F. Kennedy Jr., and it has already begun to press food makers to eliminate artificial food dyes. The new road map also underscores the Trump administration’s efforts to smooth the way for major industries with an array of efforts aimed at getting products to pharmacies and store shelves quickly.

    Some aspects of the proposals outlined in JAMA were met with skepticism, particularly the idea that artificial intelligence is up to the task of shearing months or years from the painstaking work of examining applications that companies submit when seeking approval for a drug or high-risk medical device.

    “I don’t want to be dismissive of speeding reviews at the F.D.A.,” said Stephen Holland, a lawyer who formerly advised the House Committee on Energy and Commerce on health care. “I think that there is great potential here, but I’m not seeing the beef yet.”

    Final stage capitalism: Purging all the experts (at catching bullshit from appllicants) before the agencies train the AI with newb level inputs.

  • It doesn't. I understand the actual technology. There are applications of human decision making where it's possibly better.

    LLM does no decision making. At all. It spouts (as you say) bullshit. If there is enough training data for "Trump is divine", the LLM will predict that Trump is divine, with no second thought (no first thought either). And it's not even great to use as a language-based database.

    Please don't even consider LLMs as "AI".

  • Or maybe that is part of the allure of automation: the eschewing of human responsibility, such that any bias in decision making appears benign (the computer deemed it so, no one’s at fault) and any errors - if at all recognized as such - become simply a matter of bug-fixing or model fine-tuning. The more inscrutable the model the better in that sense. The computer becomes an oracle and no one’s to blame for its divinations.

    I saw a paper a while back that argued that AI is being used as "moral crumple zones". For example, an AI used for health insurance acts allows for the company to reject medically necessary procedures without employees incurring as much moral injury as part of that (even low level customer service reps are likely to find comfort in being able to defer to the system.). It's an interesting concept that I've thought about a lot since I found it.

  • LLM does no decision making. At all. It spouts (as you say) bullshit. If there is enough training data for "Trump is divine", the LLM will predict that Trump is divine, with no second thought (no first thought either). And it's not even great to use as a language-based database.

    Please don't even consider LLMs as "AI".

    Even an RNG does decision-making.

    I know what LLMs are, thank you very much!

    If you wanted to even understand my initial point, you already would have.

    Things have become really grim if people who can't read a small message are trying to teach me on fundamentals of LLMs.

  • Oh good, a 60% chance you’ll get an ineffective or killer drug because they’ll use AI to analyze the usage and AI to report on it.

    If it actually ends up being an AI and not just some Trump cuck stooge masquerading as AI picking the drug by the company that gave the largest bribe to Trump, I 100% guarantee this AI is trained only on papers written by non-peer reviewed drug company paid "scientists" containing made up narratives.

    Those of us prescribed the drugs will be the guinea pigs because R&D costs money and hits the bottom line. The many deaths will be conveniently scape-goated on "the AI" the morons in charge promised is smarter and more efficient than a person.

    Fuck this shit.

  • Even an RNG does decision-making.

    I know what LLMs are, thank you very much!

    If you wanted to even understand my initial point, you already would have.

    Things have become really grim if people who can't read a small message are trying to teach me on fundamentals of LLMs.

    I wouldn't define flipping coins as decision making. Especially when it comes to blanket governmental policy that has the potential to kill (or severely disable) millions of people.

    You seem to not want any people to teach you anything. And are somehow completely dejected at such perceived actions.

  • Different types of AI, different training data, different expectations and outcomes. Generative AI is but one use case.

    It's already been proven a useful tool in research, when directed and used correctly by an expert. It's a tool, to give to scientists to assist them, not replace them.

    If you're goal to use AI to replace people, you've got a bad surprise coming.

    If you're not equipping your people with the skills and tools of AI, your people will become obsolete in short time.

    Learn AI and how to utilize it as a tool, you can train your own model on your own private data and locally interrogate the model to do unique analysis typically not possible in realtime. Learn the goods and bads of technology and let your ethics guide how you use it, but stop dismissing revolutionary technology because the earlier generative models weren't reinforced enough get fingers right.

    I'm not dismissing its use. It is a useful tool, but it cannot replace experts at this point, or maybe ever (and I'm gathering you agree on this).

    If it ever does get to that point, we need to also remedy the massive social consequences of revoking those same experts' ability to have sufficient income to have a reasonable living.

    I was being a little silly for effect.

  • I saw a paper a while back that argued that AI is being used as "moral crumple zones". For example, an AI used for health insurance acts allows for the company to reject medically necessary procedures without employees incurring as much moral injury as part of that (even low level customer service reps are likely to find comfort in being able to defer to the system.). It's an interesting concept that I've thought about a lot since I found it.

    I can absolutely see that. And I don’t think it’s AI-specific, it’s got to do with relegating responsibility to a machine. Of course AI in the guise of LLMs can make things worse with its low interpretability, where it might be even harder to trace anything back to an executive or clerical decision.

  • I wouldn't define flipping coins as decision making. Especially when it comes to blanket governmental policy that has the potential to kill (or severely disable) millions of people.

    You seem to not want any people to teach you anything. And are somehow completely dejected at such perceived actions.

    You seem to not want any people to teach you anything.

    No, I don't seem that. I don't like being ascribed opinions I haven't expressed.

    I wouldn’t define flipping coins as decision making. Especially when it comes to blanket governmental policy that has the potential to kill (or severely disable) millions of people.

    When your goal is to avoid a certain most harmful subset of such decisions, and living humans always being pressured by power and corrupt profit to pick that subset, flipping coins is preferable, if that's the two variants between which we are choosing.

  • Oh good, a 60% chance you’ll get an ineffective or killer drug because they’ll use AI to analyze the usage and AI to report on it.

    That is an underestimate, since it doesn't factor in the knockdown effect of the more lax regulations having, so people will try to sell all kinds of crap as "medicine".

  • Text to avoid paywall

    The Food and Drug Administration is planning to use artificial intelligence to “radically increase efficiency” in deciding whether to approve new drugs and devices, one of several top priorities laid out in an article published Tuesday in JAMA.

    Another initiative involves a review of chemicals and other “concerning ingredients” that appear in U.S. food but not in the food of other developed nations. And officials want to speed up the final stages of making a drug or medical device approval decision to mere weeks, citing the success of Operation Warp Speed during the Covid pandemic when workers raced to curb a spiraling death count.

    “The F.D.A. will be focused on delivering faster cures and meaningful treatments for patients, especially those with neglected and rare diseases, healthier food for children and common-sense approaches to rebuild the public trust,” Dr. Marty Makary, the agency commissioner, and Dr. Vinay Prasad, who leads the division that oversees vaccines and gene therapy, wrote in the JAMA article.

    The agency plays a central role in pursuing the agenda of the U.S. health secretary, Robert F. Kennedy Jr., and it has already begun to press food makers to eliminate artificial food dyes. The new road map also underscores the Trump administration’s efforts to smooth the way for major industries with an array of efforts aimed at getting products to pharmacies and store shelves quickly.

    Some aspects of the proposals outlined in JAMA were met with skepticism, particularly the idea that artificial intelligence is up to the task of shearing months or years from the painstaking work of examining applications that companies submit when seeking approval for a drug or high-risk medical device.

    “I don’t want to be dismissive of speeding reviews at the F.D.A.,” said Stephen Holland, a lawyer who formerly advised the House Committee on Energy and Commerce on health care. “I think that there is great potential here, but I’m not seeing the beef yet.”

    it's what ai is supposed to be used for, but it mabye isn't good enough

  • It doesn't. I understand the actual technology. There are applications of human decision making where it's possibly better.

    It kinda seems like you don’t understand the actual technology.

  • OpenAI’s Sam Altman warns of AI voice fraud crisis in banking

    Technology technology
    54
    1
    236 Stimmen
    54 Beiträge
    134 Aufrufe
    V
    I am rich so write what I said.
  • One Law to Rule Them All: The Iron Law of Software Performance

    Technology technology
    1
    1
    32 Stimmen
    1 Beiträge
    15 Aufrufe
    Niemand hat geantwortet
  • 68 Stimmen
    17 Beiträge
    82 Aufrufe
    H
    Set up arrs, you basically set it and forget it.
  • What was Radiant AI, anyway?

    Technology technology
    6
    1
    20 Stimmen
    6 Beiträge
    41 Aufrufe
    T
    In fact Daggerfall was almost nothing but quests and other content like that.
  • Where are all the data centres and why should you care?

    Technology technology
    5
    1
    63 Stimmen
    5 Beiträge
    35 Aufrufe
    A
    Ai says Virginia is home to the largest data center market in the world, with over 576 data centers, primarily located in Northern Virginia,
  • 92 Stimmen
    42 Beiträge
    191 Aufrufe
    G
    You don’t understand. The tracking and spying is the entire point of the maneuver. The ‘children are accessing porn’ thing is just a Trojan horse to justify the spying. I understand what are you saying, I simply don't consider to check if a law is applied as a Trojan horse in itself. I would agree if the EU had said to these sites "give us all the the access log, a list of your subscriber, every data you gather and a list of every IP it ever connected to your site", and even this way does not imply that with only the IP you could know who the user is without even asking the telecom company for help. So, is it a Trojan horse ? Maybe, it heavily depend on how the EU want to do it. If they just ask "show me how you try to avoid that a minor access your material", which normally is the fist step, I don't see how it could be a Trojan horse. It could become, I agree on that. As you pointed out, it’s already illegal for them to access it, and parents are legally required to prevent their children from accessing it. No, parents are not legally required to prevent it. The seller (or provider) is legally required. It is a subtle but important difference. But you don’t lock down the entire population, or institute pre-crime surveillance policies, just because some parents are not going to follow the law. True. You simply impose laws that make mandatories for the provider to check if he can sell/serve something to someone. I mean asking that the cashier of mall check if I am an adult when I buy a bottle of wine is no different than asking to Pornhub to check if the viewer is an adult. I agree that in one case is really simple and in the other is really hard (and it is becoming harder by the day). You then charge the guilty parents after the offense. Ok, it would work, but then how do you caught the offendind parents if not checking what everyone do ? Is it not simpler to try to prevent it instead ?
  • Meta Reportedly Eyeing 'Super Sensing' Tech for Smart Glasses

    Technology technology
    4
    1
    34 Stimmen
    4 Beiträge
    30 Aufrufe
    M
    I see your point but also I just genuinely don't have a mind for that shit. Even my own close friends and family, it never pops into my head to ask about that vacation they just got back from or what their kids are up to. I rely on social cues from others, mainly my wife, to sort of kick start my brain. I just started a new job. I can't remember who said they were into fishing and who didn't, and now it's anxiety inducing to try to figure out who is who. Or they ask me a friendly question and I get caught up answering and when I'm done I forget to ask it back to them (because frequently asking someone about their weekend or kids or whatever is their way of getting to share their own life with you, but my brain doesn't think that way). I get what you're saying. It could absolutely be used for performative interactions but for some of us people drift away because we aren't good at being curious about them or remembering details like that. And also, I have to sit through awkward lunches at work where no one really knows what to talk about or ask about because outside of work we are completely alien to one another. And it's fine. It wouldn't be worth the damage it does. I have left behind all personally identifiable social media for the same reason. But I do hate how social anxiety and ADHD makes friendship so fleeting.
  • The Document Foundation is proud to release LibreOffice 25.2.3

    Technology technology
    7
    1
    265 Stimmen
    7 Beiträge
    53 Aufrufe
    somethingburger@jlai.luS
    View -> User Interface -> Tabs It already exists but is nowhere near as good as MS Office (like everything with LO).