Skip to content

Why doesn't Nvidia have more competition?

Technology
22 14 87
  • This post did not contain any content.

    Because its competitors care about Not Invented Here instead of building common industry standards.

  • What a shit article. Doesn't explain the software situation. While CUDA is the most popular, a lot of frameworks do support AMD chips.

    A comically bad "article".

  • Naji said the firm has also “developed the broadest ecosystem” of developers and software.

    “And so it's just so much easier to … build an application, build an AI model on top of those chips,” he said.

    Expounding, Nvidia has very deeply engrained itself in educational and research institutions. People learning GPU compute are being taught CUDA and Nvidia hardware. Researchers have access to farms of Nvidia chips.

    AMD has basically gone the "build it and they will come" attitude, and the results to match.

  • Naji said the firm has also “developed the broadest ecosystem” of developers and software.

    “And so it's just so much easier to … build an application, build an AI model on top of those chips,” he said.

    It's literally the most surface level take. Does not even mention what CUDA is or AMD's efforts to run it

    https://www.xda-developers.com/nvidia-cuda-amd-zluda/

    But it is no longer funded by AMD or Intel

    AMD GPUs are still supported by frameworks like PyTorch

    While Nvidia might be the fastest, they are not always the cheapest option, especially if you rent it in the cloud. When I last checked, it was cheaper to rent AMD GPUs

  • This post did not contain any content.

    Corporate consolidation and monopoly/oligopoly is usually why.

  • This post did not contain any content.

    How to create a successful GPU company in 2025:

    • Step 1: build a time machine and go back 30 years
  • Because its competitors care about Not Invented Here instead of building common industry standards.

    Well, Intel tried with OneApi. As for AMD they go 5 minutes between every time they shoot themselves in the foot. It's unbelievable to watch.

  • Well, Intel tried with OneApi. As for AMD they go 5 minutes between every time they shoot themselves in the foot. It's unbelievable to watch.

    Intel tried with OpenAPI because ROCm was not invented here.

  • Expounding, Nvidia has very deeply engrained itself in educational and research institutions. People learning GPU compute are being taught CUDA and Nvidia hardware. Researchers have access to farms of Nvidia chips.

    AMD has basically gone the "build it and they will come" attitude, and the results to match.

    AMD has basically gone the “build it and they will come” attitude

    Except they didn't.

    They repeatedly fumble the software with little mistakes (looking at you, Flash Attention). They price the MI300X, W7900, and any high VRAM GPU through the roof, when they have every reason to be more competitive and undercut Nvidia. They have sad, incomplete software efforts divorced from what devs are actually doing, like their quantization framework or some inexplicably bad LLMs they trained themself. I think Strix Halo is the only GPU compute thing they did half right recently, and they still screwed that up.

    They give no one any reason to give them a chance, and wonder why no one comes. Lisa Su could fix this with literally like three phone calls (remove VRAM restrictions on their OEMs, cut pro card prices, fix stupid small bugs in ROCM), but she doesn't. It's inexplicable.

  • How to create a successful GPU company in 2025:

    • Step 1: build a time machine and go back 30 years

    What do i do before I did that?

  • What do i do before I did that?

    Have rich parents, pale skin, etc.

  • Have rich parents, pale skin, etc.

    No that comes afterwards

  • AMD has basically gone the “build it and they will come” attitude

    Except they didn't.

    They repeatedly fumble the software with little mistakes (looking at you, Flash Attention). They price the MI300X, W7900, and any high VRAM GPU through the roof, when they have every reason to be more competitive and undercut Nvidia. They have sad, incomplete software efforts divorced from what devs are actually doing, like their quantization framework or some inexplicably bad LLMs they trained themself. I think Strix Halo is the only GPU compute thing they did half right recently, and they still screwed that up.

    They give no one any reason to give them a chance, and wonder why no one comes. Lisa Su could fix this with literally like three phone calls (remove VRAM restrictions on their OEMs, cut pro card prices, fix stupid small bugs in ROCM), but she doesn't. It's inexplicable.

    That's basically what I said in so many words.
    AMD is doing its own thing, if you want what Nvidia offers you're gonna have to build it yourself.
    WRT pricing, I'm pretty sure AMD is typically a fraction of the price of Nvidia hardware on the enterprise side, from what I've read, but companies that have made that leap have been unhappy since AMD's GPU enterprise offerings were so unreliable.

    The biggest culprit from what I can gather is that AMD's GPU firmware/software side is basically still ATI camped up in Markham, divorced from the rest of the company in Austin that is doing great work with their CPU-side.

  • This post did not contain any content.

    People like youz starts snoopin' around askin' questions tends to fall outta windows, y'know what I'm sayin'?

  • That's basically what I said in so many words.
    AMD is doing its own thing, if you want what Nvidia offers you're gonna have to build it yourself.
    WRT pricing, I'm pretty sure AMD is typically a fraction of the price of Nvidia hardware on the enterprise side, from what I've read, but companies that have made that leap have been unhappy since AMD's GPU enterprise offerings were so unreliable.

    The biggest culprit from what I can gather is that AMD's GPU firmware/software side is basically still ATI camped up in Markham, divorced from the rest of the company in Austin that is doing great work with their CPU-side.

    WRT pricing, I’m pretty sure AMD is typically a fraction of the price of Nvidia hardware on the enterprise side

    I'm not as sure about this, but seems like AMD is taking a fat margin on the MI300X (and its sucessor?), and kinda ignoring the performance penalty. It's easy to say "build it yourself!" but the reality is very few can, or will, do this, and will simply try to deploy vllm or vanilla TRL or something as best they can (and run into the same issues everyone does).

    The 'enthusiast' side where all the university students and tinkerer devs reside is totally screwed up though. AMD is mirroring Nvidia's VRAM cartel pricing when they have absolutely no reason to. It's completely bonkers. AMD would be in a totally different place right now if they had sold 40GB/48GB 7900s for an extra $200 (instead of price matching an A6000).

    The biggest culprit from what I can gather is that AMD’s GPU firmware/software side is basically still ATI camped up in Markham, divorced from the rest of the company in Austin that is doing great work with their CPU-side.

    Yeah, it does seem divorced from the CPU division. But a lot of the badness comes from business decisions, even when the silicon is quite good, and some of that must be from Austin.

  • This post did not contain any content.

    At first I was going to say there is ATI.
    Then I realized I hadn't heard about ATI in a while and looked up what happened to it.
    Then I realized... I'm old.

  • WRT pricing, I’m pretty sure AMD is typically a fraction of the price of Nvidia hardware on the enterprise side

    I'm not as sure about this, but seems like AMD is taking a fat margin on the MI300X (and its sucessor?), and kinda ignoring the performance penalty. It's easy to say "build it yourself!" but the reality is very few can, or will, do this, and will simply try to deploy vllm or vanilla TRL or something as best they can (and run into the same issues everyone does).

    The 'enthusiast' side where all the university students and tinkerer devs reside is totally screwed up though. AMD is mirroring Nvidia's VRAM cartel pricing when they have absolutely no reason to. It's completely bonkers. AMD would be in a totally different place right now if they had sold 40GB/48GB 7900s for an extra $200 (instead of price matching an A6000).

    The biggest culprit from what I can gather is that AMD’s GPU firmware/software side is basically still ATI camped up in Markham, divorced from the rest of the company in Austin that is doing great work with their CPU-side.

    Yeah, it does seem divorced from the CPU division. But a lot of the badness comes from business decisions, even when the silicon is quite good, and some of that must be from Austin.

    The ‘enthusiast’ side where all the university students and tinkerer devs reside is totally screwed up though. AMD is mirroring Nvidia’s VRAM cartel pricing when they have absolutely no reason to. It’s completely bonkers. AMD would be in a totally different place right now if they had sold 40GB/48GB 7900s for an extra $200 (instead of price matching an A6000).

    Eh, the biggest issue here is that most (post-secondary) students probably just have a laptop for whatever small GPGPU learning they're doing, which is overwhelmingly dominated by Nvidia. For grad students they'll have access to the institution resources, which is also dominated by Nvidia (this has been a concerted effort).

    Only a few that explicitly pursue AMD hardware will end up with it, but that also requires significant foundational work for the effort. So the easiest path for research is throw students at CUDA and Nvidia hardware.

    Basically, Nvidia has entrenched itself in the research/educational space, and that space is slow moving (Java is still the de facto CS standard, with only slow movements to Python happening at some universities), so I don't see much changing, unless AMD decides it's very hungry and wants to chase the market.

    Lower VRAM prices could help, but the truth is people and intuitions are willing to pay more (obviously) for plug and play.

  • The ‘enthusiast’ side where all the university students and tinkerer devs reside is totally screwed up though. AMD is mirroring Nvidia’s VRAM cartel pricing when they have absolutely no reason to. It’s completely bonkers. AMD would be in a totally different place right now if they had sold 40GB/48GB 7900s for an extra $200 (instead of price matching an A6000).

    Eh, the biggest issue here is that most (post-secondary) students probably just have a laptop for whatever small GPGPU learning they're doing, which is overwhelmingly dominated by Nvidia. For grad students they'll have access to the institution resources, which is also dominated by Nvidia (this has been a concerted effort).

    Only a few that explicitly pursue AMD hardware will end up with it, but that also requires significant foundational work for the effort. So the easiest path for research is throw students at CUDA and Nvidia hardware.

    Basically, Nvidia has entrenched itself in the research/educational space, and that space is slow moving (Java is still the de facto CS standard, with only slow movements to Python happening at some universities), so I don't see much changing, unless AMD decides it's very hungry and wants to chase the market.

    Lower VRAM prices could help, but the truth is people and intuitions are willing to pay more (obviously) for plug and play.

    I dunno. From my more isolated perspective on GitHub and small LLM testing circles, I see a lot of 3090s, 4090s, sometimes arrays of 3060s/3090s or old P40s or MI50s, which people got basically for the purpose of experimentation and development because they can't drop (or at least justify) $5K.

    They would 100% drop that money on at least one 7900 48GB instead (as the sheer capacity is worth it over the speed hit and finickiness), and then do a whole bunch of bugfixing/testing on them. I know I would. Hence the Framework Strix Halo thing is sold out even though it's... rather compute-lite compared to a 3090+ GPU.

    It seems like a tiny market, but a lot of the frameworks/features/models being developed by humble open source devs filter up to the enterprise space. You'd absolutely see more enterprise use once the toolkits were hammered out on desktops... But they aren't, because AMD gives us no incentive to do so. A 7900 is just not worth the trouble over a 3090/4090 if its VRAM capacity is the same, and this (more or less) extends up and down the price ranges.

  • This post did not contain any content.

    It’s funny how the article asks the question, but completely fails to answer it.

    About 15 years ago, Nvidia discovered there was a demand for compute in datacenters that could be met with powerful GPU’s, and they were quick to respond to it, and they had the resources to focus on it strongly, because of their huge success and high profitability in the GPU market.

    AMD also saw the market, and wanted to pursue it, but just over a decade ago where it began to clearly show the high potential for profitability, AMD was near bankrupt, and was very hard pressed to finance developments on GPU and compute in datacenters. AMD really tried the best they could, and was moderately successful from a technology perspective, but Nvidia already had a head start, and the proprietary development system CUDA was already an established standard that was very hard to penetrate.

    Intel simply fumbled the ball from start to finish. After a decade of trying to push ARM down from having the mobile crown by far, investing billions or actually the equivalent of ARM’s total revenue. They never managed to catch up to ARM despite they had the better production process at the time. This was the main focus of Intel, and Intel believed that GPU would never be more than a niche product.
    So when intel tried to compete on compute for datacenters, they tried to do it with X86 chips, One of their most bold efforts was to build a monstrosity of a cluster of Celeron chips, which of course performed laughably bad compared to Nvidia! Because as it turns out, the way forward at least for now, is indeed the massively parralel compute capability of a GPU, which Nvidia has refined for decades, only with (inferior) competition from AMD.

    But despite the lack of competition, Nvidia did not slow down, in fact with increased profits, they only grew bolder in their efforts. Making it even harder to catch up.

    Now AMD has had more money to compete for a while, and they do have some decent compute units, but Nvidia remains ahead and the CUDA problem is still there, so for AMD to really compete with Nvidia, they have to be better to attract customers. That’s a very tall order against Nvidia that simply seems to never stop progressing. So the only other option for AMD is to sell a bit cheaper. Which I suppose they have to.

    AMD and Intel were the obvious competitors, everybody else is coming from even further behind.
    But if I had to make a bet, it would be on Huawei. Huawei has some crazy good developers, and Trump is basically forcing them to figure it out themselves, because he is blocking Huawei and China in general from using both AMD and Nvidia AI chips. And the chips will probably be made by Chinese SMIC, because they are also prevented from using advanced production in the west, most notably TSMC.
    China will prevail, because it’s become a national project, of both prestige and necessity, and they have a massive talent mass and resources, so nothing can stop it now.

    IMO USA would clearly have been better off allowing China to use American chips. Now China will soon compete directly on both production and design too.

  • How Android phones became an earthquake warning system

    Technology technology
    7
    1
    77 Stimmen
    7 Beiträge
    21 Aufrufe
    G
    Fake news! Everyone knows that "The Gays" are what causes earthquakes, tornados and floods, so just need Grindr! ::: spoiler Spoiler it's sarcasm, in case it's not obvious enough. :::
  • 797 Stimmen
    220 Beiträge
    136 Aufrufe
    uriel238@lemmy.blahaj.zoneU
    algos / AI has already been used to justify racial discrimination in some counties who use predictive policing software to adjust the sentences of convicts (the software takes in a range of facts about the suspect and the incident and compares it to how prior incidents and suspects were similar features were adjudicated) and wouldn't you know it, it simply highlighted and exaggerated the prejudices of police and the courts to absurdity, giving whites absurdly lighter sentences than nonwhites, for example. This is essentially mind control or coercion technology based on the KGB technology of компромат (Kompromat, or compromising information, or as CIA calls it biographical leverage, ) essentially, information about a person that can be used either to jeopardize their life, blackmail material or means to lure and bribe them. Take this from tradecraft and apply it to marketing or civil control, and you get things like the Social Credit System in China to keep people from misbehaving, engaging in discontent and coming out of the closet (LGBTQ+ but there are plenty of other applicable closets). From a futurist perspective, we homo-sapiens appear just incapable of noping out of a technology or process, no matter how morally black or heinous that technology is, we'll use it, especially those with wealth and power to evade legal prosecution (or civil persecution). It breaks down into three categories: Technologies we use anyway, and suffer, e.g. usury, bonded servitude, mass-media propaganda distribution Technologies we collectively decide are just not worth the consequences, e.g. the hydrogen bomb, biochemical warfare Technologies for which we create countermeasures, usually turning into a tech race between states or between the public and the state, e.g. secure communication, secure data encryption, forbidden data distribution / censorship We're clearly on the cusp of mind control and weaponizing data harvesting into a coercion mechanism. Currently we're already seeing it used to establish and defend specific power structures that are antithetical to the public good. It's currently in the first category, and hopefully it'll fall into the third, because we have to make a mess (e.g. Castle Bravo / Bikini Atol) and clean it up before deciding not to do that again. Also, with the rise of the internet, we've run out of myths that justify capitalism, which is bonded servitude with extra steps. So we may soon (within centuries) see that go into one of the latter two categories, since the US is currently experiencing the endgame consequences of forcing labor, and the rest of the industrialized world is having to bulwark from the blast.
  • Say Hello to the World's Largest Hard Drive, a Massive 36TB Seagate

    Technology technology
    261
    1
    612 Stimmen
    261 Beiträge
    1k Aufrufe
    spacecadet@feddit.nlS
    Write speeds on SMR drives start to stagnate after mere gigabytes written, not after terabytes. As soon as the CMR cache is full, you're fucked, and it stagnates to utterly unusable speeds as it's desperately trying to balance writing out blocks to the persistent area of the disk and accepting new incoming writes. I have 25 year old consumer level IDE drives that perform better than an SMR drive in this thrashing state. Also, I often use hard drives as a temporary holding area for stuff that I'm transferring around for one reason or another and that absolutely sucks if an operation that normally takes an hour or two is suddenly becoming a multi-day endeavour tying up my computing resources. I was burned once when Seagate submarined SMR drives into the Barracuda line, and I got a drive that was absolutely unfit for purpose. Never again.
  • 147 Stimmen
    55 Beiträge
    244 Aufrufe
    01189998819991197253@infosec.pub0
    I meant to download from the official Microsoft site. Kudos on getting your mum on Linux! I was unable to keep mine on it : / Maybe I'm missing something, but this is from the "Download Windows 11 Disk Image (ISO) for x64 devices" section from the official Microsoft site, but I don't see any option to buy or mention of it: Before you begin downloading an ISO Make sure you have: An internet connection (internet service provider fees may apply). Sufficient data storage available on the computer, USB, or external drive you are downloading the .iso file to. A blank DVD disc with at least 8GB (and DVD burner) to create a bootable disc. We recommend using a blank USB or blank DVD, because any content on it will be deleted during installation. If you receive a “disc image file is too large” message while attempting to burn a DVD bootable disc from an ISO file, consider using a higher capacity Dual Layer DVD.
  • One-Click RCE in ASUS's Preinstalled Driver Software

    Technology technology
    9
    30 Stimmen
    9 Beiträge
    55 Aufrufe
    M
    Yeah, Lemmy has a VERY large Linux user base, which means Windows discussions tend to get mocked or dismissed. But the reality is that Windows is still the dominant OS for the vast majority of users, by leaps and bounds. Linux runs the world’s infrastructure, but Windows is what the average user boots up every day. “This exploit only works on the average user’s OS. And it only works if the user clicks the “yes” button to escalate permissions, which they have been conditioned to always do without question. Obviously this isn’t an exploit to worry about.”
  • Where are all the data centres and why should you care?

    Technology technology
    5
    1
    63 Stimmen
    5 Beiträge
    35 Aufrufe
    A
    Ai says Virginia is home to the largest data center market in the world, with over 576 data centers, primarily located in Northern Virginia,
  • Google confirms more ads on your paid YouTube Premium Lite soon

    Technology technology
    273
    1
    942 Stimmen
    273 Beiträge
    1k Aufrufe
    undefined@lemmy.hogru.chU
    I had to look it up, what I was thinking of was MQA. Looks like they discontinued it last year though.
  • 0 Stimmen
    2 Beiträge
    21 Aufrufe
    V
    Here's how you know it's not ready: AI hasn't replaced a single CEO.