Skip to content

Apple sues YouTuber who leaked iOS 26’s new “Liquid Glass” software redesign

Technology
74 50 3
  • One Law to Rule Them All: The Iron Law of Software Performance

    Technology technology
    1
    1
    32 Stimmen
    1 Beiträge
    12 Aufrufe
    Niemand hat geantwortet
  • FairPhone AMA

    Technology technology
    5
    14 Stimmen
    5 Beiträge
    39 Aufrufe
    alcan@lemmy.worldA
    Ask Me Anything
  • The Decline of Usability: Revisited | datagubbe.se

    Technology technology
    8
    67 Stimmen
    8 Beiträge
    43 Aufrufe
    R
    I blame the idea of the 00s and 10s that there should be some "Zen" in computer UIs and that "Zen" is doing things wrong with the arrogant tone of "you don't understand it". Associated with Steve Jobs, but TBH Google as well. And also another idea of "you dummy talking about ergonomics can't be smarter than this big respectable corporation popping out stylish unusable bullshit". So - pretense of wisdom and taste, under which crowd fashion is masked, almost aggressive preference for authority over people actually having maybe some wisdom and taste due to being interested in that, blind trust into whatever tech authority you chose for yourself, because, if you remember, in the 00s it was still perceived as if all people working in anything connected to computers were as cool as aerospace engineers or naval engineers, some kind of elite, including those making user applications, objective flaw (or upside) of the old normal UIs - they are boring, that's why UIs in video games and in fashionable chat applications (like ICQ and Skype), not talking about video and audio players, were non-standard like always, I think the solution would be in per-application theming, not in breaking paradigms, again, like with ICQ and old Skype and video games, I prefer it when boredom is thought with different applications having different icons and colors, but the UI paradigm remains the same, I think there was a themed IE called LOTR browser which I used (ok, not really, I used Opera) to complement ICQ, QuickTime player and BitComet, all mentioned had standard paradigm and non-standard look.
  • 0 Stimmen
    1 Beiträge
    11 Aufrufe
    Niemand hat geantwortet
  • For All That Is Good About Humankind, Ban Smartphones

    Technology technology
    89
    1
    132 Stimmen
    89 Beiträge
    426 Aufrufe
    D
    Appreciated, but do you think the authorities want to win the war on drugs?
  • 131 Stimmen
    6 Beiträge
    43 Aufrufe
    P
    This is a tough one for me: I'm opposed to femicide, but I only wish the absolute worst on influencers.
  • Why doesn't Nvidia have more competition?

    Technology technology
    22
    1
    33 Stimmen
    22 Beiträge
    87 Aufrufe
    B
    It’s funny how the article asks the question, but completely fails to answer it. About 15 years ago, Nvidia discovered there was a demand for compute in datacenters that could be met with powerful GPU’s, and they were quick to respond to it, and they had the resources to focus on it strongly, because of their huge success and high profitability in the GPU market. AMD also saw the market, and wanted to pursue it, but just over a decade ago where it began to clearly show the high potential for profitability, AMD was near bankrupt, and was very hard pressed to finance developments on GPU and compute in datacenters. AMD really tried the best they could, and was moderately successful from a technology perspective, but Nvidia already had a head start, and the proprietary development system CUDA was already an established standard that was very hard to penetrate. Intel simply fumbled the ball from start to finish. After a decade of trying to push ARM down from having the mobile crown by far, investing billions or actually the equivalent of ARM’s total revenue. They never managed to catch up to ARM despite they had the better production process at the time. This was the main focus of Intel, and Intel believed that GPU would never be more than a niche product. So when intel tried to compete on compute for datacenters, they tried to do it with X86 chips, One of their most bold efforts was to build a monstrosity of a cluster of Celeron chips, which of course performed laughably bad compared to Nvidia! Because as it turns out, the way forward at least for now, is indeed the massively parralel compute capability of a GPU, which Nvidia has refined for decades, only with (inferior) competition from AMD. But despite the lack of competition, Nvidia did not slow down, in fact with increased profits, they only grew bolder in their efforts. Making it even harder to catch up. Now AMD has had more money to compete for a while, and they do have some decent compute units, but Nvidia remains ahead and the CUDA problem is still there, so for AMD to really compete with Nvidia, they have to be better to attract customers. That’s a very tall order against Nvidia that simply seems to never stop progressing. So the only other option for AMD is to sell a bit cheaper. Which I suppose they have to. AMD and Intel were the obvious competitors, everybody else is coming from even further behind. But if I had to make a bet, it would be on Huawei. Huawei has some crazy good developers, and Trump is basically forcing them to figure it out themselves, because he is blocking Huawei and China in general from using both AMD and Nvidia AI chips. And the chips will probably be made by Chinese SMIC, because they are also prevented from using advanced production in the west, most notably TSMC. China will prevail, because it’s become a national project, of both prestige and necessity, and they have a massive talent mass and resources, so nothing can stop it now. IMO USA would clearly have been better off allowing China to use American chips. Now China will soon compete directly on both production and design too.
  • 0 Stimmen
    2 Beiträge
    23 Aufrufe
    A
    How about right now? How's that going?