• 1 Post
  • 24 Comments
Joined 2 years ago
cake
Cake day: July 5th, 2023

help-circle


  • golli@lemm.eetoWikipedia@lemmy.worldPope Leo XIV
    link
    fedilink
    English
    arrow-up
    4
    ·
    14 days ago

    From what little I’ve read more like Francis, which isn’t surprising considering he appointed the majority of those who just voted.

    Apparently he’s a bit more moderate. Similar focus on the poor and immigrants, which apparently Leo 13 was as well (name choice is also a statement, so reading up on whoever has the name before might be worthwhile). But slightly more conservative in regard to lgbtq and women.


  • There’s currently a Kickstarter going on for a watch that aims to be modular and repairable. It’s called UNA Watch.

    Look interesting, but imo with these things it’s a bit of a chicken and egg problem, where the upgradeability/repairability only has value, if it is actually provided in the future (and economically viable). Something that can only be proven in time, but requires people to trust it before.

    I’m not in the market for a new watch right now, since I just repaired the screen on my Garmin, but am keeping an eye on it, since sadly Garmin seems to have entered the early stages of enshittification.






  • If you don’t mind Meta/Facebook, then the oculus quest headsets are also very affordable hardware and deliver a good experience. I think the issue lies with content.

    Smartphones or handhelds like the steam deck with flat screens could use plenty of already existing content made for screens. With VR you want different content that is made specifically for it. There is a decent amount of games (but still much fewer than for other devices), but honestly not that much more.

    Additionally it also can only really be used at home, where most already have other devices.

    It’s a chicken and egg problem. But imo if there were more genuine unique productivity tasks and experiences available through VR, we would see more adoption.






  • Yes, it does make a difference, but like with many other things you should not fall into the trap of decision making paralysis.

    Your current instance lemmy.world is the largest and perfectly adequate for the majority. It also costs nothing to make accounts on other instances and is done in less time than it takes to decide between them.

    That said some reasons to choose one instance over another:

    Federation with other instances

    Some instances choose not to federate with others. Common reasons being political ideology or NSFW/piracy/violent content. Others might be more liberal and leave it up to their users to block whatever within their own apps. As someone already mentioned world seems to have defederated dbzer0 the piracy focused instance. Some political instances that often also get defederated are the far left ones like hexbear and Lemmy grad (seems like those are defederated by lemmy.world as well).

    Server location and performance

    Especially country focused instances like lemmy.ca or feddit.uk will have their servers in their own countries. You can of course access them from anywhere, but a European user might have a better time choosing an instance with servers located in Europe, while someone in America might have lower latency with one located there.

    Alternate frontends

    If you are accessing Lemmy through a browser rather than an app, you might enjoy alternate frontends that change the design. Your current one lemmy.world for example offers 4 different designs that can also be found in the sidebar alexandrite, photon, voyager mobile and one looking like old.reddit.

    General ideology of the instance

    Your choice of instance might also tell others something about you. If you choose a country specific instance people seeing your profile name might assume you come from there, if you choose one with a particular political view people will probably assume you hold similar views. Same goes for other instances that are related to things like sexuality or hobbies.



  • I’d say their recent releases were quite mixed.

    The Battlemage GPUs have decent performance at an attractive price for many consumers. But at the same time the CPU overhead problems are a big issue at exactly that price point.

    Arrow Lake had some great efficiency gains, but that was because previously it was terrible. Now it’s better, but still not even close to the likes of Apple. Great improvements on the efficiency cores and with that in some productivity tasks, but not much on the performance cores and latency seems to be a big issue. So that’s pretty mixed, especially when comparing it to AMD’s offerings.

    Lunar Lake seems imo is a very interesting product, but also apparently a one off. So seems like they won’t “keep doing that”.

    Sadly i’m not too knowledgeable in the probably more important data center space. Granite/Sierra Forest seem like quite decent products, so hopefully they’ll continue to keep improving there. Gaudi 3 i really don’t know much about, but i don’t think they sold much of those. And they just canceled their Falcon shores release and the next Jaguar Shores is probably in 2026, so nothing new in this year?


  • If we are talking the manufacturing side, rather than design/software i am very curious to see how SIMC develops. You are absolutely right that there is a big advantage for the second mover, since they can avoid dead ends and already know on an abstract level what is working. And diminishing returns also help make gaps be slightly less relevant.

    However i think we can’t just apply the same timeline to them and say “they have 7nm now” and it took others x years to progress from there to 5nm or 3nm, because these steps include the major shift from DUV to EUV, which was in the making for a very long time. And that’s a whole different beast compared to DUV, where they are also probably still relying on ASML machines for the smallest nodes (although i think producing those domestically is much more feasible). Eventually they’ll get there, but i think this isn’t trivial and will take more than 2 years for sure.

    On the design side vs Nvidia the hyperscalers like Alibaba/Tencent/Baidu or maybe even a smaller newcomer might be able to create something competitive for their specific usecases (like the Google TPUs). But Nvidia isn’t standing still either, so i think getting close to parity will be extremely hard there aswell.


    Of course, the price gap will shrink at the same rate as ROCm matures and customers feel its safe to use AMD hardware for training.

    Well to what degree ROCm matures and closes the gap is probably the question. Like i said, i agree that their hardware seems quite capable in many ways, although my knowledge here is quite limited. But AMD so far hasn’t really shown that they can compete with Nvidia on the software side.


    As far as Intel goes, being slow in my reply helps my point. Just today Intel canceled their next-generation GPU Falcon Shore, making it an internal development step only. As much as i am rooting for them, it will need a major shift in culture and talent for them to right the ship. Gaudi 3 wasn’t successful (i think they didn’t even meet their target of $500mio sales) and now they probably don’t have any release in 2025, assuming Jaguar Lake is 2026 since Falcon Shore was slated for end of this year. In my books that is the definition of being behind more than 1 year, considering they are not even close to parity right now.


  • Yeah. I don’t believe market value is a great indicator in this case. In general, I would say that capital markets are rational at a macro level, but not micro. This is all speculation/gambling.

    I have to concede that point to some degree, since i guess i hold similar views with Tesla’s value vs the rest of the automotive Industry. But i still think that the basic hirarchy holds true with nvidia being significantly ahead of the pack.

    My guess is that AMD and Intel are at most 1 year behind Nvidia when it comes to tech stack. “China”, maybe 2 years, probably less.

    Imo you are too optimistic with those estimations, particularly with Intel and China, although i am not an expert in the field.

    As i see it AMD seems to have a quite decent product with their instinct cards in the server market on the hardware side, but they wish they’d have something even close to CUDA and its mindshare. Which would take years to replicate. Intel wish they were only a year behind Nvidia. And i’d like to comment on China, but tbh i have little to no knowledge of their state in GPU development. If they are “2 years, probably less” behind as you say, then they should have something like the rtx 4090, which was released end of 2022. But do they have something that even rivals the 2000 or 3000 series cards?

    However, if you can make chips with 80% performance at 10% price, its a win. People can continue to tell themselves that big tech always will buy the latest and greatest whatever the cost. It does not make it true.

    But the issue is they all make their chips at the same manufacturer, TSMC, even Intel in the case of their GPUs. So they can’t really differentiate much on manufacturing costs and are also competing on the same limited supply. So no one can offer 80% of performance at 10% price, or even close to it. Additionally everything around the GPU (datacenters, rack space, power useage during operation etc.) also costs, so it is only part of the overall package cost and you also want to optimize for your limited space. As i understand it datacenter building and power delivery for them is actually another limiting factor right now for the hyperscalers.

    Google, Meta and Amazon already make their own chips. That’s probably true for DeepSeek as well.

    Google yes with their TPUs, but the others all use Nvidia or AMD chips to train. Amazon has their Graviton CPUs, which are quite competitive, but i don’t think they have anything on the GPU side. DeepSeek is way to small and new for custom chips, they evolved out of a hedge fund and just use nvidia GPUs as more or less everyone else.



  • It’s a reaction to thinking China has better AI

    I don’t think this is the primary reason behind Nvidia’s drop. Because as long as they got a massive technological lead it doesn’t matter as much to them who has the best model, as long as these companies use their GPUs to train them.

    The real change is that the compute resources (which is Nvidia’s product) needed to create a great model suddenly fell of a cliff. Whereas until now the name of the game was that more is better and scale is everything.

    China vs the West (or upstart vs big players) matters to those who are investing in creating those models. So for example Meta, who presumably spends a ton of money on high paying engineers and data centers, and somehow got upstaged by someone else with a fraction of their resources.