• Buffalox@lemmy.world
    link
    fedilink
    English
    arrow-up
    52
    ·
    edit-2
    4 hours ago

    Chinese companies are heavily incentivized to use Chinese chips instead of American since Trump blocked trade with China.
    China used to parallel import the chips they needed, and even repackage them with more onboard RAM, making more powerful Nvidia solutions available in China than in the rest of the world.
    But Trumps behavior towards China made the Chinese government decide to limit the use of American technologies for AI.
    There was a point where Nvidia exports to China was basically at a standstill, because China forbade the purchase of a new cut down Nvidia chip made for the Chinese market to circumvent American trade restrictions.

    China is building their own complete stack now, replacing everything with Chinese technologies, right from the AI chips to the entire AI software framework.

    So not only does Nvidia and other American companies lose hardware sales, the entire stack will be threatened with a Chinese alternative, that will likely compete with American options on the international market in the future. If Cuda loses its current dominance, it will be easier for competitors to take marketshare from Nvidia.

    Hopefully this will be good for consumers worldwide.

    • mannycalavera@feddit.uk
      link
      fedilink
      English
      arrow-up
      3
      ·
      4 hours ago

      Hopefully this will be good for consumers worldwide.

      Until America decides to tariff anyone using Chinese technology.

      • FaceDeer@fedia.io
        link
        fedilink
        arrow-up
        10
        ·
        4 hours ago

        It’ll be good for consumers worldwide. America is not the whole world.

        I, for example, am in Canada. We’ve established a bunch of very nice trade deals with China recently, we’re going to end up with access to a bunch of Chinese products that Americans can’t get due to their self-imposed trade war with China.

        • BrinkBreaker@lemmy.dbzer0.com
          link
          fedilink
          English
          arrow-up
          2
          ·
          2 hours ago

          I think they mean that the US would put trade pressure on countries doing any tech trade with China, not specifically preventing it punishing American companies from using Chinese chips.

          Unfortunately the United States is still a big economy regardless of their politics and Manny is right that the US would throw their weight behind anti China policies to the detriment of other nations.

          How successful such a move would be is up to debate.

          • FaceDeer@fedia.io
            link
            fedilink
            arrow-up
            5
            ·
            2 hours ago

            The US is already trying to throw its economic weight around bullying Canada, and we’ve already settled in to an effective economic defensive posture. Those trade deals with China are actually part of it, previously we were supporting various American initiatives to tariff China but the Americans tore up a bunch of agreements with us so we responded in kind. It’s unfortunate but they started it and we’re prepared to hold our own.

      • Otter@lemmy.ca
        link
        fedilink
        English
        arrow-up
        2
        ·
        3 hours ago

        America decides to tariff anyone using Chinese technology

        America already decides to tariff countries regardless of what they’re doing

  • brucethemoose@lemmy.world
    link
    fedilink
    English
    arrow-up
    18
    ·
    edit-2
    4 hours ago

    What’s left unsaid is the software architecture is extremely interesting, and efficient.


    Ironically, the Nvidia embargo was the best thing to ever happen to the Chinese labs (which Nvidia tried to tell the US govt). It forced them to get thrifty, unlike US labs which (allegedly) fill some GPU farms with busywork for the appearance of high utilization.

  • BETYU@moist.catsweat.com
    link
    fedilink
    arrow-up
    5
    arrow-down
    1
    ·
    4 hours ago

    the only thing deepseek is really good for is to fuck over NVIDIA and chat gpt or American ai in general because deep seek is bast on chat gpt after all anything that brakes this monopoly is good and anything associated with it like ram.

  • ag10n@lemmy.world
    link
    fedilink
    English
    arrow-up
    10
    ·
    5 hours ago

    You can run it on CPU alone. Not surprising they’re building their own AI ecosystem

    • brucethemoose@lemmy.world
      link
      fedilink
      English
      arrow-up
      7
      arrow-down
      1
      ·
      edit-2
      4 hours ago

      Not at scale. Even on the new architecture, one really needs some kind of accelerator to make it economical for servers.

      Bitnet-like models might change the calculus, but no major trainer had tried that yet.

      • [object Object]@lemmy.ca
        link
        fedilink
        English
        arrow-up
        6
        ·
        edit-2
        4 hours ago

        Even with a bitnet, it’s almost definitely better to train on a high precision float then refine down to bits.

        I would expect bitnet to require more layers for equivalent quality too.

        • brucethemoose@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          1
          ·
          4 hours ago

          I just meant for mass inference serving.

          Yeah, I haven’t seen much in the way of bitnet training savings yet, like regular old QAT. It does appear that Deepseek is finetuning their MoEs in a 4-bit format now, though.

      • ag10n@lemmy.world
        link
        fedilink
        English
        arrow-up
        2
        arrow-down
        1
        ·
        4 hours ago

        Yes, you can run it at scale. Which is why it uses Huawei hardware.

        You can run it on anything, scaled or not

        • theunknownmuncher@lemmy.world
          link
          fedilink
          English
          arrow-up
          2
          arrow-down
          1
          ·
          edit-2
          2 hours ago

          Nope! You don’t know what you’re talking about. At all. But you can have fun running a 1.6 trillion parameter model on CPU at basically 0 tokens per second at scale, MoE or not.

        • brucethemoose@lemmy.world
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          1
          ·
          edit-2
          4 hours ago

          Just not power/cost efficiently on CPU only, is what I meant. CPUs don’t have the compute for batching (running generation requests in parallel). You need an accelerator, like Huawei’s, to be economical.

          It’s fine for local inference, of course.

          • ag10n@lemmy.world
            link
            fedilink
            English
            arrow-up
            1
            ·
            3 hours ago

            A whole ecosystem that can run on any hardware, efficiently or not, is a whole ecosystem developed for the Chinese market