US chip-maker Nvidia led a rout in tech stocks Monday after the emergence of a low-cost Chinese generative AI model that could threaten US dominance in the fast-growing industry.

The chatbot developed by DeepSeek, a startup based in the eastern Chinese city of Hangzhou, has apparently shown the ability to match the capacity of US AI pace-setters for a fraction of the investments made by American companies.

Shares in Nvidia, whose semiconductors power the AI industry, fell more than 15 percent in midday deals on Wall Street, erasing more than $500 billion of its market value.

The tech-rich Nasdaq index fell more than three percent.

AI players Microsoft and Google parent Alphabet were firmly in the red while Meta bucked the trend to trade in the green.

DeepSeek, whose chatbot became the top-rated free application on Apple’s US App Store, said it spent only $5.6 million developing its model – peanuts when compared with the billions US tech giants have poured into AI.

US “tech dominance is being challenged by China,” said Kathleen Brooks, research director at trading platform XTB.

The focus is now on whether China can do it better, quicker and more cost effectively than the US, and if they could win the AI race,” she said.

US venture capitalist Marc Andreessen has described DeepSeek’s emergence as a “Sputnik moment” – when the Soviet Union shocked Washington with its 1957 launch of a satellite into orbit.

As DeepSeek rattled markets, the startup on Monday said it was limiting the registration of new users due to “large-scale malicious attacks” on its services.

Meta and Microsoft are among the tech giants scheduled to report earnings later this week, offering opportunity for comment on the emergence of the Chinese company.

Shares in another US chip-maker, Broadcom, fell 16 percent while Dutch firm ASML, which makes the machines used to build semiconductors, saw its stock tumble 6.7 percent.

Investors have been forced to reconsider the outlook for capital expenditure and valuations given the threat of discount Chinese AI models,” David Morrison, senior market analyst at Trade Nation.

These appear to be as good, if not better, than US versions.”

Wall Street’s broad-based S&P 500 index shed 1.7 percent while the Dow was flat at midday.

In Europe, the Frankfurt and Paris stock exchanges closed in the red while London finish flat.

Asian stock markets mostly slid.

Just last week following his inauguration, Trump announced a $500 billion venture to build infrastructure for AI in the United States led by Japanese giant SoftBank and ChatGPT-maker OpenAI.

SoftBank tumbled more than eight percent in Tokyo on Monday while Japanese semiconductor firm Advantest was also down more than eight percent and Tokyo Electron off almost five percent.

  • UnderpantsWeevil@lemmy.world
    link
    fedilink
    English
    arrow-up
    58
    arrow-down
    15
    ·
    3 days ago

    The number of people repeating “I bet it won’t tell you about Tianamen Square” jokes around this news has - imho - neatly explained why the US tech sector is absolutely fucked going into the next generation.

      • Not_mikey@slrpnk.net
        link
        fedilink
        arrow-up
        16
        arrow-down
        3
        ·
        3 days ago

        It’s even worse / funnier in the app, it will generate the response, then once it realizes its about Taiwan it will delete the whole response and say sorry I can’t do that.

        If you ask it “what is the republic of china” it will generate a couple paragraphs of the history of China, then it’ll get a couple sentences in about the retreat to Taiwan and then stop and delete the response.

        • Womble@lemmy.world
          link
          fedilink
          English
          arrow-up
          16
          ·
          edit-2
          3 days ago

          In fairness that is also exactly what chatgpt, claude and the rest do for their online versions too when you hit their limits (usually around sex). IIRC they work by having a second LLM monitor the output and send a cancel signal if they think its gone over the line.

          • JasSmith@sh.itjust.works
            link
            fedilink
            arrow-up
            7
            arrow-down
            4
            ·
            2 days ago

            Okay but one is about puritanical Western cultural standards about sex, and one is about government censorship to maintain totalitarian power. One of these things is not like the other.

        • Womble@lemmy.world
          link
          fedilink
          English
          arrow-up
          6
          arrow-down
          2
          ·
          3 days ago

          Maybe they should have been clearer than saying people were joking about it doing something that it actually does if they wanted to make a point.

          • Bronzebeard@lemm.ee
            link
            fedilink
            English
            arrow-up
            7
            arrow-down
            4
            ·
            2 days ago

            People caring more about “China bad” instead of looking at what the tech they made can actually do is the issue.

            You needing this explicitly spelled out for you does not help the case.

            • ikt@aussie.zone
              link
              fedilink
              English
              arrow-up
              1
              ·
              2 days ago

              ngl I’m still confused

              what the tech they made can actually do

              It’s AI, it does AI things, is it because China can now do the things we do (coding/development/search queries etc) that are just as good as America that it’s a problem?

              • Bronzebeard@lemm.ee
                link
                fedilink
                English
                arrow-up
                2
                ·
                1 day ago

                It has nothing to do with it being China. They just figured out how to do it more efficiently and with lower powered chips, meaning nvidia’s market dominance in high end chips that they could charge whatever they wanted for just for is legs cut out from under them. If you don’t need as many to run AI, Nvidia won’t sell as many.

                • ikt@aussie.zone
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  1 day ago

                  So the idea with this comment:

                  The number of people repeating “I bet it won’t tell you about Tianamen Square” jokes around this news has - imho - neatly explained why the US tech sector is absolutely fucked going into the next generation.

                  is that people have misplaced their concern, not at the fact that it’s censored but that the US has lost the technology high ground and won’t get it back for at least a generation?

        • ikt@aussie.zone
          link
          fedilink
          English
          arrow-up
          4
          arrow-down
          2
          ·
          3 days ago

          I’m slow, what’s the point? how does people joking about the fact China is censoring output explain

          why the US tech sector is absolutely fucked going into the next generation

          • 小莱卡@lemmygrad.ml
            link
            fedilink
            English
            arrow-up
            2
            ·
            2 days ago

            Because they care more about the model not parroting US state dept narratives than the engineering behind it.

      • Smokeydope@lemmy.world
        link
        fedilink
        English
        arrow-up
        7
        arrow-down
        2
        ·
        edit-2
        3 days ago

        Try an abliterated version of the qwen 14b or 32b R1 distills. I just tried it out they will give you a real overview.

        Still even when abliterated its just not very knowledgeable about “harmful information”. If you want a truly uncensored model hit up mistral small 22b and its even more uncensored fine tune Beepo 22b

      • Scolding7300@lemmy.world
        link
        fedilink
        arrow-up
        2
        arrow-down
        1
        ·
        3 days ago

        That’s just dumb. It at least doesn’t suppress that when provided with search results/refuses to search (at least when integrated in Kagi)

        • Womble@lemmy.world
          link
          fedilink
          English
          arrow-up
          19
          ·
          3 days ago

          ??? you dont use training data when running models, that’s what is used in training them.

            • Womble@lemmy.world
              link
              fedilink
              English
              arrow-up
              35
              arrow-down
              1
              ·
              3 days ago

              Wow ok, you really dont know what you’re talking about huh?

              No I dont have thousands of almost top of the line graphics cards to retain an LLM from scratch, nor the millions of dollars to pay for electricity.

              I’m sure someone will and I’m glad this has been open sourced, its a great boon. But that’s still no excuse to sweep under the rug blatant censorship of topics the CCP dont want to be talked about.

              • UnderpantsWeevil@lemmy.world
                link
                fedilink
                English
                arrow-up
                5
                arrow-down
                23
                ·
                3 days ago

                No I dont have thousands of almost top of the line graphics cards to retain an LLM from scratch

                Fortunately, you don’t need thousands of top of the line cards to train the DeepSeek model. That’s the innovation people are excited about. The model improves on the original LLM design to reduce time to train and time to retrieve information.

                Contrary to common belief, an LLM isn’t just a fancy Wikipedia. Its a schema for building out a graph of individual pieces of data, attached to a translation tool that turns human-language inputs into graph-search parameters. If you put facts about Tianamen Square in 1989 into the model, you’ll get them back as results through the front-end.

                You don’t need to be scared of technology just because the team that introduced the original training data didn’t configure this piece of open-source software the way you like it.

                that’s still no excuse to sweep under the rug blatant censorship of topics the CCP dont want to be talked about.

                Wow ok, you really dont know what you’re talking about huh?

                • Womble@lemmy.world
                  link
                  fedilink
                  English
                  arrow-up
                  20
                  ·
                  edit-2
                  3 days ago

                  https://www.analyticsvidhya.com/blog/2024/12/deepseek-v3/

                  Huh I guess 6 million USD is not millions eh? The innovation is it’s comparatively cheap to train, compared to the billions OpenAI et al are spending (and that is with acquiring thousands of H800s not included in the cost).

                  Edit: just realised that was for the wrong model! but r1 was trained in the same budget https://x.com/GavinSBaker/status/1883891311473782995?mx=2

                  • UnderpantsWeevil@lemmy.world
                    link
                    fedilink
                    English
                    arrow-up
                    2
                    arrow-down
                    16
                    ·
                    edit-2
                    3 days ago

                    The innovation is it’s comparatively cheap to train, compared to the billions

                    Smaller builds with less comprehensive datasets take less time and money. Again, this doesn’t have to be encyclopedic. You can train your model entirely on a small sample of material detailing historical events in and around Beijing in 1989 if you are exclusively fixated on getting results back about Tienanmen Square.

                  • JasSmith@sh.itjust.works
                    link
                    fedilink
                    arrow-up
                    2
                    ·
                    2 days ago

                    Because the parent comment by Womble is about using the Chinese hosted DeepSeek app, not hosting the model themselves. The user above who responded either didn’t read the original comment carefully enough, or provided a very snarky response. Neither is particularly endearing.