One of those rare lucid moments by the stock market? Is this the market correction that everyone knew was coming, or is some famous techbro going to technobabble some more about AI overlords and they return to their fantasy values?
It’s quite lucid. The new thing uses a fraction of compute compared to the old thing for the same results, so Nvidia cards for example are going to be in way less demand. That being said Nvidia stock was way too high surfing on the AI hype for the last like 2 years, and despite it plunging it’s not even back to normal.
If AI is cheaper, then we may use even more of it, and that would soak up at least some of the slack, though I have no idea how much.
How is the “fraction of compute” being verified? Is the model available for independent analysis?
Its freely availible with a permissive license, but I dont think that that claim has been verified yet.
And the data is not available. Knowing the weights of a model doesn’t really tell us much about its training costs.
Most rational market: Sell off NVIDIA stock after Chinese company trains a model on NVIDIA cards.
Anyways NVIDIA still up 1900% since 2020 …
how fragile is this tower?
Emergence of DeepSeek raises doubts about sustainability of western artificial intelligence boom
Is the “emergence of DeepSeek” really what raised doubts? Are we really sure there haven’t been lots of doubts raised previous to this? Doubts raised by intelligent people who know what they’re talking about?
Ah, but those “intelligent” people cannot be very intelligent if they are not billionaires. After all, the AI companies know exactly how to assess intelligence:
Microsoft and OpenAI have a very specific, internal definition of artificial general intelligence (AGI) based on the startup’s profits, according to a new report from The Information. … The two companies reportedly signed an agreement last year stating OpenAI has only achieved AGI when it develops AI systems that can generate at least $100 billion in profits. That’s far from the rigorous technical and philosophical definition of AGI many expect. (Source)
The economy rests on a fucking chatbot. This future sucks.
On the brightside, the clear fragility and lack of direct connection to real productive forces shows the instability of the present system.
And no matter how many protectionist measures that the US implements we’re seeing that they’re losing the global competition. I guess protectionism and oligarchy aren’t the best ways to accomplish the stated goals of a capitalist economy. How soon before China is leading in every industry?
This conclusion was foregone when China began to focus on developing the Productive Forces and the US took that for granted. Without a hard pivot, the US can’t even hope to catch up to the productive trajectory of China, and even if they do hard pivot, that doesn’t mean they even have a chance to in the first place.
In fact, protectionism has frequently backfired, and had other nations seeking inclusion into BRICS or more favorable relations with BRICS nations.
Economy =/= stock market
Almost like yet again the tech industry is run by lemming CEOs chasing the latest moss to eat.
The best part is that it’s open source and available for download
So can I have a private version of it that doesn’t tell everyone about me and my questions?
Yes
Yep, lookup ollama
Can someone with the knowledge please answer this question?
Yes, you can run a downgraded version of it on your own pc.
Apparently phone too! Like 3 cards down was another post linking to instructions on how to run it locally on a phone in a container app or termux. Really interesting. I may try it out in a vm on my server.
I watched one video and read 2 pages of text. So take this with a mountain of salt. From that I gathered that deepseek R1 is the model you interact with when you use the app. The complexity of a model is expressed as the number of parameters (though I don’t know yet what those are) which dictate its hardware requirements. R1 contains 670 bn Parameter and requires very very beefy server hardware. A video said it would be 10th of GPUs. And it seems you want much of VRAM on you GPU(s) because that’s what AI crave. I’ve also read 1BN parameters require about 2GB of VRAM.
Got a 6 core intel, 1060 6 GB VRAM,16 GB RAM and Endeavour OS as a home server.
I just installed Ollama in about 1/2 an hour, using docker on above machine with no previous experience on neural nets or LLMs apart from chatting with ChatGPT. The installation contains the Open WebUI which seems better than the default you got at ChatGPT. I downloaded the qwen2.5:3bn model (see https://ollama.com/search) which contains 3 bn parameters. I was blown away by the result. It speaks multiple languages (including displaying e.g. hiragana), knows how much fingers a human has, can calculate, can write valid rust-code and explain it and it is much faster than what i get from free ChatGPT.
The WebUI offers a nice feedback form for every answer where you can give hints to the AI via text, 10 score rating thumbs up/down. I don’t know how it incooperates that feedback, though. The WebUI seems to support speech-to-text and vice versa. I’m eager to see if this docker setup even offers APIs.
I’ll probably won’t use the proprietary stuff anytime soon.
Yeah, but you have to run a different model if you want accurate info about China.
Yeah but China isn’t my main concern right now. I got plenty of questions to ask and knowledge to seek and I would rather not be broadcasting that stuff to a bunch of busybody jackasses.
I agree. I don’t know enough about all the different models, but surely there’s a model that’s not going to tell you “<whoever’s> government is so awesome” when asking about rainfall or some shit.
Unfortunately it’s trained on the same US propaganda filled english data as any other LLM and spits those same talking points. The censors are easy to bypass too.
Yes but your server can’t handle the biggest LLM.
But Chinese…
They’d need to do some pretty fucking advanced hackery to be able to do surveillance on you just via the model. Everything’s possible I guess, but … yeah perhaps not.
If they could do that, essentially nothing you do on your computer would be safe.
This just shows how speculative the whole AI obsession has been. Wildly unstable and subject to huge shifts since its value isn’t based on anything solid.
Nvidia’s most advanced chips, H100s, have been banned from export to China since September 2022 by US sanctions. Nvidia then developed the less powerful H800 chips for the Chinese market, although they were also banned from export to China last October.
I love how in the US they talk about meritocracy, competition being good, blablabla… but they rig the game from the beginning. And even so, people find a way to be better. Fascinating.
You’re watching an empire in decline. It’s words stopped matching its actions decades ago.
deleted by creator
Remember to cancel your Microsoft 365 subscription to kick them while they’re down
Joke’s on them: I never started a subscription!
I don’t have one to cancel, but I might celebrate today by formatting the old windows SSD in my system and using it for some fast download cache space or something.
Good. LLM AIs are overhyped, overused garbage. If China putting one out is what it takes to hack the legs out from under its proliferation, then I’ll take it.
Cutting the cost by 97% will do the opposite of hampering proliferation.
Possibly, but in my view, this will simply accelerate our progress towards the “bust” part of the existing boom-bust cycle that we’ve come to expect with new technologies.
They show up, get overhyped, loads of money is invested, eventually the cost craters and the availability becomes widespread, suddenly it doesn’t look new and shiny to investors since everyone can use it for extremely cheap, so the overvalued companies lose that valuation, the companies using it solely for pleasing investors drop it since it’s no longer useful, and primarily just the implementations that actually improved the products stick around due to user pressure rather than investor pressure.
Obviously this isn’t a perfect description of how everything in the work will always play out in every circumstance every time, but I hope it gets the general point across.
It’s not about hampering proliferation, it’s about breaking the hype bubble. Some of the western AI companies have been pitching to have hundreds of billions in federal dollars devoted to investing in new giant AI models and the gigawatts of power needed to run them. They’ve been pitching a Manhattan Project scale infrastructure build out to facilitate AI, all in the name of national security.
You can only justify that kind of federal intervention if it’s clear there’s no other way. And this story here shows that the existing AI models aren’t operating anywhere near where they could be in terms of efficiency. Before we pour hundreds of billions into giant data center and energy generation, it would behoove us to first extract all the gains we can from increased model efficiency. The big players like OpenAI haven’t even been pushing efficiency hard. They’ve just been vacuuming up ever greater amounts of money to solve the problem the big and stupid way - just build really huge data centers running big inefficient models.
Overhyped? Sure, absolutely.
Overused garbage? That’s incredibly hyperbolic. That’s like saying the calculator is garbage. The small company where I work as a software developer has already saved countless man hours by utilising LLMs as tools, which is all they are if you take away the hype; a tool to help skilled individuals work more efficiently. Not to replace skilled individuals entirely, as Sam Dead eyes Altman would have you believe.
LLMs as tools,
Yes, in the same way that buying a CD from the store, ripping to your hard drive, and returning the CD is a tool.
Lol serves you right for pushing AI onto us without our consent
The determination to make us use it whether we want to or not really makes me resent it.
Hello darkness my old friend
We will have true AI once it is capable of answering “I don’t know” instead of making things up
Turns out, some people i know are apparently fake AI.
Which is actually something Deepseek is able to do.
Even if it can still generate garbage when used incorrectly like all of them, it’s still impressive that it will tell you it doesn’t “know” something, but can try to help if you give it more context. which is how this stuff should be used anyway.
and it’s open-source!
The funny thing is, this was unveiled a while ago and I guess investors only just noticed it.
Deepthink R1(the reasoning model) was only released on January 20. Still took a while though.
“wiped”? There was money and it ceased to exist?
The money went back into the hands of all the people and money managers who sold their stocks today.
Edit: I expected a bloodbath in the markets with the rhetoric in this article, but the NASDAQ only lost 3% and the DJIA was positive today…
Nvidia was significantly over-valued and was due for this. I think most people who are paying attention knew that
There’s been a lot of disproportionate hype around deepseek lately
Trump counterbalance keeping it in check but my gut is saying once tariffs come in February there’s going to be a market correction. Pure speculation on my part.
To be fair, NQ futures momentarily dropped 5% before recovering some. A few days from now on would be interesting.
It’s pixie dust
“off US stocks”
As a European, gotta say I trust China’s intentions more than the US’ right now.