▲This reminds me of the recent LaurieWired video presenting a hypothetical of, "what if we stopped making CPUs":
https://www.youtube.com/watch?v=L2OJFqs8bUkSpoiler, but the answer is basically that old hardware rules the day because it lasts longer and is more reliable of timespans of decades.
DDR5 32GB is currently going for ~$330 on Amazon
DDR4 32GB is currently going for ~$130 on Amazon
DDR3 32GB is currently going for ~50 on Amazon (4x8GB)
For anyone where cost is a concern, using older hardware seems like a particularly easy choice, especially if a person is comfortable with a Linux environment, since the massive droves of recently retired Windows 10 incompatible hardware works great with your Linux distro of choice.
reply▲If everyone went for DDR4 and DDR3, surely the cost would go up. There is no additional supply there, as they are no longer being made.
reply▲Currently trying to source a large amount of DDR4 to upgrade a 3 year old fleet of servers at a very unfortunate time.
It's very difficult to source in quantity, and is going up in price more or less daily at this point. Vendor quotes are good for hours, not days when you can find it.
reply▲Have you considered dropping by in person to your nearest computer recycler/refurbisher? As a teen I worked at one, and the boxes and boxes of RAM sticks pulled from scrapped machines (usually scrapped due to bad main boards) made a strong impression. They tend to not even put the highest-spec stuff they pull into anything, as refurbished machines are mostly sold/donated in quantity (to schools and the like) and big customers want standardized SKUs with interchangeable specs for repairability more than they want performance. Workers at these places are often invited to build personal machines to take home out of these “too good” parts — and yet there’s so much that they can’t possibly use it all up that way. If someone showed up asking if they could have some DDR3, they might literally just let you dig through the box and then sell it to you by weight!
reply▲We're looking to buy 1TB of DDR5 for a large database server. I'm pretty sure that we're just going to pay the premium and move on.
reply▲georgeburdell3 hours ago
[-] And that’s why everybody’s B2B these days. The decision-making people at companies are not spending their own money
reply▲At some point that’s true, but don’t they run the n-1 or 2 generation production lines for years after the next generation launches? There’s a significant capital investment there and my understanding is that the cost goes down significantly over the lifetime as they dial in the process so even though the price is lower it’s still profitable.
reply▲Unless plans have changed, the foundries making DDR4 are winding down, with the last shipments going out as we speak.
reply▲Isn’t that mostly because Chinese manufacturers are flooding the market with cheaper products?
reply▲The specialty DRAM vendors (Nanya notably) will keep making DDR4 and earlier
reply▲This is only true as long as there's enough of a market left. You tend to end up with oversupply and excessive inventory during the transition, and that pushes profit margins negative and removes all the supply pretty quickly.
reply▲Undoubtably the cost would go up, but nobody is building out datacenters full of DDR4, either, so I don't figure it would go up nearly as much as DDR5 is right now.
reply▲reply▲Awesome charts, thanks! I think it bears out that the multiplier for older hardware isn't as extreme as the newer hardware, right?
reply▲~2.8x for DDR4, ~3.6x for DDR5. DDR5 is still being made, though, so it will be interesting to see how it changes in the future.
Either way, it's going to be a long few years at the least.
reply▲Unless the AI bubble pops.
reply▲One possible outcome is the remaining memory manufacturers have dedicated all their capacity for AI and when the bubble pops, they lose their customer and they go out of business too.
reply▲I wouldn't be too surprised to find at least some of the big ram foundries are deeply bought into the fake money circles where everybody is "paying" each other with unrealised equity in OpenAI/Anthropic/whoever, resulting in a few trillion dollars worth of on-paper "money" vanishing overnight, at which stage a whole bunch of actual-money loans will get called in and billion dollar companies get gutted by asset strippers.
Maybe larger makerspaces and companies like Adafruit, RasPi, and Pine should start stockpiling (real) money, and pick themselves up an entire fab full of gear at firesale prices so they can start making their own ram...
reply▲And then the RAM factories... get sold to another business, if that's more profitable than dismantling them.
reply▲That's the Avarage price for new DDR4 which has dwindling supply. Meanwhile used DDR4 is being retired in both desktops and data centers.
reply▲DDR4 production is winding down or done. Only "new old stock" will remain, or used DDR4 modules. Good luck buying that in quantity.
reply▲xboxnolifes3 hours ago
[-] Unfortunately, older RAM also means an older motherboard, which also means older socket and older CPUs. It works, but it's not usually a drop in replacement.
reply▲maximilianthe13 hours ago
[-] Can't you use DDR3 in DDR5 compatible board?
reply▲No, the different generations have fundamentally incompatible interfaces, not just mechanically, but in terms of voltage and signaling.
reply▲forthefuture2 hours ago
[-] Unfortunately not, each version of RAM uses a different physical slot.
reply▲Yes, DDR3 is the lowest CAS latency and lasts ALOT longer.
Just like SSDs from 2010 have 100.000 writes per bit instead of below 10.000.
CPUs might even follow the same durability pattern but that remains to be seen.
Keep your old machines alive and backed up!
reply▲Aurornis50 minutes ago
[-] > Yes, DDR3 is the lowest CAS latency and lasts ALOT longer.
DDR5 is more reliable. Where are you getting this info that DDR3 lasts longer?
DDR5 runs at lower voltages, uses modern processes, and has on-die ECC.
This is already showing up in reduced failure rates for DDR5 fleets: https://ieeexplore.ieee.org/document/11068349
The other comment already covered why comparing CAS latency is misleading. CAS latency is measured in clock cycles. Multiply by the length of a clock cycle to get the CAS delay.
reply▲CAS latency is specified in cycles and clock rates are increasing, so despite the number getting bigger there's actually been a small improvement in latency with each generation.
reply▲Not for small amounts of data.
Bandwith increases, but if you only need a few bytes DDR3 is faster.
Also slower speed means less heat and longer life.
You can feel the speed advantage by just moving the mouse on a DDR3 PC...
reply▲While it has nothing to do with how responsive your mouse feels, as that is measured in milliseconds while CAS latency is measured in nanoseconds, there has indeed been a small regression with DDR5 memory compared to the 3 previous generations. The best DDR2-4 configurations could fetch 1 word in about 6-7 ns while the best DDR5 configurations take about 9-10 ns.
https://en.wikipedia.org/wiki/CAS_latency#Memory_timing_exam...
reply▲RAM latency doesn't affect mouse response in any perceptible way. The fastest gaming mice I know of run at 8000Hz, so that's 125000ns between samples, much bigger than any CAS latency. And most mice run substantially slower.
Maybe your old PC used lower-latency GUI software, e.g. uncomposited Xorg instead of Wayland.
reply▲I only felt it on Windows, maybe tht is due to the special USB mouse drivers Microsoft made? Still motion-to-photon latency is really lower on my DDR3 PCs, would be cool to know why.
reply▲CyberDildonics2 hours ago
[-] You are conflating two things that have nothing to do with each other. Computers have had mice since the 80s.
Still motion-to-photon latency is really lower on my DDR3 PCs, would be cool to know why.
No it isn't, your computer is doing tons of stuff and the cursor on windows is a hardware feature of the graphics card.
Should I even ask why you think memory bandwidth is the cause of mouse latency?
reply▲> 100.000 writes per bit
per cell*
Also, that SSD example is wildly untrue. Especially with the context of available capacity at the time. You CAN get modern SSD's with mind boggling write endurance per cell, AND has multides more cells, resulting in vastly more durable media than what was available pre 2015. The one caveat there to modern stuff being better than older stuff is Optane (the enterprise stuff like the 905P or P5800X, not that memory and SSD combo shitshow that Intel was shoveling out the consumer door). We still haven't reached parity with the 3DXpoint stuff, and it's a damn shame Intel hurt itself in it's confusion and cancelled that, because boy would they and Micron be printing money hand over fist right now if they were still making them. Still, Point being: Not everything is a TLC/QLC 0.3DWPD disposable drive like has become standard in the consumer space. If you want write endurance, capacity, and/or performance, you have more and better options today than ever before (Optane/3DXPoint excepted).
Regarding CPU's, they still follow that durability pattern if you unfuck what Intel and AMD are doing with boosting behavior and limit them to perform with the margins that they used to "back in the day". This is more of a problem on the consumer side (Core/Ryzen) than the enterprise side (Epyc/Xeon). It's also part of why the OC market is dying (save for maybe the XOC market that is having fun with LN2), those CPU's (especially consumer ones) come from the factory with much less margin for pushing things, because they're already close to their limit without exceedingly robust cooling.
I have no idea what the relative durability of RAM is tbh, it's been pretty bulletproof in my experience over the years, or at least bulletproof enough for my usecases that I haven't really noticed a difference. Notable exception is what I see in GPU's, but that is largely heat-death related and often a result of poor QA by the AIB that made it (eg, thermal pads not making contact with the GDDR modules).
reply▲Maybe but in my experience a good old <100GB SSD from 2010-14 will completely demolish any >100GB from 2014+ in longevity.
Some say they have the opposite experience, mine is ONLY Intel drives, maybe that is why.
X25-E is the diamond peak of SSDs probably forever since the machines to make 45nm SLC are gone.
reply▲What if you overprovision the newer SSD to a point where it can run the entirety of the drive in pseudo-SLC ("caching") mode? (You'd need to store no more than 25% of the nominal capacity, since QLC has four bits per cell.) That should have fairly good endurance, though still a lot less than Optane/XPoint persistent memory.
reply▲CAS latency doesn't matter so much as ns of total random-access latency and the raw clockspeed of the individual RAM cells. If you are accessing the same cell repeatedly, RAM hasn't gotten faster in years (around DDR2 IIRC).
reply▲Nice, My current PC uses DDR4
Time to dust off my 2012 PC and put Linux on it.
reply▲I bought 384 GB of DDR5-4800 RDIMM a few months back for a Zen 4c system with lots of I/O like dual 25G NIC's and ten MCIO x8 ports... So far it has been the best value for money compared to any memory before it. The bandwidth is nuts. Power consumption went DOWN compared to DDR4. Doesn't matter much if you got two sticks, but as soon as you get into 6+ territory, it does matter a lot. The same goes for NVMe's.
reply▲a 2x16 ddr4 kit I bought in 2020 for $160 is now $220. older memory is relatively cheap but not cheaper than before at all.
reply▲I wondered how much of this is inflation -- after adjusting for CPI inflation, $160 in 2020 is worth $200 in today's dollars [$], so the price of that ddr4 kit is 10% higher in real terms.
[$] https://www.usinflationcalculator.com/
reply▲USD is also weaker than it was in the past.
reply▲A friend built a new rig and went with DDR4 and a 5800x3d just because of this as he needed a lot of ram.
reply▲I have exactly this and was planning to upgrade to AM5 soon :'(
Looks like AM4 will live on for many more years in my flat
reply▲I think the OpenAI deal to lock wafers was a wonderful coup. OpenAI is more and more losing ground against the regularity[0] of the improvements coming from Anthropic, Google and even the open weights models. By creating a chock point at the hardware level, OpenAI can prevent the competition from increasing their reach because of the lack of hardware.
[0]: For me this is really an important part of working with Claude, the model improves with the time but stay consistent, its "personality" or whatever you want to call it, has been really stable over the past versions, this allows a very smooth transition from version N to N+1.
reply▲I don't see this working for Google though, since they make their own custom hardware in the form of the TPUs. Unless those designs include components that are also susceptible?
reply▲That was why OpenAI went after the wafers, not the finished products. By buying up the supply of the raw materials they bottleneck everybody, even unrelated fields. It's the kind of move that requires a true asshole to pull off, knowing it will give your company an advantage but screw up life for literally billions of people at the same time.
reply▲Went after the right component too. RAM manufacturers love an opportunity to create as much scarcity as possible.
reply▲> By buying up the supply
We actually don't know for certain whether these agreements are binding. If OpenAI gets in a credit crunch we'll soon find out.
reply▲TPUs use HBM, which are impacted.
reply▲Still susceptible, TPUs need DRAM dies just as much as anything else that needs to process data. I think they use some form of HBM, so they basically have to compete alongside the DDR supply chain.
reply▲Could this generate pressure to produce less memory hungry models?
reply▲There has always been pressure to do so, but there are fundamental bottlenecks in performance when it comes to model size.
What I can think of is that there may be a push toward training for exclusively search-based rewards so that the model isn't required to compress a large proportion of the internet into their weights. But this is likely to be much slower and come with initial performance costs that frontier model developers will not want to incur.
reply▲Yeah that was my unspoken assumption. The pressure here results in an entirely different approach or model architecture.
If openAI is spending $500B then someone can get ahead by spending $1B which improves the model by >0.2%
I bet there's a group or three that could improve results a lot more than 0.2% with $1B.
reply▲I wonder if this maintains the natural language capabilities which are what LLM's magic to me. There is a probably some middle ground, but not having to know what expressions, or idiomatic speech an LLM will understand is really powerful from a user experience point of view.
reply▲UncleOxidant3 hours ago
[-] Or maybe models that are much more task-focused? Like models that are trained on just math & coding?
reply▲isn't that what the mixture of experts trick that all the big players do is? Bunch of smaller, tightly focused models
reply▲> exclusively search-based rewards so that the model isn't required to compress a large proportion of the internet into their weights.
That just gave me an idea! I wonder how useful (and for what) a model would be if it was trained using a two-phase approach:
1) Put the training data through an embedding model to create a giant vector index of the entire Internet.
2) Train a transformer LLM but instead only utilising its weights, it can also do lookups against the index.
Its like a MoE where one (or more) of the experts is a fuzzy google search.
The best thing is that adding up-to-date knowledge won’t require retraining the entire model!
reply▲> so that the model isn't required to compress a large proportion of the internet into their weights.
The knowledge compressed into an LLM is a byproduct of training, not a goal. Training on internet data teaches the model to talk at all. The knowledge and ability to speak are intertwined.
reply▲Is anyone else deeply perturbed by the realization that a single unprofitable corporation can basically buy out the entire world's supply of computing hardware so nobody else can have it?
How did we get here? What went so wrong?
reply▲They're simply making a bet that they can put the DRAM dies to more valuable use than any of the existing alternatives, including e.g. average folks playing the latest videogames on their gaming rig. At this kind of scale, they had better be right or they are toast: they have essentially gone all-in on their bet that this whole AI thing is not going to 'pop' anytime soon.
reply▲> They're simply making a bet that they can put the DRAM dies to more valuable use than any of the existing alternatives
They can't. They know they can't. We all know they can't. But they can just keep abusing the infinite money glitch to price everyone else out, so it doesn't matter.
reply▲Phelinofist3 hours ago
[-] > By creating a chock point at the hardware level, OpenAI can prevent the competition from increasing their reach because of the lack of hardware
I already hate OpenAI, you don't have to convince me
reply▲hnuser1234564 hours ago
[-] Sure, but if the price is being inflated by inflated demand, then the suppliers will just build more factories until they hit a new, higher optimal production level, and prices will come back down, and eventually process improvements will lead to price-per-GB resuming its overall downtrend.
reply▲Micron has said they're not scaling up production. Presumably they're afraid of being left holding the bag when the bubble does pop
reply▲reply▲I mean it says on the page
>help ensure U.S. leadership in memory development and manufacturing, underpinning a national supply chain and R&D ecosystem.
It's more political than supply based
reply▲Future demand aka DDR6.
The 2027 timeline for the fab is when DDR6 is due to hit market.
reply▲Not just Micron, SK Hynix has made similar statements (unfortunately I can only find sources in Korean).
DRAM manufacturers got burned multiple times in the past scaling up production during a price bubble, and it appears they've learned their lesson (to the detriment of the rest of us).
reply▲Hedging is understandable. But what I don't understand is why they didn't hedge by keeping Crucial around but more dormant (higher prices, less SKUs, etc)
reply▲daemonologist2 hours ago
[-] The theory I've heard is built on the fact that China (CXMT) is starting to properly get into DRAM manufacturing - Micron might expect that to swamp the low end of the market, leaving Crucial unprofitable regardless, so they might as well throw in the towel now and make as much money as possible from AI/datacenter (which has bigger margins) while they can.
But yeah even if that's true I don't know why they wouldn't hedge their bets a bit.
reply▲So position Crucial as a premium brand, raise prices 4x instead of 3x, and drastically cut down on the SKUs to reduce overhead. If they tried that and kept spiraling into fewer and fewer SKUs and sales, I could understand it. But the discontinuation felt pretty abrupt.
reply▲Memory fabs take billions of dollars and years to build, also the memory business is a tough one where losses are common, so no such relief in sight.
With a bit of luck OpenAI collapses under its own weight sooner than later, otherwise we're screwed for several years.
reply▲Chip factories need years of lead time, and manufacturers might be hesitant to take on new debt in a massive bubble that might pop before they ever see any returns.
reply▲Please explain to me like I am five: Why does OpenAI need so much RAM?
2024 production was (according to openai/chatgpt) 120 billion gigabytes. With 8 billion humans that's about 15 GB per person.
reply▲What they need is not so much memory but memory bandwidth.
For training, their models have a certain number of memory needed to store the parameters, and this memory is touched for every example of every iteration. Big models have 10^12 (>1T )parameters, and with typical values of 10^3 examples per batch, and 10^6 number of iteration. They need ~10^21 memory accesses per run. And they want to do multiple runs.
DDR5 RAM bandwidth is 100G/s = 10^11, Graphics RAM (HBM) is 1T/s = 10^12. By buying the wafer they get to choose which types of memory they get.
10^21 / 10^12 = 10^9s = 30 years of memory access (just to update the model weights), you need to also add a factor 10^1-10^3 to account for the memory access needed for the model computation)
But the good news is that it parallelize extremely well. If you parallelize you 1T parameters, 10^3 times, your run time is brought down to 10^6 s = 12 days. But you need 10^3 *10^12 = 10^15 Bytes of RAM by run for weight update and 10^18 for computation (your 120 billions gigabytes is 10^20, so not so far off).
Are all these memory access technically required : No if you use other algorithms, but more compute and memory is better if money is not a problem.
Is it strategically good to deprive your concurrents from access to memory : Very short-sighted yes.
It's a textbook cornering of the computing market to prevent the emergence of local models, because customers won't be able to buy the minimal RAM necessary to run the models locally even just the inferencing part (not the training). Basically a war on people where little Timmy won't be able to get a RAM stick to play computer games at Xmas.
reply▲Thanks - but this seems like fairly extreme speculation.
> if money is not a problem.
Money is a problem, even for them.
reply▲large language models are large and must be loaded into memory to train or to use for inference if we want to keep them fast. older models like gpt3 have around 175 billion parameters. at float32s that comes out to something like 700GB of memory. newer models are even larger. and openai wants to run them as consumer web services.
reply▲I mean, I know that much. The numbers still don't make sense to me. How is my internal model this wrong?
For one, if this was about inference, wouldn't the bottleneck be the GPU computation part?
reply▲This "memory shortage" is not about AI companies needing main memory (which you plug into mainboards), but manufacturers are shifting their production capacities to other types of memory that will go onto GPUs. That brings supply for other memory products down, increasing their market price.
reply▲Concurrency?
Suppose some some parallelized, distributed task requires 700GB of memory (I don't know if it does or does not) per node to accomplish, and that speed is a concern.
A singular pile of memory that is 700GB is insufficient not because it lacks capacity, but instead because it lacks scalability. That pile is only enough for 1 node.
If more nodes were added to increase speed but they all used that same single 700GB pile, then RAM bandwidth (and latency) gets in the way.
reply▲daemonologist3 hours ago
[-] The conspiracy theory (which, to be clear, may be correct) is that they don't actually need so much RAM, but they know they and all their competitors do still need quite a bit of RAM. By buying up all the memory supply they can, for a while, keep everyone else from being able to add compute capacity/grow their business/compete.
reply▲codybontecou4 hours ago
[-] This became very clear with the outrage, rather than excitement, of forcing users to upgrade to ChatGPT-5 over 4o.
reply▲reply▲yes. on the moore's law is dead podcast they were talking about rumors where some 'AI enterprise company's representatives' were trying to buy memory in bulk from brick and mortar stores. in some cases openai was mentioned. crazy if true. also interesting considering none of those would be ECC certified like what you would opt for for a commercial server.
reply▲The author does mention:
> The reason for all this, of course, is AI datacenter buildouts.
This bubble can't pop fast enough. I'm curious to see which actually useful AIs remain after the burst, and how expensive they are once the cash-burning subsides.
reply▲Perhaps we'll have to start optimizing software for performance and RAM usage again.
I look at MS Teams currently using 1.5GB of RAM doing nothing.
reply▲RIP electron apps and PWAs. Need to go native, as chromium based stuff is so memory hungry. PWAs on Safari use way less memory, but PWA support in Safari is not great.
reply▲I, for one, would not miss a single one of the electron apps I'm forced to use.
Every single one of them makes me feel like the vendor is telling me "we can't be bothered employing half decent developers or giving the developers we have enough time and resources to write decent software, so we're just going to use cheap and inexperience web developers and burn another gigabyte or two of your memory to run what could easily be a sub 100MB native app."
At least now I'll have significant numbers to tell my boss: "Sure, we can continue to use Slack/VSCode/Teams/Figma/Postman/ - but each of those is going to require an additional GB or two of memory on every staff member's computer - which at today's pricing is over $500 in ram per laptop which are all on a 18-24 month replacement cycle. So that's maybe a million dollars a year in hardware budget to run those 5 applications across the whole team. We'll need to ensure we have signoff on that expenditure before we renew our subscriptions for those apps."
reply▲BobaFloutist1 hour ago
[-] Your laptops have an 18-24 month replacement cycle? What are you guys doing to the poor things?
reply▲Likewise, Slack. Currently at 2.3GB.
I'm very curious if this AI wave will actually lead to more native apps being made, since the barrier to doing so will be lower.
reply▲Old codger here, happy to rewrite any of your bloated software to run on a twenty year old PC at reasonable rates..
reply▲now that's a word i haven't heard in a while...optimising
reply▲If you used it yourself, even if only to ask out loud whether the performance is acceptable enough not to require the extra work, you'd hear it all the time. Be the change you want to see!
reply▲> optimising
Usually followed by "stock holder value"
reply▲And to the tech staff, some out of context and misunderstood quote from semi technical management which starts with the word "Premature"
reply▲PCIe 5.0 NVMe drives used as page cache can help a lot.
reply▲I truly hate how bloated and inefficient MS/Windows is.
My hope is that with the growing adoption of Linux that MS takes note...
reply▲That's a very indirect way of enjoying the benefits of Linux.
reply▲Can we get web sites to optimize as well? I use a slower laptop and a lot of sites are terrible. My old Sparc station (40mhz) had a better web experience in 1997 because than people cared about this.
reply▲So you’ve hated Windows for the last 30 years?
reply▲Anyone want to start a fab with me? We can buy an ASML machine and figure out the rest as we go. Toronto area btw
reply▲A dozen or so well-resourced tech titans in China are no doubt asking themselves this same question right now.
Of course, it takes quite some time for a fab to go from an idea to mass production. Even in China. Expect prices to drop 2-3 years from now when all the new capacity comes online?
reply▲My napkin math:
According to my research, these machines can etch around 150 wafers per hour and each wafer can fit around 50 top-of-the-line GPUs. This means we can produce around 7500 AI chips per hour. Sell them for $1k a piece. That's $7.5 million per hour in revenue. Run the thing for 3 days and we recover costs.
I'm sure there's more involved but that sounds like a pretty good ROI to me.
reply▲What about the $10b to build the facility (including clean air/water/chemicals/etc)?
reply▲Rent a warehouse.
reply▲It would be cheaper to bulldoze the warehouse and start over.
reply▲Rent a warehouse in one of the non Han dominated areas of China, where you can use all you want from the city's drinking water supply and pump all your used chemicals into the nearby river. Make sure to totally automate your robotic production line so you don't need to employ any locals.
reply▲And then spend the next five years building the actual fab around it? Like what’s the plan here.
reply▲No you're right. My math is very off.
reply▲The catch is if you started today with plenty of money (billions of dollars!) and could hire the right experts as you need them (this is a big if!) there would still be a couple years between today and producing 150 wafers per house. So the question isn't what does the math look like today, it is what the math looks like in 2 years - if you could answer that why didn't you start two years ago so you could get the current prices?
reply▲A photolithography machine doesn't etch anything (well, some EUV machines do it as an unwanted side effect because of plasma generation), it just patterns some resist material. Etching is happening elsewhere. Also, keep in mind, you'll need to do multiple passes through a photolithography machine to pattern different steps of the process - it's not a single pass thing.
reply▲yeaaahhh.. these fabs don't just have a button that says "punch out a top of the line nvidia GPU"
reply▲You're right. I'm throwing away the napkin.
reply▲that's 100% yield which ain't happening
reply▲What would you expect yield to be?
reply▲danparsonson3 hours ago
[-] With no prior experience? 0%. Those machines are not just like printers :-)
reply▲We'll have to gain some experience then :)
reply▲danparsonson2 hours ago
[-] Sure - once you have dozens of engineers and 5 years under your belt you'll be good to go!
This will get you started: https://youtu.be/B2482h_TNwg
Keep in mind that every wafer makes multiple trips around the fab, and on each trip it visits multiple machines. Broadly, one trip lays down one layer, and you may need 80-100 layers (although I guess DRAM will be fewer). Each layer must be aligned to nanometer precision with previous layers, otherwise the wafer is junk.
Then as others have said, once you finish the wafer, you still need to slice it, test the dies, and then package them.
Plus all the other stuff....
You'll need billions in investment, not millions - good luck!
reply▲Thank you. This is going to be way harder than I thought.
reply▲Especially when the plan is to just run them in a random rented commercial warehouse.
I drive by a large fab most days of the week. A few breweries I like are down the street from a few small boutique fabs. I got to play with some experimental fab equipment in college. These aren't just some quickly thrown together spaces in any random warehouse.
And it's also ignoring the water manufacturing process, and having the right supply chain to receive and handle these ultra clean discs without introducing lots of gunk into your space.
reply▲Sounds like the kind of question ChatGPT would be good at answering...
reply▲At that point, it'll be the opposite problem as more capacity than demand will be available. These new fabs won't be able to pay for themselves. Every tic receives a tok.
reply▲it's just a bunch of melted sand. How hard can it be?
reply▲UncleOxidant3 hours ago
[-] I think it would be more like 5-7 years from now if they started breaking ground on new fabs today.
reply▲China cannot buy ASML machines. All advanced semiconductor manufacturing in China is done with stockpiled ASML machines from before the ban.
reply▲That restriction is only for the most advanced systems. According to ASML's Q3 2025 filing, 42% of all system sales went to China.
SK Hynix also has significant memory manufacturing presence in China; or about 40% of the company's entire DRAM capacity.
reply▲Would you really need ASML machines to do DDR5 RAM? Honest question, but I figured there was competition for the non-bleeding edge - perhaps naively so.
reply▲Yes. You need 16nm or better for DDR5.
reply▲ASML cannot be avoided for 7nm and better due to EUV.
reply▲As someone who knows next to nothing about this space, why can China not build their own machines? Is ASML the only company making those machines? If so, why? Is it a matter of patents, or is the knowledge required for this so specialized only they've built it up?
reply▲They can - if they are willing to invest a a lot of money over several years. The US got Nuclear bombs in a few years during WWII with this thinking, and China (or anyone else) could too. This problem might be harder than a bomb, but the point remains, all it takes is a willingness to invest.
Of course the problem is we don't see what would be missed by doing this investment. If you put extra people into solving this problem that means less people curing cancer or whatever. (China has a lot of people, but not unlimited)
reply▲Yes. ASML is the only company making these machines. And both, they own thousands of patents and are also the only ones with the institutional knowledge required to build them anyway.
reply▲I thought Fermi paradox is about nukes; I increasingly think it's about chips
reply▲Designing such an advanced lithography machine is probably a Manhattan Project scale task. China is indeed trying to make progress and they will probably get there someday but for now ASML is the only company in the world that knows how to build these or anything remotely close.
reply▲GeekFortyTwo4 hours ago
[-] As someone with no skills in the space, no money, and lives near Ottawa: I'd love to help start a fab in Ontario.
reply▲Right on, partner. I think there's a ton of demand for it tbh
I'll do the engineering so we're good on that front. Just need investors.
reply▲I think I have a few dollars left on a Starbucks gift card. I'm in!
reply▲Hear me out- artisanal DRAM.
reply▲Every transistor carefully hand-wired with love, giving each bit its own individual characteristic, for more organic, natural data.
reply▲This existed and it was called LOL memory (little old lady memory). It was knitted by little old ladies, being careful of which wires passed through which cores of course. It was ROM, not RAM.
reply▲You've cracked the funding problem!
You need to sell the artisanal DRAM to HiFi hucksters. Get Monster Cable on board to say their cables can only reach peak warmth, punch, and clarity - when using gear with artisanal DRAM. You'll easily sell into that market at $10k per GB...
reply▲Locally sourced artisanal DRAM
reply▲I hope you have very deep pockets. But I'm cheering you on from the sidelines.
reply▲Just need a half billion in upfront investment. And thank you for the support :)
reply▲jeroenhd22 minutes ago
[-] I can't invest, your company doesn't have "AI" in the name and your business plan doesn't seem to involve shoving unwanted chatbots down everyone's throats at all. That's no way to run a modern business!
reply▲I'll happily invest $100 into this project in exchange for stock. let me know.
reply▲Kickstarter time? That would set some records.
reply▲So, playing the Mega Powerball are you?
reply▲reply▲I suppose it would help if I could read the whole page: I cannot see the left few characters on Firefox on Android. What did you make this with?
Note that fixing the site won't increase my chances of donating, I'm from the ASML country ;)
reply▲That's annoying... I made it with Next.js and Tailwind CSS tho. Hosted on Vercel.
reply▲wonder which one will find a winner faster...
reply▲that is an understatement
reply▲But what if it's a bubble driven by speculation?
It wouldn't pay off.
Starting a futures exchange on RAM chips, on the other hand...
reply▲Only if you put up the 10 billion dollars.
reply▲Machines are less than 400 million.
reply▲You're just talking about a lithography machine. Patterning is one step out of thousands in a modern process (albeit an important one). There's plenty more stuff needed for a production line, this isn't a 3D printer but for chips. And that's just for the FEOL stuff, then you still need to do BEOL :). And packaging. And testing (accelerated/environmental, too). And failure analysis. And...
Also, you know, there's a whole process you'll need to develop. So prepare to be not making money (but spending tons of it on running the lines) until you have a well tested PDK.
reply▲> 3D printer but for chips
how about a farm of electron microscopes? these should work
reply▲Canon has been working on an alternative to EUV lithography called nanoimprint lithography. It would be a bit closer to the idea of having an inkjet printer make the masks to etch the wafers. It hasn't been proven in scale and there's a lot of thinking this won't really be useful, but it's neat to see and maybe the detractors are wrong.
https://global.canon/en/technology/nil-2023.html
https://newsletter.semianalysis.com/p/nanoimprint-lithograph...
They'll still probably require a good bit of operator and designer knowledge to work around whatever rough edges exist in the technology to keep yields high, assuming it works. It's still not a "plug it in, feed it blank wafers, press PRINT, and out comes finished chips!" kind of machine some here seem to think exist.
reply▲And the cost of the people to run those machines, and the factories that are required to run the machines?
reply▲I'll do it for free. And I'm sure we could rent a facility.
reply▲I think you are vastly underestimating how a modern fab works. The average cost is 10 billion and takes a about 10 years to build. They are very complex and even the slightest issue can ruin yields. There is a reason why there are so few.
reply▲Yeah you're right. I was way off.
reply▲deuplonicus4 hours ago
[-] Sure, we can work on brining in TinyTapeout to modern fab
reply▲I wonder if these RAM shortages are going to cause the Steam Machine to be dead on arrival. Valve is probably not a big enough player to have secured production guarantees like Sony or Nintendo would have. If they try to launch with a price tag over $750, they're probably not going sell a lot.
reply▲Yeah, I think (sadly) this kills the Steam Machine in the short term if the competition is current consoles.
At least until the supply contracts Sony & Microsoft have signed come up for renewal, at which point they’re going to be getting the short end of the RAM stick too.
In the short term the RAM shortage is going to kill homebrew PC building & small PC builders stone dead - prebuilts from the larger suppliers will be able to outcompete them on price so much that it simply won’t make any sense to buy from anyone except HP, Dell etc etc. Again, this applies only until the supply contracts those big PC firms have signed run out, or possibly only until their suppliers find they can’t source DDR5 ram chips for love nor money, because the fabs are only making HBM chips & so they have to break the contracts themselves.
It’s going to get bumpy.
reply▲reply▲Eesh.
The fight over RAM supply is going to upend a lot of product markets. Just random happenstance over whether a company decided to lock in supply for a couple of years is going to make or break individual products.
reply▲jeroenhd21 minutes ago
[-] Valve has already commented informally that Steam Machines will be priced on par with gaming PCs of similar hardware specs.
They won't be able to defeat console makers on the short term, but PC gamers will be paying the same price, so for those the value proposition remains unchanged.
reply▲seanw44413 minutes ago
[-] Or Valve draws from their deep pockets to sell them at a loss in the short term to gain product loyalty in the long term.
reply▲If anyone that has hidden cash reserves that could buy out even Apple. It would probably be Valve.
Lol, wacky reality if they say "hey we had spare cash so we bought out Micron to get DDR5 for our gaming systems"
reply▲To save my fellow ignorami some math, Valve is estimated to be valued at approximately 3.2% of Micron, or 0.2% of Apple. :)
reply▲ActorNightly2 hours ago
[-] Value =/= cash reserves. Valve has ran VERY lean for quite some time.
reply▲Sure, but valuation should always exceed cash reserves. It's very odd when it does not. I think I recall that that SUNW (by then JAVA probably) was at one point valued below cash, prior to ORCL acquisition.
If Valve's secrecy is so good that they have (substantially more than) 30-500x cash stashed away in excess of their public valuation estimates, then perhaps I underestimate Valve's secrecy!
More likely, it was an obviously-humorous exaggeration, but I wasn't sure -- I am quite ignorant of the games industry. :)
reply▲quickthrowman1 hour ago
[-] Valve is worth
maybe 0.1% (single digit billions is what I’d guess) of Apple, which made $112B in
net income in 2025. That’s profit, not revenue.
Apple could probably buy Valve for 30 days of its net income, which is around $9.3B ($306M per day in profit, including weekends)
There’s zero chance that Gaben has squirreled away Four Trillion Dollars in cash.
reply▲maybe if he didn't buy all those yachts and submarines. alas.
reply▲> And those companies all realized they can make billions more dollars making RAM just for AI datacenter products, and neglect the rest of the market.
> So they're shutting down their consumer memory lines, and devoting all production to AI.
Okay this was the missing piece for me. I was wondering why AI demand, which should be mostly HBM, would have such an impact on DDR prices, which I’m quite sure are produced on separate lines. I’d appreciate a citation so I could read more.
reply▲Just like the GPUs.
NVIDIA started allocating most of the wafer capacity for 50k GPU chips. They are a business, its a logical choice.
reply▲It's kind of a weird framing. Of course RAM companies are going to sell their limited supply to the highest bidder!
reply▲Red chip supply problems in your factory are usually caused by insufficient plastic bars, which is usually caused by oil production backing up because you're not consuming your heavy oil and/or petroleum fast enough.
Crack heavy oil to light, and turn excess petroleum into solid fuel. As a further refinement, you can put these latter conversions behind pumps, and use the circuit network to only turn the pumps on when the tank storage of the respective reagent is higher than ~80%.
hth, glhf
reply▲reply▲Hadn't heard of it before..
> Foxhole is a massively multiplayer game where
nope, not my cup of tea, but thanks for the "if you like this you might like this" rec :)
reply▲Yup, and stockpiling solid fuel is not a waste because you need it for rocket fuel later on. Just add more chests.
reply▲Not a bad time for the secondary market to be created. We keep buying everything new, when the old stuff works just as well. There is a ton of e-waste. The enthusiast market can benefit, while the enterprise market can just eat the cost.
Also, a great incentive to start writing efficient software. Does Chrome really need 5GB to run a few tabs?
reply▲"dig into that pile of old projects you never finished instead of buying something new this year."
You don't need a new PC. Just use the old one.
reply▲I just bought some 30pin SIMMs to rehab an old computer. That market is fine.
reply▲I have a bag of SIMMs that I saved, no idea why, because I clearly wrote BAD on the mylar bag.
At time time I was messing around with the "badram" patch for Linux.
reply▲It's a lot harder to source ram with known errors than it is to get clean stuff, cause most of the stuff with known errors will be tossed. Makes sense to save it.
reply▲Can someone explain why OpenAI is buying DDR5 RAM specifically? I thought LLMs typically ran on GPUs with specialised VRAM, not on main system memory. Have they figured out how to scale using regular RAM?
reply▲They're not. They are buying wafers / production capacity to make HBM so there is less DDR5 supply.
reply▲OK, fair enough, but what are OpenAI doing buying production capacity rather than, say, paying NVIDIA to do it? OpenAI aren’t the ones making the hardware?
reply▲Just because Nvidia happily sells people discrete GPU's, DGX systems, etc., doesn't mean they would turn down a company like OpenAI paying them $$$ for just the packaged chips and the technical documentation to build their own PCBs; or, let OpenAI provide their own DRAM supply for production on an existing line.
If you have a potentially multi-billion dollar contract, most businesses will do things outside of their standard product offerings to take in that revenue.
reply▲> doesn't mean they would turn down a company like OpenAI paying them $$$ for just the packaged chips and the technical documentation to build their own PCBs
FWIW, this was the standard state of affairs of the GPU market for a long while. nVidia and AMD sold the chips they paid someone to produce to integrators like EVGA, PNY, MSI, ZOTAC, GIGABYTE, etc. Cards sold under the AMD or nVidia name directly were usually partnered with one of these companies to actually build the board, place the RAM, design the cooling, etc. From a big picture perspective, it's a pretty recent thing for nVidia to only really deliver finished boards.
On top of this, OpenAI/Sam Altman have been pretty open about making their own AI chips and what not. This might point to them getting closer to actually delivering on that (pure speculation) and wanting to ensure they have other needed supplies like RAM.
reply▲Because they can provide the materials to NVIDIA for production and prevent Google, Anthropic, etc from having them.
reply▲> OpenAI aren’t the ones making the hardware?
how surprised would you be if they announced that they are?
reply▲They didn't buy DDR5 - they bought raw wafer capacity and a ton of it at that.
reply▲I wonder if Apple will budge. The margins on their RAM upgrades were so ludicrous before that they're probably still RAM-profitable even without raising their prices, but do they want to give up those fat margins?
reply▲I know contract prices are not set in stone. But if there’s one company that probably has their contract prices set for some time in the future, that company is Apple, so I don’t think they will be giving up their margins anytime soon.
reply▲Night_Thastus4 hours ago
[-] RAM upgrades are such a minor, insignificant part of Apple's income - and play no part in plans for future expansion/stock growth.
They don't care. They'll pass the cost on to the consumers and not give it a second thought.
reply▲>
I wonder if Apple will budge.Perhaps I don't understand something so clarification would be helpful:
I was under the impression that Apple's RAM was on-die, and so baked in during chip manufacturing and not a 'stand alone' SKU that is grafted onto the die. So Apple does not go out to purchase third-party product, but rather self-makes it (via ASML) when the rest of the chip is made (CPU, GPU, I/O controller, etc).
Is this not the case?
reply▲mixermachine2 hours ago
[-] Sadly everything in the general direction of RAM or SSD chips is getting more expensive because a lot of production capacity is redistributed to serve AI chips and everything around.
Even lower end GPUs are getting more expensive even if they are not really useful for AI.
But they still contain <some> chips and ram which is in high demand.
So yes, Apple will likely also have to pay higher priceses when they renew their contracts.
reply▲I'd like to believe that their pricing for ram upgrades are like that so the base model can hit a low enough of a price. I don't believe they have the same margin for the base model compared to the base model + memory upgrade.
reply▲I read online that Apple uses three different RAM suppliers supposedly? I wonder if Apple has the ability to just make their own RAM?
reply▲Apple doesn't own any foundries, so no. It's not trivial to spin up a DRAM foundry either. I do wonder if we'll see TSMC enter the market though. Maybe under pressure from Apple or nvidia...
reply▲There are no large scale pure play DRAM fabs that I’m aware of, so Apple is (more or less) buying from the same 3 companies as everyone else.
reply▲Apple doesn't own semiconductor fabs. They're not capable of making their own RAM.
reply▲on one hand they are loosing profit, on the other hand they are gaining on market share. They will probably wait a short while to assess how much they are willing to sacrifice profits for market share
reply▲I am fully expecting a 20%+ price bump on new mac hardware next year.
reply▲Not me. It’s wildly unusual for Apple to raise their prices on basically anything… in fact I'm not sure if its ever happened. *
It’s been pointed out by others that price is part of Apple's marketing strategy. You can see that in the trash can Mac Pro, which logically should have gotten cheaper over the ridiculous six years it was on sale with near-unchanged specs. But the marketing message was, "we're selling a $3000 computer."
Those fat margins leave them with a nice buffer. Competing products will get more expensive; Apple's will sit still and look even better by comparison.
We are fortunate that Apple picked last year to make 16gb the new floor, though! And I don't think we're going to see base SSDs get any more generous for a very, very long time.
* okay I do remember that Macbook Airs could be had for $999 for a few years, that disappeared for a while, then came back
reply▲loloquwowndueo4 hours ago
[-] It’s 4D chess my dude, they were just training people to accept those super high ram prices. They saw this coming I tell you!
reply▲reply▲mixermachine1 hour ago
[-] Got to say, those prices were quite cheap.
I also upgraded my home server to 32GB RAM and paid something like 55€.
Now we can just wait for some bubble to pop...
reply▲>
And those companies all realized they can make billions more dollars making RAM just for AI datacenter products, and neglect the rest of the market.I wouldn't ascribe that much intent. More simply, datacenter builders have bought up the entire supply (and likely future production for some time), hence the supply shortfall.
This is a very simple supply-and-demand situation, nothing nefarious about it.
reply▲That makes it sound like they are powerless, which is not the case. They don’t
have to have their capacity fully bought out, they could choose to keep a proportion of capacity for maintaining the existing PC market, which they would do if they thought it would benefit them in the long term.
They’re not doing that, because it benefits them not to.
reply▲$20B, 5 years and you can have your own DDR5 fab to print money with.
jokes aside, if the AI demand actually materializes, somebody will look at the above calculation and say 'we're doing it in 12 months' with a completely straight face - incumbents' margin will be the upstart's opportunity.
reply▲stevenjgarner4 hours ago
[-] My understanding is that this is primarily hitting DDR5 RAM (or better). With prices so inflated, is there an argument to revert and downgrade systems to DDR4 RAM technology in many use cases (which is not so inflated)?
reply▲DDR 4 shot up too. It was bad enough that instead of trying to put together a system with the AM4 m/b I already have, I just bought a Legion Go S.
reply▲tencentshill4 hours ago
[-] It will be hit just as hard, they have stopped new DDR4 production to focus on DDR5 and HBM.
reply▲geerlingguy4 hours ago
[-] Linked in the article, DDR4 and LPDDR4 are also 2-4x more expensive now, forcing smaller manufactures to raise prices or cancel some products entirely.
reply▲DDR4 manufacturing is mostly shut down, so if any real demand picks starts there the prices will shoot up
reply▲No DDR4 is affected too. It's a simple question of production and demand, and the biggest memory manufacturers are all winding down their DDR4/DDR5 memory production for consumers (they still make some DDR5 for OEMS and servers).
reply▲DDR4 prices have gone up 4x in the last 3 months.
reply▲I think I paid like $500 ($1300 today) in 1989ish to upgrade from 2MB to 5MB of ram (had to remove 1MB to add 4MB)
reply▲I just gathered enough money to build my new PC. I'll even go to another country to pay less taxes, and this spike hit me hard. I'll buy anyway because I don't believe it will slow down so soon. But yeah, for me is a lot of money
reply▲Me too. I had saved up to make a small server. I guess I'll have to wait up until 2027–2028 at this rate.
reply▲Buy used gear and rip out the guts???
reply▲I wonder if we'll start to see instance type shortages or price increases from EC2 and GCP if they can't get enough DRAM for latest gen servers.
reply▲> maybe it's a good time to dig into that pile of old projects you never finished instead of buying something new this year.
Always good advice.
reply▲Called it! About a year ago (or more?) I thought nVidia was overpriced and if AI was coming to PCs RAM would be important and it might be good to invest in DRAM makers. As usual I didn't do anything with my insight, and here we are. Micron has more than doubled since summer.
reply▲The article suggests that because the power and cooling are customized, it would take a ton of effort to run the new AI servers in a home environment, but I'm skeptical of that. Home-level power and cooling are not difficult these days. I think when the next generation of AI hardware comes out (in 3-5 years), there will be a large supply of used AI hardware that we'll probably be able to repurpose. Maybe we'll sell them as parts. It won't be plug-and-play at first, but companies will spring up to figure it out.
If not, what would these AI companies do with the huge supply of hardware they're going to want to get rid of? I think a secondary market is sure to appear.
reply▲A single server is 20 KW. A rack is 200 KW.
These are not the old CPU servers of yesterday.
reply▲At minimum, you'd need to wire in new 240V circuits, and you could only run one or two of these servers before you'd need a service upgrade.
Then you'd have to deal with noise from a literal wall of fans, or build a separate high capacity water cooling system (and still deal with dumping that heat somewhere).
reply▲quickthrowman1 hour ago
[-] A utility is most likely only going to offer a 240V 400A single-phase service at best for a residence in the US, 320A can be used continuously. If you need more they’ll usually offer multiple 400A services.
I’ve heard stories about people convincing their utility to install three-phase service drops in their homes, but usually it’s not an option.
Anyways, 320A continuous load at 240V single-phase is 76.8kW, if you assume 25kW per server (20 kW for server, 5kW for cooling), you can run (3) servers and 15 tons of cooling and still have just enough left for one 120V 15A circuit to charge your phone and power a light.
reply▲anybody care to speculate on how long this is likely to last? is this a blip that will resolve itself in six months, or is this demand sustainable and we are talking years to build up new manufacturing facilities to meet demand?
reply▲Pure speculation, nobody can say say for sure, but my guess is 2-3 years.
reply▲The big 3 memory manufacturers (SK Hynix, Samsung, Micron) are essentially all moving upmarket. They have limited capacity and want to use it for high margin HBM for GPUs and ddr5 for servers. At the same time CXMT, Winbond and Nanya are stepping in at the lower end of the market.
I don't think there is a conspiracy or price fixing going on here. Demand for high profit margin memory is insatiable (at least until 2027 maybe beyond) and by the time extra capacity comes online and the memory crunch eases the minor memory players will have captured such a large part of the legacy/consumer market that it makes little sense for the big 3 to get involved anymore.
Add to that scars from overbuilding capacity during previous super memory super cycles and you end up with this perfect storm.
reply▲mixermachine1 hour ago
[-] Let's see if this demand truly holds.
I'm still unsure.
Currently nobody makes money with AI.
There will be a correction (if I can trust my magic crystal ball ;) )
reply▲If the OpenAI Hodling Company buys and warehouses 40% of global memory production or 900,000 memory wafers (i.e. not yet turned into DDR5/DDR6 DIMMs) per month at price X in October 2025, leading to supply shortages and tripling of price, they have the option of later un-holding the warehoused memory wafers for a profit.
https://news.ycombinator.com/item?id=46142100#46143535
Had Samsung known SK Hynix was about to commit a similar chunk of supply — or vice-versa — the pricing and terms would have likely been different. It’s entirely conceivable they wouldn’t have both agreed to supply such a substantial part of global supply if they had known more...but at the end of the day - OpenAI did succeed in keeping the circles tight, locking down the NDAs, and leveraging the fact that these companies assumed the other wasn’t giving up this much wafer volume simultaneously…in order to make a surgical strike on the global RAM supply chain..
What's the economic value per warehoused and insured cubic inch of 900,000 memory wafers? Grok response:
> As of late 2025, 900,000 finished 300 mm 3D NAND memory wafers (typical high-volume inventory for a major memory maker) are worth roughly $9 billion and occupy about 104–105 million cubic inches when properly warehoused in FOUPs.
→ Economic value ≈ $85–90 per warehoused cubic inch.
reply▲reply▲> To save the situation, a consortium of US banks provided a $1.1 billion line of credit to the brothers which allowed them to pay Bache which, in turn, survived the ordeal.
It seems once you amass a certain amount of wealth, you just get automatically bailed out from your mistakes
reply▲I know this is mostly paranoid thinking on my behalf, but it almost feels like this is a conscious effort to attempt to destroy "personal" computing.
I've been a huge advocate for local, open, generative AI as the best resistance to massive take-over by large corporations controlling all of this content creation. But even as it is (or "was" I should say), running decent models at home is prohibitively expensive for most people.
Micron has already decided to just eliminate the Crucial brand (as mentioned in the post). It feels like if this continues, once our nice home PCs start to break, we won't be able to repair them.
The extreme version of this is that even dumb terminals (which still require some ram) will be as expensive as laptops today. In this world, our entire computing experience is connecting a dumb terminal to a ChatGPT interface where the only way we can interact with anything is through "agents" and prompts.
In this world, OpenAI is not overvalued, and there is no bubble because the large LLM companies become computing.
But again, I think this is mostly a dystopian sci-fi fiction... but it does sit a bit too close to the realm of possible for my tastes.
reply▲clusterhacks3 hours ago
[-] I share your paranoia.
My kids use personal computing devices for school, but their primary platform (just like their friends) is locked-down phones. Combining that usage pattern with business incentives to lock users into walled gardens, I kind of worry we are backing into the destruction of personal computing.
reply▲Wouldn't the easy answer to this be increased efficiency of RAM usage?
RAM being plentiful and cheap led to a lot of software development being very RAM-unaware, allowing the inefficiencies of programs to be mostly obfuscated from the user.
If RAM prices continue rising, the semi-apocalytic consumer fiction you've spun here would require that developers not change their behaviors when it comes to software they write.
There will be an equillibrium in the market that still allows the entry of consumer PC's it will just mean devices people buy will have less available RAM than is typical. The demand will eventually match up to the change in supply as is typical of supply/demand issues and not continuously rise into an infinite horizon.
reply▲171862744042 minutes ago
[-] That "dumb terminals" still need to run a modern web browser (likely Chrome) on a modern OS (likely Windows), these aren't exactly efficient with the available computing sources, so you could give up a lot of resources before you would actually trade it of for computing ability. Also resources have been risen exponentially for the last decades.
reply▲I believe that while centralized computing excels at specific tasks like consumer storage, it cannot
compete with the unmatched diversity and unique intrinsic benefits of personal computing. Kindle cannot replace all e-readers. Even Apple’s closed ecosystem cannot permit it to replace macOS with iPadOS. These are not preferences but constraints of reality.
The goal shouldn’t be to eliminate one side or the other, but to bridge the gap separating them. Let vscode.dev handle the most common cases, but preserve vscode.exe for the uncommon yet critical ones.
reply▲The first "proper" "modern" computer I had, initially came with 8 megabytes of RAM.
It's not a lot, but it's enough for a dumb terminal.
reply▲Narishma50 minutes ago
[-] A dumb terminal doesn't even need a tenth of that.
reply▲That's not disproving OP's comment; OpenAI is, in my opinion, making it untenable for a regular Joe to build a PC capable of running local LLM model. It's an attack on all our wallets.
reply▲Why do you need a LLM running locally so much that's the inflated RAM prices are an attack on your wallet? One can always opt not to play this losing game.
I remember when the crypto miners rented a plane to deliver their precious GPUs.
reply▲I don’t think you need a conspiracy theory to explain this. This is simply capitalism, a system that seems less and less like the way forward. I’m not against markets, but I believe most countries need more regulations targeted at the biggest companies and richest people. We need stronger welfare states, smaller income gaps and more democracy. But most countries seems to vote in the absolute opposite direction.
reply▲BizarroLand4 hours ago
[-] The end goal of capitalism is the same as the end goal of monopoly.
1 person has all the money and all the power and everyone else is bankrupt forever and sad.
reply▲It’s not a conspiracy, it’s just typical dumb short term business decisions amplified and enabled by a cartel supply market.
If Crucial screws up by closing their consumer business they won’t feel any pain from it because the idea of new competitors entering the space is basically impossible.
reply▲I can't help but be the pessimist angle. RAM production will need to increase to supply AI data centers. When the AI bubble bursts (and I do believe it will), the whole computing supply chain, which has been built around it, will take a huge hit too. Excess production capacity.
Wonder what would happen if it really takes a dive. The impact on the SF tech scene will be brutal. Maybe I'll go escape on a sailboat for 3 years or something.
Anyway, tangential, but something I think about occasionally.
reply▲Prices are high because no one believes it's not a bubble. Nvidias strategy has been careful careful with volume this whole time as well.
The thing is it's also not a conventional looking bubble: what we're seeing here is cashed up companies ploughing money into the only thing in their core business they could find to do so with, rather then a lot of over exuberant public trading and debt financing.
reply▲Bullwhip effect on this will be funny. At least, we are in for some cheap ram in like… a dozen months or so.
reply▲Forgeties793 hours ago
[-] I’m really hoping this doesn’t become another GPU situation where every year we think it’s going to get lower and it just stays the same or gets worse
reply▲I wonder how much of that RAM is sitting in GPUs in warehouses waiting for datacenters to be built or powered?
reply▲internet20003 hours ago
[-] Hopefully Apple uses their volume as leverage to avoid getting affected by this for as long as possible. I can ride it out if they manage to.
reply▲mikelitoris4 hours ago
[-] Companies are adamant about RAMming AI down our throats, it seems.
reply▲Is this a shortage of every type of RAM simultaneously?
reply▲Every type of DRAM is ultimately made at the same fabs, so if one type is suddenly in high demand then the supply of everything else is going to suffer.
reply▲Wait, really? For CPUs each generation needs basically a whole new fab, I thought… are they more able to incrementally upgrade RAM fabs somehow?
reply▲mixermachine1 hour ago
[-] Not quickly but if somebody puts enough money on the table, the fabs change too.
All about cost and return.
Micron just axed their brand crucial (end customer RAM and SSD) because they will only sell to database centers from now on.
Crazy times.
reply▲selectodude4 hours ago
[-] The old equipment is mothballed because china is the only buyer and nobody wants to do anything that the Trump admin will at some point decide is tariff-worthy. So it all sits.
reply▲internetter4 hours ago
[-] Essentially yes, not necessarily equivalently but every type has increased substantially
reply▲Ha! Maybe Javascript developers will finally drop memory usage! You need to display the multiplication table? Please allocate 1GB of RAM. Oh, you want alternate row coloring? Here is another 100MB of CSS to do that.
edit: this is a joke
reply▲phantasmish4 hours ago
[-] I do sometimes reflect on how 64MB of memory was enough to browse the Web with two or three tabs open, and (if running BeOS) even play MP3s at the same time with no stutters. 128MB felt luxurious at that time, it was like having no (memory-imposed) limits on personal computing tasks at all.
Now you can't even fit a browser doing nothing into that memory...
reply▲HN works under Dillo and you don't needs JS at all. If some site needs JS, don't waste your time. Use mpv+yt-dlp where possible.
reply▲Haha, the amount of downvotes of your very true comment just proves how many web developers are there on HN.
reply▲Narishma45 minutes ago
[-] One even responded to an earlier comment of mine that we shouldn't be optimizing software for low-end machines because that would be going against technological progress...
reply▲Its funny because the blogpost author makes the same joke
reply▲>
But I've already put off some projects I was gonna do for 2026, and I'm sure I'm not the only one.Let's be honest here - the projects I'm going to do in 2026, I bought the parts for those back in 2024. But this is definitely going to make me put off some projects that I might have finally gotten around to in 2028.
reply▲Also I feel that when this bubble bursts, we'll have much bigger problems than some expensive ram. More like the big 2007 crisis.
reply▲So, buy RAM now or gold? ;)
reply▲time to stop using python boys, it's zig from here on out
reply▲panem et circenses
But what will happen when people are priced out from the circus?
reply▲I am very excited for a few years when the bubble bursts and all this hardware is on the market for cheap like back in the early to mid 2000's after that bubble burst and you had tons of old servers available for homelabs. I can't wait to fill a room with 50kW of bulk GPUs on a pallet and run some cool shit.
reply▲1. Buy up 40% of global fab capacity
2. Resell wafers at huge markup to competitors
3. Profit
reply▲Disingenuous title and no mention of those directly responsible, Sam Altman/OpenAI. Once again Jeff is desperate for clicks
reply▲Only a matter of time before supply catches up and then likely overshoots (maybe combined with AI / datacenter bubble popping), and RAM becomes dirt cheap. Sucks for those who need it now though.
reply▲If openai was building out with DDR5 that would create a big ol supply of DDR5 to make RAM cheap.
No. They bought the wafers for HBM. The only thing thats going to get cheap when openai implodes are server stuff for homelab bros. The RAM manufacturers are using their production capacity to produce HBM for openai.
reply▲I grabbed a framework desktop with 128GB due to this. I can't imagine they can keep the price down for the next batches. If you bought 128GB of ram with "close" specs to the one used just that would be 1200 EUR at retail (who are obviously taking advantage).
reply▲Every shortage is followed by a glut. Wait and see for RAM prices to go way down. This will happen because RAM makers are racing to produce units to reap profits from the higher price. That overproduction will cause prices to crash.
reply▲They aren't overproducing consumer modules, they're actively cutting production of those. They're producing datacenter/AI specific form factors that won't be compatible with consumer hardware.
reply▲somebody will step up to pick up the free money if this continues.
reply▲32GB should be more than enough.
You can go 16GB if you go native and throw some assembly in the mix. Use old school scripting languages. Debloat browsers.
It has been long delayed.
reply▲Night_Thastus4 hours ago
[-] 16GB is more than fine if you're not doing high-end gaming, or heavy production workloads. No need for debloating.
But it doesn't matter either way, because both 16 and 32GB have what, doubled, tripled? It's nuts. Even if you say "just buy less memory", now is a horrible time to be building a system.
reply▲I found web browser tabs eating too much memory when you only have 16GB
reply▲171862744038 minutes ago
[-] With that much RAM, you could fit multiple OSes entirely into memory. That is simply a lot.
reply▲prmoustache3 hours ago
[-] Use adblock, stop visiting nasty websites and open less tabs. Problem is easily solved.
reply▲Night_Thastus4 hours ago
[-] That's strange. How many tabs are we talking? Are they running something intense or just ordinary webpages?
reply▲felixfurtak4 hours ago
[-] “640K ought to be enough for anybody.” - Bill Gates
reply▲there's no reliable evidence he ever uttered that phrase
reply▲franciscojs4 hours ago
[-] I bought a motherboard to build a DIY NAS... takes DDR5 SO-DIMM RAM and only 16gb costs more than double the motherboard (which includes an intel processor)
reply▲16GB is more than enough on Linux, but Win11 eats resources like crazy
reply▲Forgeties793 hours ago
[-] Sure but 32GB DDR5 ram has just jumped from ~$100 to $300+ in a flash. The 2x16GB I have in my recent build went from $105 for the pair to $250
each. $500 total!
SSD’s are also up. Hell I am seeing refurbished enterprise HDD’s at 2x right now. It’s sharp increases basically across the board except for CPU’s/GPU’s.
Every PC build basically just cranked up $400-$600 easily, and that’s not accounting for the impact of inflation over the last few years weakening everyone’s wallets. The $1600 machine I spec’d out for my buddy 5 weeks ago to buy parts for this Black Friday now runs $2k even.
reply▲I'm using 1GB with TWM, DIllo, TUI tools, XTerm, MuPDF and the like. As most tools are small
from
https://t3x.org,
https://luxferre.top and
https://howerj.github.io/subleq.htm
with EForth, (and I try to use cparser instead of) clang, my requeriments are really tiny.
You can achieve a lot by learning Klong and reading the intro on statistics. And xargs to paralelize stuff. Oh, and vidir to edit directories at crazy speeds with any editor, even nano or gedit if you like them.
reply▲Very happy to have bought a MBP M4 with 64 gb of Ram last year.
reply▲And sharing the RAM between your CPU/GPU/NPU instead of using separate memories.
reply▲citizenpaul4 hours ago
[-] > Micron's killing the Crucial brand of RAM and storage devices completely,
More rot economy. Customers are such a drag. Lets just sell to other companies for billion dollar deals at once. These AI companies have bottomless wallets. No one has thought of this before we will totally get rich.
reply▲altmanaltman4 hours ago
[-] I mean what's the big deal, can't we just download more ram
reply▲mschuster914 hours ago
[-] > The reason for all this, of course, is AI datacenter buildouts. I have no clue if there's any price fixing going on like there was a few decades ago—that's something conspiracy theorists can debate—but the problem is there's only a few companies producing all the world's memory supplies.
So it's the Bitcoin craze all over again. Sigh. The bubble will eventually collapse, it has to - but the markets can stay irrational longer than you can stay solvent... or, to use a more appropriate comparison, have a working computer.
I for myself? I hope once this bubble collapses, we see actual punishments again. Too-large-to-fail companies broken up, people getting prosecuted for the wash trading masquerading itself as "legitimate investments" in the entire bubble (that more looks like the genetic family table of the infamously incestuous Habsburg family), greedy executives jailed or, at least where national security is impacted due to chip shortages, permanently gotten rid of. I'm sick and tired of large companies being able to just get away with gobbling up everything, killing off the economy at large, they are not just parasites - they are a cancer, killing its host society.
reply▲Pricing compute out of the average persons budget to prop up investment in data centers, stocks, ultimately control agency
If an RTX 5000 series price topped out at historical prices no one would need hosted AI
Then it came to be that models were on a path to run well enough loaded into RAM... uh oh
This is in line with ISPs long ago banning running personal services and the long held desire to sell dumb thin clients that must work with a central service
Web developers fell for confidence games of old elders hook line and sinker. Nothing but the insane ego and vanity of some tech oligarchs driving this. They cannot appear weak. Vain aura farming, projection of strength.
reply▲reply▲benlivengood4 hours ago
[-] If your argument is that value produced per-cpu will increase so significantly that the value produced by AGI/ASI per unit cost exceeds what humans can produce for their upkeep in food and shelter, then yes that seems to be one of the significant risks long term if governments don't intervene.
If the argument is that prices will skyrocket simply because of long-term AI demand, I think that ignores the fact that manufacturing vastly more products will stabilize prices up to the point that raw materials start to become significantly more expensive, and is strongly incentivized over the ~10-year timeframe for IC manufacturers.
reply▲>the value produced by AGI/ASI per unit cost exceeds what humans can produce for their upkeep in food and shelter
The value of AGI/ASI is not only defined by its practical use,
It is also bounded by the purchasing power of potential consumers.
If humans aren’t worth paying, those humans won’t be paying anyone either. No business can function without customers, no matter how good the product.
reply▲benlivengood3 hours ago
[-] Precisely the place where government intervention is required to distribute wealth equitably.
reply▲captainkrtek4 hours ago
[-] I'm no economist, but if (when?) the AI bubble bursts and demand collapses at the price point memory and other related components are at, wouldn't price recover?
not trying to argue, just curious.
reply▲I'm no economist either, but I imagine the manufacturing processes for the two types of RAM are too different for supply to quickly bounce back.
reply▲IF a theoretical AI bubble bursts sure. However the largest capitalized companies in the world and all the smartest people able to do cutting edge AI research are betting otherwise. This is also what the start of a takeoff looks like
reply▲Why should we believe in another apocalypse prediction?
reply▲One of them has to be right, eventually!
reply▲Because the collapse of complex societies is real -
https://github.com/danielmkarlsson/library/blob/master/Josep...Unbounded increases in complexity lead to diminishing returns on energy investment and increased system fragility which both contribute to an increased likelihood of collapse as solutions to old problems generate new problems faster than new solutions can be created since energy that should be dedicated to new solutions is needed to maintain the layers of complexity generated by the layers of previous solutions.
reply