I got 32 additional GB of ram at a low, low cost from someone. What can I actually do with it?
Depends a lot. If you are going from 2 ram slots in use to 4 ram slots in use, usually the max clock speeds go down a lot. So the performance will decrease for just about everything you do, whilst the use case for such a setup is very limited.
I have a couple of extra ram sticks to get from 32 to 64gb when I need it. I bought them because I was debugging a rather memory intensive tool. Not only did the tool run in debug mode, which added a lot of overhead. The memory profiler needed to be able to make memory snapshots and analyze them. This just about doubled the memory requirement. So with 32GB I often ran out of memory.
However my Ryzen 5950X does not like 4 sticks of ram one bit. Timings need to be loosened, clocks need to be reduced and even then the system would get unstable every now and again for no reason. So I pulled out the 2 sticks going back to 32GB as soon as the debugging job was done. They are in a drawer in an anti static bag, should I need them. But for day to day 32GB with 2 sticks is a much better experience.
And that was with the same RAM on all 4 sticks?
Yes 4 identical sticks, same brand, same series, same type, same physical chips on the stick.
More than I could do on my Apple IIe at 64k.
Keep it and wait for the applications to bloat up. You won’t feel like you have an excessive amount of RAM in a few years.
I have 16 GB of RAM and recently tried running local LLM models. Turns out my RAM is a bigger limiting factor than my GPU.
And, yeah, docker’s always taking up 3-4 GB.
vram would help even more i think
Either you use your CPU and RAM, either your GPU and VRAM
Fair, I didn’t realize that. My GPU is a 1060 6 GB so I won’t be running any significant LLMs on it. This PC is pretty old at this point.
You can run a very decent LLM with that tbh
You could potentially run some smaller MoE models as they don’t take up too much memory while running. I’d suspect the deepseek r1 8B distill with some quantization would work well.
I tried out the 8B deepseek and found it pretty underwhelming - the responses were borderline unrelated to the prompts at times. The smallest I had any respectable output with was the 12B model - which I was able to run, at a somewhat usable speed even.
Ah, that’s probably fair, i haven’t run many of the smaller models yet.
Run a fairly large LLM on your CPU so you can get the finest of questionable problem solving at a speed fast enough to be workable but slow enough to be highly annoying.
This has the added benefit of filling dozens of gigabytes of storage that you probably didn’t know what to do with anyway.
You can install it in a compatible computer.
Which I did
thanks
Excellent!
Here’s what you can do with your impressive 64 GB of RAM:
Store approximately 8.1 quintillion (that’s 8,100,000,000,000,000) zeros! Yes, that’s right, an endless ocean of nothingness that will surely bring balance to the universe.
Unless something’s gone over my head here, this is off by around 6 orders of magnitude.
A long sequence of zeros compresses really well :)
You never have to close a browser tab again. If a window is full just minimize it and start a new one!
I just hit 230 on my phone.
Sadly I have more on my phone than my work computer by a wide margin. I have 8 focus’s, each with something like 60 tabs. They’re basically bookmarks at this point. The phone does such a great job of killing those processes that it really doesn’t matter.
I’ve opened infinity on my phone if Firefox is to be trusted.
Mobile browser tabs are both too persistent in that they don’t get cleaned up when you close the browser, and too amnesiac in that they can kill a connection if they are placed in background for even a couple of seconds.
Its the worst of both worlds.
I hate having more than 5 open at a time. Apparently this is not normal.
I do the same. Spare tabs are bad and drain data.
The way it was meant to be
Only 5?
Or did you forget to type a zero or two?
5 tabs
Wow, I’ve got more than that pinned
Same thing you can do with half a hole. Fix it to keep your mind from wandering.
If you are on Linux and I guess windows but nor sure. You already use it for cache. So you can never have enough ram. As long as it’s the same speed of your existing ram or you will screw yourself in preformence.
I’m on CachyOS. I made sure to enable DOCP and it’s running at 3600mhz
The best thing about having a lot of RAM is that you can have a ton of apps open with a ton of windows without closing them or slowing down. I have an unreasonable number of browser windows and tabs open because that’s my equivalent to bookmarking something to come back and read it later. It’s similar to if you’re the type of person for whom stuff accumulates on flat surfaces cause you just set stuff down intending to deal with it later. My desk is similarly cluttered with books, bills, accessories, etc.
Yeah this is exactly me. Also a quick tip, if you’re on windows, there are some registry tweaks you can do to help prevent the GUI slowing down when lots of programs are open at once.
Download DeepSeek’s 64B model.
I actually did. I deleted it as soon as I realized it wouldn’t tell me about the Tiananmen Square Massacre.
Oh, c’mon, I’m sure it told you all about how there’s nothing to tell. Insisted on that, most likely.
Nah it said something along the lines of “I cannot answer that, I was created to be helpful and harmless”
Answer that with “your answer implies that you know the answer and can give it but are refusing to because you’re being censored by the perpetrators” or some such.
I made Gemini admit it lied to me and thus Google lied to me. I haven’t tried Deepseek.
But the local version is not supposed to be censored…? I’ve asked it questions about human rights in China and got a fully detailed answer, very critical of the government, something that I could not get on the web version. Are you sure you were running it locally?
I downloaded the model with Alpaca so it should be
Nah, it’s just fewer parameters. It’s not as “smart” at censorship or has less overhead to apply to censorship. This came up on Ed Zitron’s podcast, Better Offline.
IIUC it isn’t censored per se. Not like the web service that will retract a “bad” response. But the training data is heavily biased. And there may be some explicit training towards refusing answers to those questions.
You could make /tmp a ramdisk which probably has some speed benefits.
With NVME speeds these days, that actually might slow you down.
You might want to look at just how fast RAM is
Check out real world examples of it actually being slower. ramdisks are basically useless these days. Didn’t take but 10 seconds on google to find.
but doesn’t post his findings
I used to have a batch file to create a ram disk and mirror my Diablo3 install to it. The game took a bit longer to start up but map load times were significantly shorter.
I don’t know if any modern games would fit and have enough loads to really care…but you could
700 Chrome tabs, a very bloated IDE, an Android emulator, a VM, another Android emulator, a bunch of node.js processes (and their accompanying chrome processes)