

+1 for immutable in general.
+1 for immutable in general.
Yeah, that’s problematic, heh.
To be fair, I do wish more privacy-friendly browsers took DDG mobile’s approach, namely torch all sites but make it really easy to (and prompt you to) whitelist frequently used ones.
You can permanently disable the chatbot in full DDG search. Click the little gear.
It does make me wonder what API they use. I thought it was huggingface (which would be less bad), but they don’t say it explicitly.
Yeah. But it also messes stuff up from the llama.cpp baseline, and hides or doesn’t support some features/optimizations, and definitely doesn’t support the more efficient iq_k quants of ik_llama.cpp and its specialzied MoE offloading.
And that’s not even getting into the various controversies around ollama (like broken GGUFs or indications they’re going closed source in some form).
…It just depends on how much performance you want to squeeze out, and how much time you want to spend on the endeavor. Small LLMs are kinda marginal though, so IMO its important if you really want to try; otherwise one is probably better off spending a few bucks on an API that doesn’t log requests.
In case I miss your reply, assuming a 3080 + 64 GB of RAM, you want the IQ4_KSS (or IQ3_KS, for more RAM for tabs and stuff) version of this:
https://huggingface.co/ubergarm/GLM-4.5-Air-GGUF
Part of it will run on your GPU, part will live in system RAM, but ik_llama.cpp does the quantizations split and GPU offloading in a particularly efficient way for these kind of ‘MoE’ models. Follow the instructions on that page.
If you ‘only’ have 32GB RAM or less, that’s tricker, and the next question is what kind of speeds do you want. But it’s probably best to wait a few days and see how Qwen3 80B looks when it comes out. Or just go with the IQ4_K version of this: https://huggingface.co/ubergarm/Qwen3-30B-A3B-Thinking-2507-GGUF
And you don’t strickly need the hyper optimization of ik_llama.cpp for a small model like Qwen3 30B. Something easier like lm studio or the llama.cpp docker image would be fine.
Alternatively, you could try to squeeze Gemma 27B into that 11GB VRAM, but it would be tight.
How much system RAM, and what kind? DDR5?
ik doesn’t have great documentation, so it’d be a lot easier for me to just point you places, heh.
At risk of getting more technical, ik_llama.cpp has a good built in webui:
https://github.com/ikawrakow/ik_llama.cpp/
Getting more technical, its also way better than ollama. You can run models way smarter than ollama can on the same hardware.
For reference, I’m running GLM-4 (667 GB of raw weights) on a single RTX 3090/Ryzen gaming rig, at reading speed, with pretty low quantization distortion.
And if you want a ‘look this up on the internet for me’ assistant (which you need for them to be truly useful), you need another docker project as well.
…That’s just how LLM self hosting is now. It’s simply too hardware intense and ad hoc to be easy and smart and cheap. You can indeed host a small ‘default’ LLM without much tinkering, but its going to be pretty dumb, and pretty slow on ollama defaults.
Yeah, see, that makes sense. A random app and an optional account number are not reliable notification systems. They can’t just assume everyone will opt into those.
I’m sick of finger wagging over what other nations are doing juxtaposed with shameless fawning over the righteousness their own governments.
I’m sick of obsessions over how awful NATO is for their involvement in Ukraine, from people who turn around, and in the same breath, praise the virtues of yet another post WWII ‘liberation war’, like the Ukraine War will somehow not end in disaster like every. Single. One. Can you even name a post WWII intervention that ended well?
But noooo, those last hundred interventions were totally different than this one! Yeah, that lines up with historical precedent, sure.
You have no right to finger wag about westerners living in a bubble. Or intervening in another country.
And no. I am not a Nazi.
So Lemmy.ml can either get an ounce of self humility, not a lot, just a little. Or they can take their downvote barrage.
through phone if you have a phone on your water account, through a system no one knew existed
I interpreted this as one system. So its:
Water website, you’d have to happen to stumble upon
Obscure opt-in phone system
If that’s the case, the complaint is reasonable, as the water service is basically assuming Facebook (and word of mouth) are the only active notifications folks need.
But yeah, if OP opted out of SMS warnings or something, that’s more on them.
I’m sorry, but what’s the track record for ‘foreign intervention to fight imperialism’ after WWII?
I’ll tell you: disaster. Every, single, one, western or not.
I find this:
Russia invades the nazi puppet government of the US empire to prevent further empire NATO imperialism and reestablish the Ukrainian peoples proper government
So ironic it goes from being painful to hilarious to painful again. Somehow, this one liberation war, among all in the last 100 years, will be the justified and successful one. Like how Americans were so happy being stuck in what was supposed to be a quick ‘liberation war’ in Vietnam. To quote Johnson:
Why are these realities our concern? Why are we in South Vietnam? We are there because we have a promise to keep. Since 1954 every American president has offered support to the people of South Vietnam. We have helped to build, and we have helped to defend. Thus, over many years, we have made a national pledge to help South Vietnam defend its independence. And I intend to keep that promise…
…But yeah. Oh, this one is the exception if we keep going, sure. Nods head and pats back. We’ll talk again in a few years…
I think Memes@lemmy.ml doesn’t understand why so many of their posts get mass downvoted, like pointing out this revelation to other instances…
We. Freaking. KNOW!
Western fediverse instances know there was a freakin Nazi battalion in Ukraine, and rampant corruption, and other problems. We know ‘the west’ has been, and is, imperialist, hypocritical, genocidal and transactionally shit to other countries. This isn’t /r/conservative with bots preaching on about the virtues of western patriotism in 2025, this is a refuge for deep internet dwellers either watching their countries burn down around them (like me in the US), or, according to polls, barrel headfirst into the same techno-fascist utopias (largely Europe right now).
Memes like this are pretentious. It’s insulting, and most of all, insensitive.
How ridiculous would it look if Americans repeatedly pointed out how awful a trading partner China is in the geopolitical space compared to the US? Or if we constantly filled Lemmy with memes about how inclusive and virtuous and definitely-not-racist the American megacorps are compared to, say, North Korea? What if we set Ukraine aside for a sec and started memeing on the crises in Russia’s backyard, while utterly refusing to even acknowledge the dumpster fires in our own?
…So, yeah.
Even setting a ton of stuff aside, you’re preaching to the choir, in a really unfun way.
This meme is also wrong, at least on the top half, because apparently the majority of my country thinks rebadged, redneck/cyberpunk flavored neo-nazism is the way forward :(
I guess it’d be nice if I had a victim complex big enough to obsess over other countries instead of my own, but I’m sorry, I can barely think over Fox News raging about how white Christians and Donald Trump are victims in the background.
I’m a massive fan of CachyOS, personally! Installed it years ago, kept the same image since then and haven’t even considered switching.
Different philosphies, I suppose. I suspect Bazzite may work better if you want stuff to just work, while Cachy is more tweaking focused and gets quite rapid updates, though is still quite set up out-of-the-box.
Yeah. Distros are basically just preconfigured sets of Linux, with the communities focusing on what they are interested in.
For gaming? You need a distro that does stuff for you!
To elaborate, if you’re using wine bottles, you’ve gone waaay into the land of manual from-scratch configuration, when you should just use stuff from a community that spends thousands of man hours figuring it out and packaging it.
Try CachyOS or Bazzite! They have a bunch of packages like advanced versions of preconfigured Proton one install away.
For docker… yeah, it’s a crazy learning curve if you just want to try one small thing. It’s honestly annoying to go through all the setup and download like 100 gigabytes of files just to run a python script or whatever.
You can often set up the environment yourself without docker, though.
And to reiterate, I’m very much against the ethos of “you should learn how to do everything yourself!” I get the sentiment, but honestly, this results in suboptimal configurations for most people vs simply using the packages others have spent thousands of hours refining.
Yep.
FYI, rumors suggest the AI Max/Strix Halo successor won’t be coming out till H2 2027, aka nearly 2028 (as Strix Halo techically launched in January this year, but as you can see takes time to actually make it into laptops):
Anyway, what I’m saying is it won’t go obsolete anytime soon, and it will be quite strong for many years to come if you get one.
There is a 14" HP laptop with the same chip:
https://www.ultrabookreview.com/70442-amd-strix-halo-laptops/
And a handheld, heh: https://gpdstore.net/gpd-handheld-gaming-pcs/gpd-win-5/
There may be more.
TBH, it may be prudent to wait a month or two for more “AI Max” chips to show up in laptops. It’s pretty new; Asus is just super early with it like they usually are.
Oh, and one more thing. There’s a sizable linux community specifically built around Asus ROG laptops. Look up ‘linux rog’ and you will find associated gitlabs and a Discord specifically built up around them. It’s still a fantastic resource for my 2020 G14.
The Z13 is especially good for linux, as it has discrete-gpu-class performance on the IGP, so you don’t have to fuss with a dual GPU setup on linux (which can be a tremendous headache, especially with Nvidia cards).
As for a distro, I adore CachyOS for ML stuff, and its well suited for gaming. But its really down to your personal experience and taste.
Ah, crap, you’re in Europe.
So basically the only laptop worth anything for AI is one with the new Strix Halo AMD chips, and the closest to what you want is the Asus Z13: https://www.notebookcheck.net/Asus-ROG-Flow-Z13-GZ302EA-Convertible-Review-AMD-s-Strix-Halo-GPU-is-neck-and-neck-with-the-RTX-4070-Laptop.963266.0.html
https://shop.asus.com/us/rog/90nr0jy1-m00670-rog-flow-z13-2025.html?config=90NR0JY1-M00670
Specifically the 128GB version if you can save up, or at least the 64GB version. While most laptops are useless for ML, this one utterly blows my desktop out of the water: it’s like an of magnitude better than the Frameowrk 13 at that.
Even more importantly, LLM devs are targeting the Strix Halo chips, so they will be well supported. You can spin up a vllm, exllama or llama.cpp-rocm image on them right now, whereas you will struggle to get things up and running on most laptops older IGPs.
Coincidentally, you won’t find anything 13" that can game better either. Its a surface-like tablet too, and franky its cooling is way better than a Framework 13. It’s perfect!
…Problem is, I don’t know if you can even get it in Europe. But historically, I know Asus laptops tend to be proportionally more expensive than they are in the US for some reason, so even if you can, I’m afraid the 64GB/128GB versions would be cost prohibitive.
Yeah.
The ‘dreamer’ part of me pictures this as enabling solo devs with masterpieces in their heads to finally make the game they want. Or small studios to undermine AAAs even more.
…But grifters gonna grift.
And apparently people buying slop can’t help themselves, whether the slop is AI or not.