Once configured, Tor Hidden Services also just work (you may need to use some fresh bridges in certain countries if ISPs block Tor there though). You don’t have to trust any specific third party in this case.
Once configured, Tor Hidden Services also just work (you may need to use some fresh bridges in certain countries if ISPs block Tor there though). You don’t have to trust any specific third party in this case.
If your CPU isn’t ancient, it’s mostly about memory speed. VRAM is very fast, DDR5 RAM is reasonably fast, swap is slow even on a modern SSD.
8x7B is mixtral, yeah.
Mostly via terminal, yeah. It’s convenient when you’re used to it - I am.
Let’s see, my inference speed now is:
As of quality, I try to avoid quantisation below Q5 or at least Q4. I also don’t see any point in using Q8/f16/f32 - the difference with Q6 is minimal. Other than that, it really depends on the model - for instance, llama-3 8B is smarter than many older 30B+ models.
Have been using llama.cpp, whisper.cpp, Stable Diffusion for a long while (most often the first one). My “hub” is a collection of bash scripts and a ssh server running.
I typically use LLMs for translation, interactive technical troubleshooting, advice on obscure topics, sometimes coding, sometimes mathematics (though local models are mostly terrible for this), sometimes just talking. Also music generation with ChatMusician.
I use the hardware I already have - a 16GB AMD card (using ROCm) and some DDR5 RAM. ROCm might be tricky to set up for various libraries and inference engines, but then it just works. I don’t rent hardware - don’t want any data to leave my machine.
My use isn’t intensive enough to warrant measuring energy costs.
Disabling root login and password auth, using a non-standard port and updating regularly works for me for this exact use case.
Alas, their IP requirements are too much for me.
Anyway, is there much of xenharmonic sheet music there? Take, for example, Easley Blackwood (but not his books).
Downloading there is straightforward: look at network requests, redownload svg’s of individual pages with wget
and reassemble those into a pdf. I did that today and the resulting quality wasn’t exactly low - though I didn’t examine it too closely. Readability was perfect.
Probably could be automated, but I’m not bothered enough to do so yet.
Or rather Electron apps.
Passwords are safe, it’s temporary session tokens (already invalidated) that were exposed.
Alternatively, ffmpeg -protocol_whitelist file,crypto,data,https,tls,tcp -stats -i <URL.m3u8> -codec copy <FILE.mp4>
.
Also, some m3u8’s are just files containing redirects to other m3u8’s in various resolutions. You might want to extract the one you need and download that.
I have a MediaWiki instance on my laptop (I’ve found the features of all other wikis/mindmaps/knowledge databases decisively insufficient after having a taste of MW templates, Semantic MediaWiki and Scribunto).
Also some smaller things like pihole-standalone, Jellyfin and dictd.
It would. But it’s a good option when you have computationally heavy tasks and communication is relatively light.