

It stores the actual HTML structure and assets, so you can still view the page as it was more-or-less intended instead of it getting split up across print pages.
Interests: programming, video games, anime, music composition
I used to be on kbin as e0qdk@kbin.social before it broke down.
It stores the actual HTML structure and assets, so you can still view the page as it was more-or-less intended instead of it getting split up across print pages.
I’m not sure, but this is what my map looks like currently.
Huh. Maybe I’m just too early in the game still (despite having put 10+ hours into it) but crafting materials are like, the one thing I’m not hurting for. It’s got the Zelda rupee problem for me, at least at this point in the game – I’m constantly pegged at the max capacity, and it feels like I have basically nothing to use it on. (I mean, I do use the tools I have found so far, situationally, but I don’t think I’ve been down by more than ~100 or so from max other than for that one wish in the starting area.)
Edit: Ok, I’m a bit further in now, and I see what people mean… -.-
There’s something else going on there besides base64 encoding of the URL – possibly they have some binary tracking data or other crap that only makes sense to the creator of the link.
It’s not hard to write a small Python script that gets what you want out of a URL like that though. Here’s one that works with your sample link:
#!/usr/bin/env python3
import base64
import binascii
import itertools
import string
import sys
input_url = sys.argv[1]
parts = input_url.split("/")
for chunk in itertools.accumulate(reversed(parts), lambda b,a: "/".join([a,b])):
try:
text = base64.b64decode(chunk).decode("ascii", errors="ignore")
clean = "".join(itertools.takewhile(lambda x: x in string.printable, text))
print(clean)
except binascii.Error:
continue
Save that to a file like decode.py
and then you can you run it on the command line like python3 ./decode.py 'YOUR-LINK-HERE'
e.g.
$ python3 ./decode.py 'https://link.sfchronicle.com/external/41488169.38548/aHR0cHM6Ly93d3cuaG90ZG9nYmlsbHMuY29tL2hhbWJ1cmdlci1tb2xkcy9idXJnZXItZG9nLW1vbGQ_c2lkPTY4MTNkMTljYzM0ZWJjZTE4NDA1ZGVjYSZzcz1QJnN0X3JpZD1udWxsJnV0bV9zb3VyY2U9bmV3c2xldHRlciZ1dG1fbWVkaXVtPWVtYWlsJnV0bV90ZXJtPWJyaWVmaW5nJnV0bV9jYW1wYWlnbj1zZmNfYml0ZWN1cmlvdXM/6813d19cc34ebce18405decaB7ef84e41'
https://www.hotdogbills.com/hamburger-molds/burger-dog-mold
This script works by spitting the URL at ‘/’ characters and then recombining the parts (right-to-left) and checking if that chunk of text can be base64 decoded successfully. If it does, it then takes any printable ASCII characters at the start of the string and outputs it (to clean up the garbage characters at the end). If there’s more than one possible valid interpretation as base64 it will print them all as it finds them.
It doesn’t actually include all the media, and – I think – edit history. It does give you a decent offline copy of the articles with at least the thumbnails of images though.
Edit: If you want all the media from Wikimedia Commons (which may also include files that are not in Wikipedia articles directly) the stats for that are:
Total file size for all 126,598,734 files: 745,450,666,761,889 bytes (677.98 TB).
according to their media statistics page.
Nginx is running in Docker
Are you launching the container with the correct ports exposed? You generally cannot make connections into a container from the outside unless you explicitly tell Docker that you want it to allow that to happen… i.e. assuming you want a simple one-to-one mapping for HTTP and HTTPS standard ports are you passing something like -p 80:80 -p 443:443
to docker run
on the command line, adding the appropriate ports in your compose file, or doing something similar with another tool for bringing the container up?
I’ve put drives into standby mode with the gnome disks GUI tool on my regular desktop when they were being noisy and I wanted some peace for a while. If the drive was mounted before I put it to sleep, trying to access something on the disk will cause it to spin back up.
I got the KVM used in good condition. It’s an older model – but I went with it anyway since it was a drop-in replacement for my 2-port setup and getting it used was much cheaper than their newer models. The newer ones support higher resolution/frame rates though, I think; I know this one won’t do frame rates above 60FPS properly even though my monitor is capable of it when plugged in directly.
There’s a few quirks with this setup. Probably most annoying is that moving the mouse typically causes computers to wake from sleep (like pressing a key on a keyboard normally does…); I think there’s a way to mask that event off with udev rules but, eh, even a decade or so after getting the original 2-port KVM I haven’t cared enough to actually bother working it out, so I guess it’s not that big of an issue to me… :p
I upgraded my KVM switch recently to one with more ports (from 2 to 4). I used to have to get up and physically rewire the audio and switch the video input selection on my monitor when I wanted to use it instead of my main Linux desktop or ancient Win7 PC before. Now that I can just type a couple keystrokes to switch back and forth I’m actually using my Deck way more often…
Are you running different versions of the software? (e.g. different versions of ffmpeg, maybe?)
I don’t like Anubis because it requires me to enable JS – making me less secure. reddthat started using go-away recently as an alternative that doesn’t require JS when we were getting hammered by scrapers.
netstat -tp
– that’ll show you TCP connections and the associated program, doing a DNS lookup for the IPs they’re connected to. You may need elevated permissions to see what some processes are.
There are a bunch of other options (e.g. -n
to get numeric output instead of looking up names, -l
to get programs listening for incoming connections, etc); check the man pages for more details.
I just right click on the terminal to change the profile to whatever I feel like it should be in the moment (usually red). I do it by reflex, basically. I never felt the need to try to set up automation for different servers, but I expect there’s probably a way to do that if you really wanted to.
“You love the robot more than me!” 💔️
I (manually) set the background color of my terminals depending on the machine I connect to. I currently have profiles for red, green, blue, and black backgrounds with black as my default. I usually use red for ones I want to be especially careful of.
I’ve been trying to figure out a related sort of video streaming setup for work (without Owncast, but with a similar sort of 24/7 goal plus other considerations) and have been looking into using ffmpeg’s capabilities to output either HLS or DASH segments + manifests. (FFMPEG can do both but I don’t know which would be better for my needs yet.) The sources I’m working with are RTSP/RTP instead of RTMP and I only need streaming to browser clients currently – although it working with VLC naturally by pointing it to the manifest is nice.
HLS and DASH work by having videos split into small chunks that can be downloaded over HTTP, so just replacing the manifest allows for continuous streaming (the client pulls it repeatedly) without the server needing to maintain a continuous connection to the client.(Fan out to CDNs works naturally since the video chunks are just files that can be served by any web server.)
It should be possible to do some creative things by either creating / modifying the manifests myself with scripting or by piping chunks into another instance of ffmpeg from a script. (I’ve done something similar using -f image2pipe
in the past, but that was for cases where I need to do things like create a video from an image gallery dynamically.) That’s as far as I’ve gotten with it myself though.
I don’t know what the right answer is either, but I’m also interested in finding out and hopeful you get additional responses.
I have an older version of TrueNAS on it from when it was still FreeBSD based (instead of Linux). I might replace it with Scale whenever I get around to doing maintenance on it next – or maybe just go to stock Debian or something since I don’t use most of the bells-and-whistles.
I run my NAS that way too. I just mount it and play videos with VLC if I want to watch something I have on it. The main reason I have a NAS is because I ran out of drive bays in my main system a few years ago… Works fine for my needs currently; no need to make it more complicated.
You can use tesseract -l jpn input.png -
on the command line to have it print out the text from input.png into the console if you’ve got the language files for Japanese installed. (There’s also language files for vertical text and a few others for script in my package manager.) Alternatively give the filename (w/o extension) instead of -
to write the output into a .txt
file.
On Mint, I think I did sudo apt install tesseract-ocr tesseract-ocr-jpn
to get it working for the simple case of horizontal text; been a while though.
This takes a snapshot of the HTML elements from when they were loaded in your browser. If the page loads content dynamically, HTTrack won’t save it but this can. (i.e. this works better on crappy modern sites that need JS to even just load the article text…)