Just another Swedish programming sysadmin person.
Coffee is always the answer.
And beware my spaghet.
Going to be really amazing to play Factorio again without knowing how to solve everything.
Well, one part of it is that Flatpak pulls data over the network, and sometimes data sent over a network doesn’t arrive in the exact same shape as when it left the original system, which results in that same data being sent in multiple copies - until one manages to arrive correctly.
We’re mirroring the images internally, not just because their mirrors suck and would almost double the total install time when using them, but also because they only host the images for the very latest patch version - and they’ve multiple times made major version changes which have broken the installer between patches in 22.04 alone.
What is truly bloated is their network-install images, starting with a 14MB kernel and 65MB initrd, which then proceeds to pull a 2.5GB image which they unpack into RAM to run the install.
This is especially egregious when running thin VMs for lots of things, since you now require them to have at least 4GB of RAM simply to be able to launch the installer at all.
Compare this to regular Debian, which uses an 8MB kernel and a 40MB initrd for the entire installer.
Or some larger like AlmaLinux, which has a 13MB kernel and a 98MB initrd, and which also pulls a 900MB image for the installer. (Which does mean a 2GB RAM minimum, but is still almost a third of the size of Ubuntu)
If you’re going to post release notes for random selfhostable projects on GitHub, could you at least add the GitHub About text for the project - or the synopsis from the readme - into the post.
Been using the KeyDB fork for ages anyway, mainly because it supports running in a multi-master / active-active setup, so it scales and clusters without the ridiculousness that is HA Redis.
I think the only project I’ve seen so far where I’ve felt that a blockchain has actually been the correct choice is Alfis, which is a decentralized DNS that uses the blockchain as the public append-only ledger that it is, and it uses proof-of-work to add arbitrary costs to updates - to make spamming or namesquatting expensive.
I feel like this could go really well together with Piet.
Just imagine; an album consisting of a bunch of Velato programs with Piet code as the artwork.
Well, there are people running Linux in all manner of ways, like VRChat shaders.
If you build a linked list in C, and put the pointer to the next entry as the first element in your struct, then you only need a single variable (and two comparisons) to do sorted insertion into the list.
Steve Rogers might have America’s Ass, but Mothman has the Buns of Steel.
People really have no love for JPEG-XL - though to be fair that’s mainly Google’s fault at the moment.
Haven’t really used any proper JMAP clients - since the setup is broken anyway, so mainly just curl.
You could also just run IMAP/JMAP/SMTP as separate components, I can’t see any place in the Stalwart documentation - or in the Docker image itself - where monolith is the only option.
I haven’t tested the setup myself yet, but me and another root are planning on testing a setup of Stalwart to replace a semi-broken IMAP/JMAP setup for a computer club, keeping the SMTP as is.
Reading the Dockerfile in their repo, it’s simply a clean debian:slim with four compiled rust binaries placed into it. There’s no services, no supervisord, nothing except the mail server binaries themselves.
I can already imagine so many fun ways this could be used.
“We interrupt your regular scheduling to bring you this additional bit of Factorio hype.”
Let me tell you about Bumblebee and their issue #123, though that one’s even worse seeing as installing system packages are done as root.
(Their install/update commands included rm -rf /usr /lib/nvidia-current/xorg/xorg
)
Probably not what you’re looking for, but I’m going to note that Turris make some great OpenWRT routers.
Currently running theTurris Omnia, and using both Wireguard and Yggdrasil through it.
It’s somewhat amusing how Itanium managed to completely miss the mark, and just how short its heyday was.
It’s also somewhat amusing that I’m still today helping host a pair of HPE Itanium blades - and two two-node DEC Alpha servers - for OpenVMS development.