Which compression level are you using? My old server is able to compress flac’s at the highest (and therefore “slowest”) compression level at >50x speed, so bumping the level up shouldn’t be too hard on your CPU.
Which compression level are you using? My old server is able to compress flac’s at the highest (and therefore “slowest”) compression level at >50x speed, so bumping the level up shouldn’t be too hard on your CPU.
Have you tried the official guide from the jellyfin website?
As for the guide this AI generated: it bothers me that they instruct you to use chocolatey for the *arrs, but still advice you to install docker, qbittorrent and jellyfin manually (all of which have chocolatey packages). I disagree with the comment that external storage would be recommended, as internal storage is generally more reliable (depending on a lot of factors of course). Also, I believe the “adding a library”-section of the jellyfin setup is a bit too short to be of any use, and would recommend referring to the jellyfin docs instead.
This guide also doesn’t explain how to make jellyfin accessible outside of your LAN. Once again, I’d recommend referring to the jellyfin docs if you want to do this.
I personally have only set up qbittorrent, jellyfin and docker (not the *arr suite), so I can’t comment on the completeness of the guide, but I wouldn’t trust it too much (seeing the previous oversights).
And finally, as someone who started their selfhosted server journey on windows: don’t. There is a reason why almost all guides are written for linux, as it is (in my humble opinion) vastly superior for server usage once you get used to it.
didn’t know that was a part of bisexuality
I should probably flee before I get eaten by an army of blahåjar (apparently that’s the correct plural?)
Oh I don’t mind the nitpicking, thanks for the explanation! I (apparently erroneously) thought “demake” and “decompile” were synonyms. Guess I’m one of today’s 10000.
In that case the (now taken down, but forked a gazillion times) portal64 project would be a correct example of a demake, right?
interested in females
Username checks out, though I’m assuming you meant “demakes”?
Anyways, the demake I’m most familiar with is the in-progress Lego island. The YouTuber behind it documented part of the process in vlogs (linked on the GitHub page), so that might be an interesting starting point.
I believe SSD’s don’t actually experience wear when reading data, only when writing. Loading more data from SSD’s shouldn’t cause any premature failure. Overwriting more data each update could cause the drive to fail slightly earlier, but if that’s really that big of a concern, you’d be best of moving to Debian stable (no updates means no SSD writes).
If SSD wear prevention is really that big of a concern, you might be interested in profile-sync-daemon (https://wiki.archlinux.org/title/Profile-sync-daemon). It reduces writes to hard drives by keeping your browser profile in RAM, and only periodically syncing it to disk.
Though I must add that SSD’s wearing out really isn’t that much of an issue with modern drives. With normal usage, a drive will become obsolete long before it actually wears out.
Not OP (OC? Not the person you were helping, you get what I mean), are you sure you meant df -h
? fd -H
seems more useful for to me when trying to find a specific file in a dotfolder, though even that didn’t work on my system. fd
ignores ~/.config
by default, so you need to use fd -u
(which is an alias for fd -I -H
) to find the correct files.
Anyways, from your description it seems like the correct file would be ~/.config/kwinrc
, which exists on my system.
You could look at the awesome-selfhosted list, specifically these two sections:
https://awesome-selfhosted.net/tags/recipe-management.html
https://awesome-selfhosted.net/tags/task-management--to-do-lists.html
I don’t have any experience with any of those, but there might be something that fits your needs.
If the installer is small enough (<650MB I believe), you can upload it to virustotal.com to have it be scanned by ~65 antivirus programs
Ah, it looks like we have a small misunderstanding. I thought you were talking about uncompressed video, which is enormous. This is only used in HDMI cables for example. A 1080p60 uncompressed video is 2.98Gbit/s, or about 1.22 terabytes per hour.
A remux is “uncompressed” in the sense that it isn’t recompressed, or in this case transcoded. A remux is still compressed, just to a lesser degree than a transcode. This means the files are indeed larger, but the quality is also better than transcodes.
To clarify the article’s confusing statement: they claim that remuxes can reduce size by throwing away some audio streams, while keeping the original video. This is true, but the video itself hasn’t gotten any smaller: you are simply throwing away other information.
Remuxes aren’t uncompressed, nor are they losslessly compressed. They’re just a 1:1 direct copy from some other medium (generally blu-rays or DVD’s).
The more you compress the longer and more CPU intensive it is to decompress
I believe this is becoming less and less true with modern algorithms. Take for example ZSTD: while the compression speeds differs by several orders of magnitude between the fastest and slowest modes, the decompression difference is only about 20%. The same holds true for flac, where the decompression speed is pretty uniform across all compression levels.
These algorithms probably aren’t used by repacked like fitgirl (so your answer is generally correct in the context of repacks). I do believe it is still interesting to see these new developments in compression techniques.
YouTube would be smart enough not to advertise Adobe creative cloud in the pre-roll ads of this video, right? Right???