I'm new to docker and all of my shit stopped working recently. Just wouldn't load. Took about a half hour to find out that old images were taking up about 63GB on my 100GB boot partition, resulting in it being completely full.
I added the command to prune 3 month old images to my update scripts.
One of the things I dislike about Rust is the massive amount of disk space and time it takes to do a download, compile, test run.
2GB of dependencies and build files for a 200K binary is a bit much.
Clean all the cache downloads of Arch Linux Packages
pacman -Scc
Remove unused docker networks and images
docker system prune --all
Cleanup untracked git files that might be in .gitignore such as build and out directories (beware of losing data, use "n" instead of "f" for a dry run)
git clean -xdf
Do an aggresive pruning of objects in git (MIGHT BE VERY SLOW)
My dad's Linux setup couldn't log in. After a bit of investigation, starting the session manually and so on, i got a hunch and indeed; i saw in Baobab that the backup script took the wrong disk, filled up the one with home, making it slow, so the log-in thingie timed out, failing the session.
Ncdu is my go-to tool. Can't live without it on the servers I administer. However from this thread I've also learned about gdu, diskonaut and du-dust that I need to check out.
45347740 bytes is 43.247 megabytes. That is to say, the entire install of filelight is only 43 megabytes.
KDE packages have many dependencies, which cause the packages themselves to be extremely tiny. By sharing a ton of code via libraries, they save a lot of space.
Basically all KDE apps have the same dependency set. So install one and the next ones will only install the app most likely. On KDE itself you'd already have these.
I don't think you'll need to do that, unless you are planning to download files that are over 4Gb long and/or you are using a potato that has less than 1 Gb of ram.
t. I've set my entire ram into a ramdisk, and the performance actually IMPROVED compared to not setting a ramdisk at all.
If you need a more interactive method, gdu is awesome. And if you're using btrfs, btdu gives preliminary results instantly (which get more precise over time).
I was trying to remember the name of kdirstat ladt night when I stumbled across filelight and made use of that instead.
And now there's a thread on this exact topic. Y'all need to quit it with all this Truman Show nonsense, Baader-Meinhof alone isn't enough to explain how frequently shit like this happens. XD
I'm here to promote fclones. I've used it twice and recovered over a terabyte on my NAS the last time I used it. I'm not affiliated. Hyperspace for Mac is similar (but different) and I haven't used it, but it was developed by my favorite nerd podcast host. I'm planning to test it out eventually, but the latest fclones run was only about a month ago, so it doesn't make sense to try it yet.
I'm still pretty new to Linux so I break stuff pretty often, like recently I was trying to get opencl working with my amd gpu and I ended up causing every video I played to stutter constantly.
And I've been trying out new software to control fans or rgb and following guides making me enter commands until I figure out something that works I note it down so when I do a fresh install again I can easily configure it without all the trial and error etc and install only the software I found that I liked
I usually keep important stuff on my server but things like games and stuff I purge with the fresh install and just download the games I'm actively playing, also helps clear up any issues from installing random junk during the months between as I settle on what programs I like