• 1 Post
  • 37 Comments
Joined 1 year ago
cake
Cake day: June 7th, 2023

help-circle












  • You’re not wrong, but there’s a kind of irony in it when you talk about ending humanity because of it. There’s a lot to hate about humanity if you have humanity and have human values. There’s nothing objectively wrong about being cruel or destructive or dishonest or greedy or abusive or murderous and I imagine most hypothetical alien species would look at those things and say “what’s wrong with any of that?”

    But because humans evolved as social creatures and our survival depended upon trusting one another, we’re constantly trying to judge ourselves against values that can’t actually be met. So we look at ourselves and say we’re a really horrible species, but that statement only makes sense because ironically we’re a really glorious species that’s fabricated these completely irrational things like love and compassion and empathy and honesty and sacrifice that no other species has (though many other social species do have bits and pieces of them).

    And we’ll forever hate ourselves for not being able to live up to our own values.




  • It’s a bit more complicated than that. System load is a count of how many processes are in an R state (either "R"unning or "R"eady). If a process does disk I/O or accesses the network, that is not counted towards load, because as soon as it makes a system call, it’s now in an S (or D) state instead of an R state.

    But disk I/O does affect it, which makes it a bit tricky. You mentioned swapping. Swapping’s partner in crime, memory-mapped files, also contribute. In both of those cases, a process tries to access memory (without making a system call) that the kernel needs to do work to resolve, so the process stays in an R state.

    I can’t think of a common situation where network activity could contribute to load, though. If your swap device is mounted over NFS maybe?

    Anyway, generally load is measuring CPU usage, but if you have high disk usage elsewhere (which is not counted directly) and are under high memory pressure, that can contribute to load. If you’re seeing a high load with low CPU utilization, that’s almost always due to high memory pressure, which can cause both swapping and filesystem cache drops.



  • BitWarden+PiHole+NextCloud+Wireguard combined will add to like maybe 100MB of RAM or so.

    Where it gets tricky, especially with something like NextCloud, is the performance you see from NextCloud will depend tremendously on what kind of hard drives you have and how much of it can be cached by the OS. If you have 4GB of RAM, then like 3.5GB-ish of that can be used as cache for NextCloud (and whatever else you have that uses considerable storage). If you have tiny NextCloud storage (like 3.5GB or less), then your OS can keep the entire storage in cache, and you’ll see lightning-fast performance. If you have larger storage (and are actually accessing a lot of different files), then NextCloud will actually have to touch disk, and if you’re using a mechanical (spinning rust) hard drive, you will definitely see the 1-second lag here and there for when that happens.

    And then if you have something like Immich on top of that…

    And then if you have transmission on top of that…

    Anything that is using considerable filesystem space will be fighting over your OS’s filesystem cache. So it’s impossible to say how much RAM would be enough. 512MB could be more than enough. 1TB could be not enough. It depends on how you’re using it and how tolerant you are of cache misses.

    Mostly you won’t have to think about CPU. Most things (like NextCloud) would be using like <0.1% CPU. But there are some exceptions.

    Notably, Wireguard (or anything that requires encryption, like an HTTPS server) will have CPU usage that depends on your throughput. Wireguard, in particular, has historically been a heavy CPU user once you get up to like 1Gbit/s. I don’t have any recent benchmarks, but if you’re expecting to use Wireguard beyond 1Gbit/s, you may need to look at your CPU.


  • Yes, with some big "if"s. NextCloud can work very well for a large organization if that large organization has a “real” IT department. I use “real” to describe how IT departments used to work 20+ years ago, where someone from IT was expected to be on call 24/7, they built and configured their own software, did daily checks and maintenance, etc. Those sorts of IT departments are rare these days. But if they have the right personnel, it can definitely be done. NextCloud can be set up with hot failovers and fancy stuff like that if you know what you’re doing.