You make it sound as if it’s a religion … UNIX isn’t a dogma handed down by an infallible being, just a piece of software that made sense for its time. Todays needs are different than the needs 40 years ago, so ofc things have to change.
You make it sound as if it’s a religion … UNIX isn’t a dogma handed down by an infallible being, just a piece of software that made sense for its time. Todays needs are different than the needs 40 years ago, so ofc things have to change.
Actually most of the money are just 1‘s and 0‘s in a computer, coming into existence from nothing and vanishing into nothing. Fiat money backed by “trust”. As Henry Ford once said:
It is well enough that people of the nation do not understand our banking and monetary system, for if they did, I believe there would be a revolution before tomorrow morning.
No, that’s a very bad idea. BTRFS has deduplication, without that the snapshots would take up way to much space. Also it’s too many writes since ext4 doesn’t use cow and would have to do distinct writes for every snapshot.
The 240 gb are plenty for a root system without /home and years worth of snapshots on a btrfs volume, only the changes take up space so the amount of snapshots hardly matters.
For /home either ext4, xfs or btrfs is fine. Personally I only use a single btrfs volume and put certain folders in their own subvolumes so they can have different settings for snapshots(no snapshots for /home, tmp and cache folders).
That 40 year old X protocol might be the issue here, use wayland for multi monitor with different resolutions.
Sounds like a lawsuit waiting to happen. It’s pretty much the definition of unfair competition, I mean Apple makes cables and artificially hampers competitors cables unless they pay some money?
Maybe the fines are too low so far if they test us like that.
So it’s people borking it and not the “system itself”. You have control over which people are involved in the software on your system ne it affects the likelihood of it ending up borked.
Nope, it doesn’t. It always requires human assistance or random hardware failure. It’s either the user, the distro, package maintainer or upstream fucking up.
Personally I blame half on users for picking the wrong distro(not suited for beginners) and half on the linux community giving poor advice(use the terminal). Not everyone has the time or inclination to become a power user and if people wouldn’t be so thickheaded and recommending the same problematic distros over and over to these people it wouldn’t be such a mess.
I have a 80 year old neighbour whose old windows laptop was a mess and who was open to trying a new OS(because he couldn’t operate windows either anyway). I setup a MicroOS system for him, put a taskbar extension on it and showed him how to install software from gnome-software(which only has flatpaks). ZERO problems in half a year. He doesn’t have to do anything nor learn anything. He happily installed some card games, reads the few websites he follows and that’s it.
Moderating large subreddits is a ton of work, chances are the people willing to do that work are already doing it and will thus either be part of the ones being removed or to busy with their current workload elsewhere. I mean it’s basicly a unpaid volunteer job that doesn’t benefit your resume, doesn’t exactly sound like something people stand in line for.
This was refuted long ago. Systemd isn’t a single binary file doing everything, it’s a project that has many different binaries doing many different things. The only difference is that they are developed under one project to ensure they integrate well with each other. What your doing is like complaining that glibc tries to do everything, I mean it does open, read, write, malloc, printf, getaddrinfo, dlopen, pthread_create, crypt, login, exit and more… Xorg would be another example of a project that does many things instead of one very well.