• 0 Posts
  • 125 Comments
Joined 3 years ago
cake
Cake day: June 14th, 2023

help-circle
  • Menu bar at the top at least makes some sense - it’s easier to mouse to it, since you can’t go too far. Having menus per-window like Linux, or like Windows used to before big ugly ribbons became the thing, is easier to overshoot. (Which is why I always open my menu bars by pressing ‘alt’ with my left thumb, and then using the keyboard shortcuts that are helpfully underlined. Window likes to hide those from you now since they’re ‘ugly’, and also makes you mouse over the pretty icons to get the tooltip that tells you what they are, which is just a PITA. Pretty != usable.)

    Mac OS has had the menu at the top since before it was a multitasking OS. They had them there on the first Mac I ever used, a Mac Classic 2 back in 1991 or so, and it was probably like that before then too. It’s not like they’ve been ‘innovating’ that particular feature and annoying their users.


  • Generally, companies are trying to maximise profit, which means that the price will be reduced only when it’s stopped selling at the previous and they want to make sales the next, more price-conscious, segment of the market. They might want some quick bucks if the company is in financial trouble, or to ‘make the news’ with a sale if they need some publicity.

    BG3 sold shedloads, is still selling shedloads, was on multiple games-of-the-year list and generally ranks amongst the best games of all time, often at the top; and Larian seem sufficiently flush with cash from the success of it. So like you say, don’t hold your breath waiting for a big sale, it doesn’t make sense for them to do that.


  • Data centre GPUs tend not to have video outputs, and have power (and active cooling!) requirements in the “several kW” range. You might be able to snag one for work, if you work at a university or at somewhere that does a lot of 3D rendering - I’m thinking someone like Pixar. They are not the most convenient or useful things for a home build.

    When the bubble bursts, they will mostly be used for creating a small mountain of e-waste, since the infrastructure to even switch them on costs more than the value they could ever bring.



  • systemd-networkd gets installed by default by Arch, integrates a bit better with the rest of SystemD, doesn’t have so many VPN surprises, and the configuration is a bit more obvious to me - a few config files rather than NetworkManager’s “loads of scripts” approach. Small niggles rather than big issues.

    Really, I just don’t want duplication of services - more stuff to keep up-to-date. And if I’ve got SystemD anyway, might as well use it…


  • NetworkManager dependencies can now be disabled at build time…

    Nice. It was a damned nuisance that Cinnamon brought its own network stack with it. All my headless servers and my Plasma gaming desktop use systemd-networkd, which meant that my Cinnamon laptop needed different configuration. Now they can all be the same.

    Hopefully the new release will bash a few of the remaining Wayland bugs; Plasma is great but I prefer Cinnamon for work, and it’s just too buggy for gaming on a multi-monitor setup at the moment.


  • Yeah. You know the first time you install Arch (btw), and you realise you’ve not installed a working network stack, so you need to reboot from the install media, remount your drives, and pacstrap the stuff you forgot on again? Takes, like, three minutes every time? Imagine that, but you’ve got a kernel compile as well, so it takes about half an hour.

    Getting Gentoo so that it’ll boot to a useful command line took me a few hours. Worthwhile learning experience, understand how boot / the initramfs / init and the core utilities all work together. Compiling the kernel is actually quite easy; understanding all the options is probably a lifetime’s work, but the defaults are okay. Setting some build flags and building ‘Linux core’ is just a matter of watching it rattle by, doesn’t take long.

    Compiling a desktop environment, especially a web browser, takes hours, and at the end, you end up with a system with no noticeable performance improvements over just installing prebuilt binaries from elsewhere.

    Unless you’re preparing Linux for eg. embedded, and you need to account for basically every byte, or perhaps you’re just super-paranoid and don’t want any pre-built binaries at all, then the benefits of Gentoo aren’t all that compelling.





  • The router provided with our internet contract doesn’t allow you to run your own firmware, so we don’t have anything so flexible as what OpenWRT would provide.

    Short answer; in order to Pi-hole all of the advertising servers that we’d be connecting to otherwise. Our mobile phones don’t normally allow us to choose a DNS server, but they will use the network-provided one, so it sorts things out for the whole house in one go.

    Long, UK answer: because our internet is being messed with by the government at the moment, and I’d prefer to be confident that the DNS look-ups we receive haven’t been altered. That doesn’t fix everything - it’s a VPN job - but little steps.

    The DHCP server provided with the router is so very slow in comparison to running our own locally, as well. Websites we use often are cached, but connecting to something new takes several seconds. Nothing as infuriating as slow internet.


  • Big shout out to Windows 11 and their TPM bullshit.

    Was thinking that my wee “Raspberry PI home server” was starting to feel the load a bit too much, and wanted a bit of an upgrade. Local business was throwing out some cute little mini PCs since they couldn’t run Win11. Slap in a spare 16 GB memory module and a much better SSD that I had lying about, and it runs Arch (btw) like an absolute beast. Runs Forgejo, Postgres, DHCP, torrent and file server, active mobile phone backup etc. while sipping 4W of power. Perfect; much better fit than an old desktop keeping the house warm.

    Have to think that if you’ve been given a work desktop machine with a ten-year old laptop CPU and 4GB of RAM to run Win10 on, then you’re probably not the most valued person at the company. Ran Ubuntu / GNOME just fine when I checked it at its original specs, tho. Shocking, the amount of e-waste that Microsoft is creating.



  • Only has the functionality that you need, everything is obviously in its place. Runs incredibly quickly without using a lot of resources, and then gets out your way when you’re trying to do stuff. No settings hidden away because they might confuse novice users. No bullshit shoehorned in by managers.

    Apart from the ugly font rendering, this might be as good as the Windows UI ever got. WinNT looks the same and has almost incomparable stability improvements, but only if you’ve the right hardware to run it. WinXP starts the downhill slide with ‘appearance over functionality’ and the hot mess of the control panel.

    I could live with how OP has things set up here; my own copy of Plasma doesn’t look a million miles from this.





  • I understand that things have changed a bit since I first moved over to Linux - moving from Red Hat Linux to Ubuntu ‘Warty Warthog’ was such a revelation in overall user-friendliness and usability, back in the day. But upgrading my graphics card from an NVidia one to an AMD was a similar change. I might have only just installed the base operating system and a desktop environment and haven’t got around to a web browser yet, but I’ve already got full hardware accelerated graphics - that’s crazy.

    Most distros now make the NVidia drivers a complete non-issue, I think? My 6600XT is requiring just a few too many compromises on new games, so I’ll need something new too, sooner or later. I used to hold off on graphics cards updates until I could get something twice as good so that it was a noticeable upgrade, but I could buy a pretty decent second-hand car for all the ones which are ‘twice as good’ now.

    An upgrade from a 1050 Ti shouldn’t be such a problem. Well done on keeping it alive so long - I had a GeForce GTX 970 that would have been a similar age, but it let out its magic smoke years ago.


  • Indeed - most Java IDEs have FernFlower built in, so it’s dead easy.

    Decompiled Java is surprisingly close to the original, especially compared to eg. decompiled C++; good luck with that. You get all the class, function and variable names back on the original line numbers.

    What you do not get back is any comments. So you can see what and how, but not why. Admittedly, most comments are kind of useless and do not explain ‘why’ very well, but for weird-but-critical code they can be essential.


  • Indeed - I’ve seen more people recommend Hannah Montana Linux (apt-based) than any of those for newcomers recently.

    You are entirely right that a Linux distribution is really just its package manager, the default packages installed, and some remote repositories which may (or may not) have had some customisation applied, which will have been pulled and built from a source repository somewhere. All that’s really needed to swap between eg. Arch, Manjaro or Cachy is to update the repo files and issue a package manager update command, although I’d probably like to verify my backups and get a stiff drink first.

    The House of Linux is built out of bricks, and the bricks aren’t that scary - you can take them to bits and look at them if you like, they’re usually zipped-up folders of text files and the binaries you’d get from compiling them yourself. But if that’s not what you’re used to, then yeah - 🤯 .

    In all seriousness, I wish that most distros had art half as good as what Void Linux has - got some really gifted people, there.