Replacing garbage with sewer water. Not exactly an improvement.
Bazzite comes with wine all setup by default. KDE’s file managerl can integrate running exe with wine on a default prefix automatically.
Most distributions and DEs already package wine in a set it and forget it configuration. Wine by default has a system wide prefix such that clicking on any exe in the file system automatically runs it on the default prefix. This way of doing things predates wsl by a long time. It is just safer and better practice to setup a new prefix for every software, specially if they are games.
My biggest fear is that so far Nvidia has a track record of introducing regressions and new bugs with each new driver version. Just a week ago all my flatpaks weren’t working on Wayland, again. It happens almost with every single update. Some games that are native or platinum randomly stop working and it takes several updates before they start working again. While on AMD everything just works all the time and regressions are solved in a day not weeks. It’s just annoying.
If gaming is top priority. Go all amd, disregard Nvidia. AMD has extraordinary linux support and if it runs on the steam deck it will run on any all AMD machine.
Always remember to disable secureboot and remove bitlocker before installing linux on a oem windows machine. They make it hell to remove that malware from newer machines.
Please remember that no one is taking anything away from you. If you don’t like it, you don’t have to use it. Traditional package managing isn’t going away any time soon. You are safe. Others can have their preferred tech, and you don’t have to like it. It’s ok to have different tastes.
It’s not a flaw. Ostree is a last resort, you should be using containerized software. Layering a package should only be done when strictly necessary and not as the regular way to manage packages. If you need an overtly customized system, you use Nix or universal blue to design your new system declaratively and create your custom image.
Then just install KDE in your Arch install. Or use endeavorOS with KDE, or any other Arch based OS with KDE. Don’t be dismissive of other people’s interests.
Settings live in user space. Software exist in containers like AppImage, Flatpak or Distrobox. If something need deep system integration, they can be layered on top of the system in the user layer. Immutable does NOT mean less control. Just exerting control over the system in a different, usually more systematic, automatic and deterministic way.
your base distro is immutable, then any extra changes go on an additional mutable layer
That is exactly how OsTree and other layering solutions work. Only Nix requires a whole distro rebuild.
There’s KDE Neon already. The whole point of this distribution is the atomic immutable part.
Maybe they’ll fix the sddm custom theming? It’s currently broken on all immutables and doesn’t allow custom themes.
On Mint, you troubleshoot the wifi antenna following a guide once and then you’re done. On Bazzite you probably just needed to click to change to X11 instead of plain Plasma, on the login screen. I would bet money that you have an Nvidia GPU. Sometimes Nvidia breaks the drivers support on Wayland. They intentionally neglect it in order to keep your kind of mentality around.
On Windows, MS is going to eventually fix the workarounds so you can’t update your computer anymore.
How is having to apply workarounds to keep windows working on old machines any different from troubleshooting the occasional linux issue? It’s a rethorical question, the difference is that the workaround on Windows is mandatory while the Linux troubleshoot is nowadays rare and usually related to edge cases.
Some of the workarounds in this article are far more involved and convoluted than what I’ve ever had to do in 15 years of linux. Some are even dangerous for system stability and security. My very recent install of bazzite in a new laptop has been a perfectly out of the box it just works experience. Not even having to open the terminal. 100% friendly GUI without compromising flexibility, power and customizability. Today, suggesting linux with a solid desktop environment like KDE plasma is just foolproof. The end user will be using exactly the same knowledge and habits of Windows, without the harassment machine that is MS now. The change is not learning a new OS, is just switching a few assumptions on how some advanced things work.
If this were done by MS or Apple, who lack any shred of respect left, sure. If it were a material change on how the code works, certainly it would be most concerning. But what happened was blown entirely out of proportion for who Bitwarden has been and how they’ve acted in the past. They are still ethically very solid. And it was an immaterial change in the build tools, that could very well have been neglectful or accidental.
That’s a poor understanding of the situation. Nothing in the licensing changed. The SDK has always been the proprietary business to business secrets management product. The client integrates with and can use that SDK to provide the paid service to businesses. The client and the server side management of password has always been and still is FOSS.
This was apparently an accidental change in the build code (not the client code, just the building scripts) that required the inclusion of the SDK to build the client when actually it has never and doesn’t really need any of that code. It prevented building the client without accepting the SDK license. Which it shouldn’t.
This was fixed and some things will be put in place so it doesn’t happen again. Nothing in the licensing scheme changed, at all. This is not a catastrophic enshittification event. A Dev was just being lazy and forgot to check the dependencies on the build chain before their commit.
Firefox has all those things already (albeit vertical tabs are experimental still). Just without the AI marketing and the built-in identity theft.
Well, you see, that’s the really hard part of LLMs. Getting good results is a direct function of the size of the model. The bigger the model, the more effective it can be at its task. However, there’s something called compute efficient frontier (technical but neatly explained video about it). Basically you can’t make a model more effective at their computations beyond said linear boundary for any given size. The only way to make a model better, is to make it larger (what most mega corps have been doing) or radically change the algorithms and method underlying the model. But the latter has been proving to be extraordinarily hard. Mostly because to understand what is going on inside the model you need to think in rather abstract and esoteric mathematical principles that bend your mind backwards. You can compress an already trained model to run on smaller hardware. But to train them, you still need the humongously large datasets and power hungry processing. This is compounded by the fact that larger and larger models are ever more expensive while providing rapidly diminishing returns. Oh, and we are quickly running out of quality usable data, so shoveling more data after a certain point starts to actually provide worse results unless you dedicate thousands of hours of human labor producing, collecting and cleaning the new data. That’s all even before you have to address data poisoning, where previously LLM generated data is fed back to train a model but it is very hard to prevent it from devolving into incoherence after a couple of generations.