How it shall look… # Linux & BSDs # Windows # macOS # State on Fedora 40 Workstation & XFCE Spin… # Screenshots taken from the GNOME bugtracker, copies to not stall their GitLab instance.
More closed and non-customizable systems are much more stable. I guess that’s what GNOME devs are trying to achieve and I don’t really mind it. We have other options for those who need customization. The most used and mainstream one really should be focused on stability. Though I don’t think anyone tried breaking icons before. It’s a bit too much. The app devs will need to make multiple icons for different DEs which is a good thing but shouldn’t be forced like that
You’re talking about two different kinds of stability. They are talking about development stability. You are talking about runtime stability.
One thing is to not break applications that use your library because of changes you introduce to it. Specifically changes that go against the standard you’re supposed to be following.
Another thing altogether is to not go outside the memory limits of the application so it doesn’t get yeeted by the kernel.
The standards are supposed to be the stable thing. If some part of GNOME advertises itself as following a specific standard then it should remain stable in following that standard.
That is a misconception that 99% of the devs don’t understand. Sometimes you do need major changes that break stuff to upgrade the base. GNOME started doing it recently. Keeping old bases for a very long time makes them bloated, hacky, slow and unstable
Implementing old standards does not magically result in unstable software. I can create software today that implements decades old standards using whatever whiz-bang tech is in vogue.
I do lot accept that “old bases” have to succumb to any of the things you suggest either. Refactoring is a thing. You can remove dead code, you can adopt new patterns, you make code modular, you can even extend using new tech if you want.
Linux is 30 years old ( the basic design is decades older ). Should we throw it out? I vote no but allowing Rust into the kernel seems like a good idea. How old is GCC? How old is Microsoft Office? How old is Firefox? This is software you may use every day. Trust me, your life relies on software that is much, much older. How often do you think they rewrite air traffic control systems or core financial software to to make it more “stable” as you suggest?
I mostly hear your argument when devs want to try new tech and cannot justify it any other way. Most often the result is something that is far buggier and missing many features. By the time the features return, the new code is at least as bloated as the original. Around then, somebody usually suggests a total rewrite.
Old architectures are a different story. Sometimes things are not worth fixing in place. In my experience though, this is fairly rare. Even then, in-place migration to something else often makes more sense.
In my view, if you cannot modernize an old code base, it is a skills issue.
In that case, you implement the old API or other interfaces so older things will continue working, while having the new one alongside it, and then phase the old one out when nobody is using it anymore. It’s not that hard to emulate an older API with a newer subsystem. Just a shitload of function wrappers and things so that the things your program used to call now transparently use the new system while the program is unaware anything changed from its perspective.
That’s what happens with the Linux kernel. Linus would go apeshit if one of the devs straight up broke a ton of user programs with a change. He’s already demonstrated his commitment to not doing that in one of his mailing list rants. Because unlike GNOME, the kernel is running some pretty critical things all around the world.
GNOME seems to be treating their DE like their own little pet project that they’re tinkering with alone in the basement without caring that millions are relying on it every day. Breaking a large portion of programs on a regular basis is what I do in the evenings. Not professionals.
Linus Torvalds is author and maintainer of one of the most successful pieces of software ever written ( software that is decades old and still growing in popularity ).
What does Linus says about your philosophy that “Sometimes you do need major changes that break stuff to upgrade the base”? I think his first sentence explains where he stands but he expands on his initial point.
More closed and non-customizable systems are much more stable. I guess that’s what GNOME devs are trying to achieve and I don’t really mind it. We have other options for those who need customization. The most used and mainstream one really should be focused on stability. Though I don’t think anyone tried breaking icons before. It’s a bit too much. The app devs will need to make multiple icons for different DEs which is a good thing but shouldn’t be forced like that
Don’t advertise the theme as standards compliant then.
If stability was their aim, they wouldn’t be breaking stuff all the time…
I would argue that gnome is pretty stable in recent years. Don’t remember when was the last time something crashed.
This might would probably be true for Extensions.
KDE has been unstable for me on Wayland in the past.
You’re talking about two different kinds of stability. They are talking about development stability. You are talking about runtime stability.
One thing is to not break applications that use your library because of changes you introduce to it. Specifically changes that go against the standard you’re supposed to be following.
Another thing altogether is to not go outside the memory limits of the application so it doesn’t get yeeted by the kernel.
The standards are supposed to be the stable thing. If some part of GNOME advertises itself as following a specific standard then it should remain stable in following that standard.
That is a misconception that 99% of the devs don’t understand. Sometimes you do need major changes that break stuff to upgrade the base. GNOME started doing it recently. Keeping old bases for a very long time makes them bloated, hacky, slow and unstable
Implementing old standards does not magically result in unstable software. I can create software today that implements decades old standards using whatever whiz-bang tech is in vogue.
I do lot accept that “old bases” have to succumb to any of the things you suggest either. Refactoring is a thing. You can remove dead code, you can adopt new patterns, you make code modular, you can even extend using new tech if you want.
Linux is 30 years old ( the basic design is decades older ). Should we throw it out? I vote no but allowing Rust into the kernel seems like a good idea. How old is GCC? How old is Microsoft Office? How old is Firefox? This is software you may use every day. Trust me, your life relies on software that is much, much older. How often do you think they rewrite air traffic control systems or core financial software to to make it more “stable” as you suggest?
I mostly hear your argument when devs want to try new tech and cannot justify it any other way. Most often the result is something that is far buggier and missing many features. By the time the features return, the new code is at least as bloated as the original. Around then, somebody usually suggests a total rewrite.
Old architectures are a different story. Sometimes things are not worth fixing in place. In my experience though, this is fairly rare. Even then, in-place migration to something else often makes more sense.
In my view, if you cannot modernize an old code base, it is a skills issue.
In that case, you implement the old API or other interfaces so older things will continue working, while having the new one alongside it, and then phase the old one out when nobody is using it anymore. It’s not that hard to emulate an older API with a newer subsystem. Just a shitload of function wrappers and things so that the things your program used to call now transparently use the new system while the program is unaware anything changed from its perspective.
That’s what happens with the Linux kernel. Linus would go apeshit if one of the devs straight up broke a ton of user programs with a change. He’s already demonstrated his commitment to not doing that in one of his mailing list rants. Because unlike GNOME, the kernel is running some pretty critical things all around the world.
GNOME seems to be treating their DE like their own little pet project that they’re tinkering with alone in the basement without caring that millions are relying on it every day. Breaking a large portion of programs on a regular basis is what I do in the evenings. Not professionals.
I won’t even try to explain. People here only want to argue
I imagine most readers will assume that response indicates that you have no argument.
That is totally fine. I doubt anybody wants to fight. I am glad we were able establish the value you had to add.
If you don’t want to argue (which I highly appreciate) it doesn’t mean others don’t too. This place is quite toxic
No, you’re just dense.
Linus Torvalds is author and maintainer of one of the most successful pieces of software ever written ( software that is decades old and still growing in popularity ).
What does Linus says about your philosophy that “Sometimes you do need major changes that break stuff to upgrade the base”? I think his first sentence explains where he stands but he expands on his initial point.
https://lkml.org/lkml/2012/12/23/75
Ahh, the old “it’s too difficult for us.”
If the gnome devs are so incompetent, why don’t they just get another job?