Are we stuck with the same Desktop UX forever? [video]
87 comments
·December 13, 2025linguae
I enjoyed this talk, and I want to learn more about the concept of “learning loops” for interface design.
Personally, I wish there were a champion of desktop usability like how Apple was in the 1980s and 1990s. I feel that Microsoft, Apple, and Google lost the plot in the 2010s due to two factors: (1) the rise of mobile and Web computing, and (2) the realization that software platforms are excellent platforms for milking users for cash via pushing ads and services upon a captive audience. To elaborate on the first point, UI elements from mobile and Web computing have been applied to desktops even when they are not effective, probably to save development costs, and probably since mobile and Web UI elements are seen as “modern” compared to an “old-fashioned” desktop. The result is a degraded desktop experience in 2025 compared to 2009 when Windows 7 and Snow Leopard were released. It’s hamburger windows, title bars becoming toolbars (making it harder to identify areas to drag windows), hidden scroll bars, and memory-hungry Electron apps galore, plus pushy notifications, nag screens, and ads for services.
I don’t foresee any innovation from Microsoft, Apple, or Google in desktop computing that doesn’t have strings attached for monetization purposes.
The open-source world is better positioned to make productive desktops, but without coordinated efforts, it seems like herding cats, and it seems that one must cobble together a system instead of having a system that works as coherently as the Mac or Windows.
With that said, I won’t be too negative. KDE and GNOME are consistent when sticking to Qt/GTK applications, respectively, and there are good desktop Linux distributions out there.
gtowey
It's because companies are no longer run by engineers. The MBAs and accountants are in charge and they could care less about making good products.
At Microsoft, Satya Nadella has an engineering background, but it seems like he didn't spend much time as an engineer before getting an MBA and playing the management advancement game.
Our industry isn't what it used to be and I'm not sure it ever could.
linguae
I feel a major shift happened in the 2010s. The tech industry became less about making the world a better place through technology, and more about how to best leverage power to make as much money as possible, making a world a better place be damned.
This also came at a time when tech went from being considered a nerdy obsession to tech being a prestigious career choice much like how law and medicine are viewed.
Tech went from being a sideshow to the main show. The problem is once tech became the main show, this attracts the money- and career-driven rather than the ones passionate about technology. It’s bad enough working with mercenary coworkers, but when mercenaries become managers and executives, they are now the boss, and if the passionate don’t meet their bosses’ expectations, they are fired.
I left the industry and I am now a tenure-track community college professor, though I do research during my winter and summer breaks. I think there are still niches where a deep love for computing without being overly concerned about “stock line go up” metrics can still lead to good products and sustainable, if small, businesses.
jack_tripper
>The tech industry became less about making the world a better place through technology
When the hell was even that?
Normal_gaussian
It's great to hear from someone who thinks these people still care! It has rarely been my experience, but I haven't been everywhere yet.
scottjenson
I've given dozens of talks, but this one seems to have struck a chord, as it's my most popular video in quite a while. It's got over 14k views in less than a day.
I'm excited so many people are interested in desktop UX!
NetOpWibby
Fantastic talk, I found myself nodding in agreement a lot. In my research on next-generation desktop interfaces, I was referred to Ink & Switch as well and man, I sure wish they were hiring. I missed out on the Xerox and Bell Labs eras. I'm also reading this book, "Inventing the Future" by John Buck that details early Apple (there's no reason the Jonathan Computer wouldn't sell like hotcakes today, IMHO).
In my downtime I'm working on my future computing concept[1]. The direction I'm going for the UI is context awareness and the desktop being more of an endless canvas. I need to flesh out my ideas into code one of these days.
P.S. Just learned we're on the same Mastodon server, that's dope.
---
calmbonsai
I concur though per my earlier post I do feel "desktop stagnation" is inevitable and we're already there. You were channeling Don Norman https://jnd.org/ in the best of ways.
az09mugen
Thanks for that nice talk, it felt like a breeze of fresh air with basic & simple yet powerful but alas "forgotten" concepts of UX.
Will look into your other talks.
null
pjmlp
It was quite interesting.
analogpixel
Why didn't Star Trek ever tackle the big issues, like them constantly updating the LCARS interface every few episodes to make it better, or having Geordi La Forge re-writing the warp core controllers in Rust?
thaumaturgy
Because, something that a lot of tech-obsessed Trek fans never seem to really come to terms with, is that Trek didn't fetishize technology.
In the Trek universe, LCARS wasn't getting continuous UI updates because they would have advanced, culturally, to a point where they recognized that continuous UI updates are frustrating for users. They would have invested the time and research effort required to better understand the right kind of interface for the given devices, and then... just built that. And, sure, it probably would get updates from time to time, but nothing like the way we do things now.
Because the way we do things now is immature. It's driven often by individual developers' needs to leave their fingerprints on something, to be able to say, "this project is now MY project", to be able to use it as a portfolio item that helps them get a bigger paycheck in the future.
Likewise, Geordi was regularly shown to be making constant improvements to the ship's systems. If I remember right, some of his designs were picked up by Starfleet and integrated into other ships. He took risks, too, like experimental propulsion upgrades. But, each time, it was an upgrade in service of better meeting some present or future mission objective. Geordi might have rewritten some software modules in whatever counted as a "language" in that universe at some point, but if he had done so, he would have done extensive testing and tried very hard to do it in a way that wouldn't've disrupted ship operations, and he would only do so if it gained some kind of improvement that directly impacted the success or safety of the whole ship.
Really cool technology is a key component of the Trek universe, but Trek isn't about technology. It's about people. Technology is just a thing that's in the background, and, sometimes, becomes a part of the story -- when it impacts some people in the story.
cons0le
>Because the way we do things now is immature. It's driven often by individual developers' needs to leave their fingerprints on something, to be able to say, "this project is now MY project", to be able to use it as a portfolio item that helps them get a bigger paycheck in the future.
AKA resume-driven development. I personally know several people working on LLM products, where in private they admit they think LLMs are scams
amelius
I still wonder why not everybody was lingering in the holodeck all the time.
(equivalent of people being glued to their smartphones today)
(Related) This is one explanation for the Fermi paradox: Alien species may isolate themselves in virtual worlds
d3Xt3r
Most likely because this was a star ship (or space station) with a limited number of personnel, all of whom have fixed duties that need to be done. You simply can't afford to waste your time away in holodecks.
The people we saw on screen most of the time also held important positions on the ship (especially the bridge, or engineering) and you can't expect them to just waste significant chunks of time.
Also, don't forget that these people actually like their jobs. They got there because they sincerely wanted to, out of personal interest and drive, and not because of societal pressures like in our present world. They already figured out universal basic income and are living in an advanced self-sufficient society, so they don't even need a job to earn money or live a decent life - these people are doing their jobs because of their pure, raw passion for that field.
RedNifre
The lack of capitalism meant that the holodeck program authors had no need to optimize their programs for user retention to show them more ads. So much fewer people suffer from holodeck addiction in Star Trek than are glued to their screens in our world.
jfengel
Most of Trek's tech is just a way to move the story along. Transporters were introduced to avoid having to land a shuttle. Warp drive is just a way to get to the next story. Communicators relay plot points.
Stories which focus on them as technology are nearly always boring. "Oh no the transporter broke... Yay we fixed it".
null
dragonwriter
> In the Trek universe, LCARS wasn't getting continuous UI updates
In the Trek universe, LCARS was continuously generating UI updates for each user, because AI coding had reached the point that it no longer needs specific direction, and it responds autonomously to needs the system itself identifies.
bena
LCARS was technically a self-adapting system that was personalized to a degree per user. So it was continuously updating itself. But in a way to reduce user frustration.
Now, this is really because LCARS is "Stage Direction: Riker hits some buttons and stuff happens".
Mistletoe
Isn't it probably just that they don't really have money in Star Trek so there is no contract promising amazing advances in the LCARS if we just pay this person or company to revamp it? If someone has money to be made from something they will always want to convince you the new thing is what you need.
krapp
Remember that in Star Trek humans have evolved beyond the desire to work for money or personal gain, so everyone just volunteers their time, and somehow this just always works.
krapp
>In the Trek universe, LCARS wasn't getting continuous UI updates because they would have advanced, culturally, to a point where they recognized that continuous UI updates are frustrating for users.
Not to be "that guy" but LCARS wasn't getting continuous UI updates because that would have cost the production team money and for TNG at least would have often required rebuilding physical sets. It does get updated between series because as part of setting the design language for that series.
And Geordi was shown constantly making improvements to the ship's systems because he had to be shown "doing engineer stuff."
rzerowan
Mostly i believe its that the writers envisioned and were able to wrldbuildinsucha way that the tech was not a subject but was rather a part of the scenery/background with the main object being the people and their relationships. Additionally in some cases where alien tech was interfaced with the characters inthe storysome UI/code rewites were written in, for example in DS9 where the Cardassian interfaces/AI are frustrating to Chief O'Brien and his efforts to remedy/upgrade such gets a recurring role in the story.
Conversly recent versions have taken the view of foregrounding tech aidied with flashy CGI to handwave through a lot.Basically using it as a plot device when the writing is weak.
RedNifre
Because the LCARS GUI is only for simple recurring tasks, so it's easy to find an optimal interface.
Complex tasks are done vibe coding style, like La Forge vibe video editing a recording to find an alien: https://www.youtube.com/watch?v=4Faiu360W7Q
I do wonder if conversational interfaces will put an end to our GUI churn eventually...
JuniperMesos
Man, I should hope that the warp core controllers on the USS Enterprise were not written in C.
On the other hand, if the writers of Star Trek The Next Generation were writing the show now, rather than 35-40 years ago - and therefore had a more expansive understanding of computer technology and were writing for an audience that could be relied upon to understand computers better than was actually the case - maybe there would've been more episodes involving dealing with the details of Future Sci-Fi Computer Systems in ways a programmer today might find recognizable.
Heck, maybe this is in fact the case for the recently-written episodes of Star Trek coming out in the past few years (that seem to be much less popular than TNG, probably because the entire media environment around broadcast television has changed drastically since TNG was made). Someone who writes for television today is more likely to have had the experience of taking a Python class in middle school than anyone writing for television decades ago (before Python existed), and maybe something of that experience might make it into an episode of television sci-fi.
As an additional point, my recollection is that the LCARS interface did in fact look slightly different over time - in early TNG seasons it was more orange-y, and in later seasons/Voyager/the TNG movies it generally had more of a purple tinge. Maybe we can attribute this in-universe to a Federation-wide UX redesign (imagine throwing in a scene where Barclay and La Forge are walking down a corridor having a friendly argument about whether the new redesign is better or worse immediately before a Red Alert that starts the main plot of the episode!). From a television production standpoint, we can attribute this to things like "the set designers were actually trying to suggest the passage of time and technology changing in the context of the show", or "the set designers wanted to have fun making a new thing" or "over the period of time that the 80s/90s incarnations of Star Trek were being made, television VFX technology itself was advancing rapidly and people wanted to try out new things that were not previously possible" - all of which have implications for real-world technology as well as fake television sci-fi technology.
Findecanor
I have often thought that Star Trek is supposed to show a future in which computer technology and user interfaces have evolved to a steady state that don't need to change that much, and which is superior to our own in ways that we don't yet understand. And because it hasn't been invented yet, the show does not invent it either.
It is for the audience to imagine that those printed transparencies back-lit with light bulbs behind coloured gel are the most intuitive, easy to use, precise user interfaces that the actors pretend that they are.
calmbonsai
Trek needs to visibly "sci-fi-up" extant tech in order to have the poetic narrative license to tell its present-day parables.
Things just need to "look futuristic". The don't actually need to have practical function outside whatever narrative constraints are imposed in order to provide pace and tension to the story.
I forget who said it first, but "Warp is really the speed of plot".
AndrewKemendo
Because it’s a fantasy space opera show that has nothing to do with reality
christophilus
No, we’re not. Niri + Dank Material Shell is a different and mostly excellent approach.
xnx
I felt rage baited when he crossed out Jakob Nielsen and promoted Ed Zitron (https://youtu.be/1fZTOjd_bOQt=1852). Bad AI is not good UI, but objecting based on AI being "not ethically trained" and "burning the planet" aren't great reasons.
GaryBluto
https://www.youtube.com/watch?v=1fZTOjd_bOQ&t=1852s You're missing the ampersand.
It's really strange how he spins off on this mini-rant about AI ethics towards the end. I clicked on a video about UI design.
xnx
Same. AI is absolutely the future of human computer interaction (exactly the article from Jakob Nielsen that he crossed out). Even the father of WIMP, Douglas Engelbart, thought it was flawed: ""Here's the language they're proposing: You point to something and grunt". AI finally gives us the chance to instruct computers as humans.
whatever1
Desktop is dead. Gamers will move to consoles and Valve-like platforms. Rest of productivity is done on a single window browser anyway. Llms will accelerate this
Coders are the only ones who still should be interested in desktop UX, but even in that segment many just need a terminal window.
linguae
Is it dead because people don’t want the desktop, or is it dead because Big Tech won’t invest in the desktop beyond what’s necessary for their business?
Whether intentional or not, it seems like the trend is increasingly locked-down devices running locked-down software, and I’m also disturbed by the prospect of Big Tech gobbling up hardware (see the RAM shortage, for example), making it unaffordable for regular people, and then renting this hardware back to us in the form of cloud services.
It’s disturbing and I wish we could stop this.
xnx
Desktop is all about collaboration and interaction with other apps. The ideal of every contemporary SaaS is that you can never download your "files" so you stay locked in.
snovv_crash
For content consumption sure.
For content creation though, desktop still rules.
immibis
Sounds like a dead market. Nobody needs to create content any more now that we have AI.
sprash
It's not dead. It's being murdered. Microsoft, Apple, Gnome and KDE are making the experience worse with each update. Productive work becomes a chore. And the last thing we need is more experiments. We need more performance, responsiveness, consistency and less latency. Everything got worse on all 4 points for every desktop environment despite hardware getting faster by several orders of magnitude.
This also means that I heavily disagree with one of the points of the presenter. We should not use the next gen hardware to develop for the future Desktop. This is the most nonsensical thing I heard all day. We need to focus on the basics.
kvemkon
> Gnome
I can't imagine what I'd be doing without MATE (GNOME 2 fork ported to GTK+ 3).
Recently I've stumbled upon:
> I suspect that distro maintainers may feel we've lost too many team members so are going with an older known quantity. [1]
This sounds disturbing.
[1] https://github.com/mate-desktop/caja/issues/1863#issuecommen...
silisili
I agree with this. I remember when Gnome 3 came out, there were a lot of legitimate complaints that were handwaved away by the developers as "doesn't work well on a mobile interface", despite Gnome having approximately zero install cases onto anything mobile. AFAICT that probably hasn't changed, all these years later.
hollerith
>productivity is done on a single window browser anyway
When I need to get productive, sometimes I disable the browser to stop myself from wasting time on the web.
whatever1
And you likely open the browser that happens to be called VS Code, Figma etc
hollerith
The point though is that my vscode window does not have an address bar I can use to visit Youtube or Pornhub at any time.
I guess the larger point is that you need a desktop to run vscode or Figma, so the desktop is not dead.
joelkesler
Great talk about the future of desktop user-interfaces.
“…Scott Jenson gives examples of how focusing on UX -- instead of UI -- frees us to think bigger. This is especially true for the desktop, where the user experience has so much potential to grow well beyond its current interaction models. The desktop UX is certainly not dead, and this talk suggests some future directions we could take.”
“Scott Jenson has been a leader in UX design and strategic planning for over 35 years. He was the first member of Apple’s Human Interface group in the late '80s, and has since held key roles at several major tech companies. He served as Director of Product Design for Symbian in London, managed Mobile UX design at Google, and was Creative Director at frog design in San Francisco. He returned to Google to do UX research for Android and is now a UX strategist in the open-source community for Mastodon and Home Assistant.”
mattkevan
Really interesting. Going to have to watch in detail.
I’m in the process of designing an os interface that tries to move beyond the current desktop metaphor or the mobile grid of apps.
Instead it’s going to use ‘frames’ of content that are acted on by capabilities that provide functionality. Very much inspired by Newton OS, HyperCard and the early, pre-Web thinking around hypermedia.
A newton-like content soup combined with a persistent LLM intelligence layer, RAG and knowledge graphs could provide a powerful way to create, connect and manage content that breaks out of the standard document model.
sprash
Unpopular take: Windows 95 was the peak of Desktop UX.
GUI elements were easily distinguishable from content and there was 100% consistency down to the last little detail (e.g. right click always gave you a meaningful context menu). The innovations after that are tiny in comparison and more opinionated (things like macos making the taskbar obsolete with the introduction of Exposé).
fragmede
Heh, the number of points you've probably gotten for that comment, I don't think that it's that unpopular. Win 98 was my jam but it looks hella dated today, but as you said, buttons were clearly marked, but also menus were navigatible via keyboard, soms support for themes and custom coloring, UIs were designable via a GUI builder in VB or Visual Studio using MFC which was very resource friendly compared to using Electron today. Because smartphones and tablets, but even the wide variety of screen sizes also didn't exist so it was a simpler time. I can't believe how much of a step back Electron is for UI creation compared to MFC, but that wasn't cross-platform and usually elements were absolute positioned instead of the relative resizable layout that's required today.
migueldeicaza
Scrubbed the talk, saw “M$” in a slide, flipped the bozo bit
DonHopkins
Golan Levin quotes Joy Mountford in his "TED Talk, 2009: Art that looks back at you":
>A lot of my work is about trying to get away from this. This a photograph of the desktop of a student of mine. And when I say desktop, I don't just mean the actual desk where his mouse has worn away the surface of the desk. If you look carefully, you can even see a hint of the Apple menu, up here in the upper left, where the virtual world has literally punched through to the physical. So this is, as Joy Mountford once said, "The mouse is probably the narrowest straw you could try to suck all of human expression through." (Laughter)
https://flong.com/archive/texts/lectures/lecture_ted_09/inde...
https://en.wikipedia.org/wiki/Golan_Levin
I understand the desire to want to fix user pain points. There are plenty to choose from. I think the problem is that most of the UI changes don't seem to fix any particular issue I have. They are just different, and when some changes do create even more problems there's never any configuration to disable them. You're trying to create a perfect, coherent system for everyone absent the ability to configure it to our liking. He even mentioned how unpopular making things configurable is in the UI community.
A perfect pain point example was mentioned in the video: Text selection on mobile is trash. But each app seems to have different solutions, even from the same developer. Google Messages doesn't allow any text selection of content below an entire message. Some other apps have opted in to a 'smart' text select which when you select text will guess and randomly group select adjacent words. And lastly, some apps will only ever select a single word when you double tap which seemed to be the standard on mobile for a long time. All of this is inconsistent and often I'll want to do something like look up a word and realize oh I can't select the word at all (G message), or the system 'smartly' selected 4 words instead, or that it did what I want and actually just picked one word. Each application designer decided they wanted to make their own change and made the whole system fragmented and worse overall.