Tag Archives: automation

I went back to Uni. What I learned on my Human-AI Interaction course.

The bit that mattered was never the screen.

Take this example. You’re looking to buy a used car. You shortlist three cars and start nudging the finance up and down. Nothing dramatic, just enough to see what an additional £30 a month buys you. A slightly newer plate perhaps. Fewer miles. A bit more punch, or practicality.

Then you click through to a retailer. Your shortlist disappears. Your modelled finance examples disappear. You’re on a blank form, being asked questions your behaviour has already answered. Nothing failed in any obvious way. The system simply dropped the ball precisely at the point it needed to hang on to it.

That is still how a great deal of digital experience works.

You can now generate a respectable interface in seconds, in Figma Make or hell even Gemini. It’ll be clean, structured, and broadly acceptable. Enough hierarchy to feel coherent, enough restraint to avoid embarrassment. It looks like design, which is why people are getting carried away and, frankly, why people are being laid off.

There is too a growing belief that interfaces are becoming fluid. Instead of designing fixed screens, we define components and let agents assemble the right interface on demand. Ask for something, get a form. No navigation, no structure, just translate intent directly into interaction.

Concepts like this are gaining column inches, conference talks and LinkedIn views because they’re brief and intelligible. However, as soon as the interaction extends beyond that moment, the weaknesses become apparent. Context isn’t carried forward properly. Earlier decisions aren’t respected. The system produces something plausible but slightly off, and the user ends up repairing it, effectively doing the work we’d hoped to have eliminated. The surface has improved. The underlying behaviour hasn’t.

Over the past eight weeks, I’ve been working through a Human–Computer Interaction and AI course with the University of Cambridge (Advance Online), and it has been mildly uncomfortable in the way good things often are. It stripped things back to a version of UX that feels almost unfashionable: define the problem before you touch a solution; model the system in terms of what it actually does, not how it looks; decide where control sits between human and machine and how that moves about; draw a boundary and accept responsibility for what happens inside it; then prove the thing works rather than just assuming it does.

As I worked through the modules, it was charmingly familiar and obvious that none of this was new. It just hasn’t been particularly visible for a while. The industry has been busy polishing UI surfaces and calling it progress. AI hasn’t changed that per se, but has made the gap harder to ignore. You can now generate something that looks finished without having done any of the thinking that would make it hold together.

That’s also why a lot of current AI design work feels slightly misplaced, and why the obsession with UI product designers irritates me. There are people doing very good work on interfaces and component systems. Zander Whitehurst is one of them, and they look excellent. But that work sits downstream of where the real difficulty now is. You can refine the surface as much as you like; if the system doesn’t carry intent, it won’t survive contact with actual use.

The “democratisation of design” line doesn’t stand up to the faintest scrutiny. Sure, more people can produce interfaces, and there’s been a lifting of all boats in terms of aesthetics. That’s true. Almost none of those interfaces are grounded in any understanding of what the user is actually trying to do. They meet a baseline. They look competent. They solve very little.

What has been democratised is production.

The discipline that underpins it, the work of defining what matters, what to surface, what to hide, what to carry forward, and where the system should stop and ask, has not been automated. It has been skipped.

The shift to intent-based interaction makes that gap more obvious. You are no longer stepping through a process one action at a time. You state an outcome, and the system attempts to get you there. That changes the shape of the problem. The system has to interpret intent, apply constraints, decide what to do next, and show enough of its reasoning that you can tell whether it has misunderstood you. When it gets that wrong, it doesn’t look like a broken interface. It looks like a reasonable answer that doesn’t quite fit, which is both harder to spot and harder to recover from.

This is where the work is moving.

If interfaces can be assembled on demand, the value shifts into what sits behind them: how intent is captured, how memory is handled, what the system is allowed to assume, when it must ask, and how it behaves when it reaches the edge of its understanding.

Friction becomes part of that. For routine actions, speed is fine. For anything with consequence, removing every pause produces something that feels smooth and behaves carelessly. A system that never slows you down also never asks you to think.

There is a second effect that is easier to miss. If the system does more of the execution, the user does less of the thinking that used to go with it. Over time, that changes behaviour. Trade-offs become less visible. Assumptions go unchallenged. Outputs are accepted because they look plausible..

You end up with users who are comfortable approving things they don’t fully understand.

The risk here isn’t replacement. It’s avoidance. If the problem isn’t defined, if the constraints aren’t understood, if the system isn’t designed to carry intent across time, then the speed of output doesn’t help. It just produces more surface, faster, over the same unresolved issues.

That is already visible. The used car journey doesn’t fail because the interface is ugly. It fails because no one took responsibility for the whole. The same pattern is now being reproduced with better tools and more convincing output.

If that continues, the work doesn’t disappear.

It just gets left to the user.

AI: I used ChatGPT to tidy up some grammar and I used Gemini to pick some holes in the piece to strengthen the arguments.

Tagged , , , , , , , , , , , , , , , , , , ,

Time Returned, Time Resold

Rain-blurred motorway at dusk viewed through a windscreen; dashboard lights glow amber in an empty driver’s seat — a quiet image of autonomy and time unclaimed.
Autonomy promised freedom. Instead, it gave us metrics.

Every few years a new invention turns up promising to give us time back. The dishwasher did it, then the calendar app, now the self-driving car. Efficiency, they say, is liberation. But the minutes never come home. They’re quietly re-employed: answering messages that weren’t urgent until we saw them, scrolling through news we already half-read. We don’t get more time. It just comes back wearing a different outfit.

Design now speaks the intoxicating language of generosity. We’ll save you clicks. We’ll make it seamless. Lovely words, but they come with a tempo you didn’t choose. The system nudges, reminds, congratulates you on your streak. Even the oven chirps when it’s done pre-heating. Helpful, yes – in the way a personal trainer is helpful when all you wanted was a walk.

Efficiency was meant to hush the world, not make it chatter. Parcels update you mid-journey, cars suggest faster routes, TV apps interrupt the credits to make sure you don’t go off to bed just yet. You start to feel managed by your own apps and appliances. Is it me, or do they all sound slightly pleased with themselves?

Still, there is a deeper promise in all this autonomy. Because the best thing about a self-driving car isn’t speed, it’s permission. The choice to drive when you want to: for rhythm, for presence, for drivers like me who relish the satisfaction of line and camber, and to switch off when you don’t. The long crawl north to the Lakes. The dawn blast to the airport. The late-night, rain-spray-soaked slog home when you’d gladly hand over the wheel and let the motorway unspool while you exhale, watch the window-light flicker, maybe half-doze through an episode of something forgettable. Control should be optional, not constant.

That’s what the technology could be about: selective surrender or a quieter freedom. But for some unfathomable reason, the marketing and product design departments have decided autonomy is best packaged as constant optimisation. That means another dashboard app full of metrics and prompts and juanty reminders. We built cars clever enough to drive themselves, then gave them personalities that never stop talking.

Real luxury now isn’t speed but discretion: the right to decide how long something should take. To drive when you feel like driving. To look out of the window when you don’t. Technology can make both possible.

Convenience promised to return our hours, but mostly it’s taught us to account for them. Every minute feels spoken for. Perhaps the odd thing is how willingly we’ve agreed to it and the peculiar pleasure we take in shaving seconds off tasks we didn’t enjoy anyway.

Maybe the best thing a self-driving car could do is forget the ETA and let us forget, too.

AI: This piece was refined with AI, for the image prompt, tags, excerpt, and a little sub-editing. The ideas, references, and rhythm are mine.

Tagged , , , , , , , , , , , , , ,