Cursor, vibe coding, and Manus: the UX revolution that AI needs
uxdesign.cc
Its time to move past the command-line era ofAI.Photo by Brett Jordan onUnsplashWe keep hearing that AI is for everyone. No coding required. Your new coworker. An assistant for the masses. And yet, to get consistently useful results from todays best models, you still need to know the secret handshakes: the right phrases, the magic tags, the unspoken etiquette ofLLMs.The power is there. The intelligence is real. But the interface? Still stuck in thepast.And that disconnect became painfully clear one day, when a friend of mine tried to do something absurdlysimpleWhen ChatGPT needed a magicspellMy friendlets call him Johndecided to use ChatGPT to count the names in a big chunk of text. He pasted the text into ChatGPT and asked for a simple count. GPT-4o responded with a confident-sounding number that was laughably off. Confusion gave way to frustration as John tried new phrasing. Another round of nonsense numbers. By the fifth attempt, he looked ready to fling his keyboard out thewindow.I ambled over and suggested a trick: wrap the block of text in <text> tags, just as if we were feeding the model some neat snippet of XML. John pressed Enter and, with that tiny tweak, ChatGPT nailed the correct count. Same intelligence, but now it was cooperating.John was both relieved and annoyed. Why did a pair of angle brackets suddenly turn ChatGPT into a model of precision? The short answer: prompt engineering. That term might sound like a fancy discipline, but in practice, it can look like rummaging around for cryptic phrases, tags, or formatting hacks that will coax an LLM into giving the right answer. It is reminiscent of a fantasy novel where you have infinite magical power, yet must utter each syllable of the summoning chant or risk conjuring the wrongmonster.https://medium.com/media/54f0195b29bc89292f7db656514a657a/hrefIn an era when we supposedly have next-level AI, its hilarious that we still rely on cryptic prompts to ensure good results. If these models were truly intuitive, they would parse our intentions without requiring special incantations. Instead, were forced to memorize half-hidden hints, like travelers collecting obscure tourist phrases before visiting a foreign land. That is a design problem, not an intelligence problem.Sound familiar?If this rings a bell, its because weve seen a similar dynamic before in the early days of personal computing. As usability expert Jakob Nielsen explains, the very first user-interface paradigm was batch processing, where you had to submit an entire batch of instructions upfront (often as punched cards) and wait hours or even days for the results. This eventually gave way to command-based interfaces (like DOS or Unix) in the 1960s, where youd type in a single command, see an immediate result, and then type another command. Back then, command lines ruled the UI landscape, demanding meticulous keystrokes that only power users could navigate.Then the graphical user interface arrived, with clickable icons and drag-and-drop metaphors that pulled computing out of the realm of black-screen mysticism and into the daylight. You no longer had to type COPY A: C:\DOC.DAT; you just dragged a file into a folder. That wasnt a minor convenienceit was a revolution that let everyday people, not just specialists, feel in control of the machine for the firsttime.Early IBM PC with a command line like interface from WikipediaWe saw the same leap again and again. WYSIWYG editors let people format documents without memorizing tags or markup. Mosaic put the internet behind glass and made it something your mom could use. The smartphone touchscreen transformed phones from keypad-laden gadgets into intuitive portals you could navigate with a swipe and a tap. In every case, the breakthrough wasnt just better tech under the hoodit was better ways to interact withit.Mosaic browser gave a GUI to the Internet, from Web DirectionsThats what made ChatGPT the spark that set off the LLM explosion. Sure, the underlying tech was impressive, but GPT-style models had been kicking around for years. What ChatGPT nailed was the interface. A simple chatbox. Something anyone could type into. No setup. No API key. No notebook to clone. It didnt just show people that LLMs could be powerfulit made that power feel accessible. That was the GUI moment for large languagemodels.ChatGPT was a GUI revolution, yet its inaccessibility is remnisicent of command lineUIYet, as many people (Nielsen included) have noted, were still in the early days. Despite ChatGPTs accessibility, it still expects precise prompts and overly careful phrasing. Most people are left relying on Reddit threads, YouTube hacks, or half-remembered prompt formats to get consistent results. Were still stuck in a command line moment, wearing wizard hats just to ask a machine forhelp.The tech is here. The intelligence is here. Whats missing is the interface leap, the one that takes LLMs from impressive to indispensable. Once we design a way to speak to these models that feels natural, forgiving, and fluid, well stop talking about prompt engineering and start talking about AI as a true collaborator. Thats when the real revolution begins.Examples of emerging UX inAICursor = AI moves into the workspaceMost people still interact with LLMs the same way they did in 2022: open a chat window, write a prompt, copy the answer, paste it back into whatever tool they were actually using. It works, but its awkward. The AI sits off to the side like a consultant you have to brief every five minutes. Every interaction starts from zero. Every result requires translation.Cursor flips that model inside out. Instead of forcing you to visit the AI, it brings the AI into your workspace. Cursor is a code editor, an IDE, or integrated development environment, the digital workbench where developers write, test, and fix code. But unlike traditional editors, Cursor was built from the ground up to work hand-in-hand with an AI assistant. You can highlight a piece of code and say, Make this run faster, or Explain what this function does, and the model responds in place. No switching tabs. No copy-paste gymnastics.Demo of Cursors Changelog where codebase and AI assistant canco-existThe magic isnt in the model. Cursor uses off-the-shelf LLMs under the hood. The real innovation is how it lets humans talk to the machine: directly, intuitively, with no rituals or spellbooks. Cursor doesnt ask users to understand the quirks of the LLM. It absorbs the context of your project and adapts to your workflow, not the other wayaround.Its a textbook example of AI becoming more powerful not by getting smarter, but by becoming easier to work with. This is what a UX breakthrough looks like: intelligence that feels embedded, responsive, and natural, not something you summon from a command line with just the right phrasing.So what happens when we go one step further, and make the interaction evenlooser?Vibe coding = casual commandlineVibe coding, a term popularized by Andrej Karpathy, takes that seamless interaction and loosens it even more. Forget structured prompts. Forget formal requests. Vibe coding is all about giving the LLM rough direction in natural language, Fix this bug, Add a field for nickname, Reduce the sidebar padding, and trusting it to figure out thedetails.By Jay Thakur on HackerNoonAt its best, its fluid and exhilarating. LLMs like a talented engineers who have read your codebase and can respond to shorthand with useful action. That sense of flowof staying in the creative zone without stopping to translate your thoughts into machine-friendly commandsis powerful. It lets you focus on what you want to build, not how to phrase theask.But heres the catch: you still have to know how to talk to themachine.Vibe coding isnt magic. Its an interaction style built on top of a chat interface. If you dont already understand how LLMs behave, what theyre good at, where they stumble, how to steer them with subtle rewordings, the magic breaks down. Youre still typing into a text box, hoping your intent comes through. Its just that now, the prompts are written in lowercase and goodvibes.https://medium.com/media/06036727978af6a2ca6b2de416bbfd0f/hrefSo while vibe coding lowers the friction for seasoned users, it still relies on unspoken rules and learned behavior. The interface is friendlier, but its not yet accessible. It shows that AI can feel conversationalbut still speak a language only some people understand.If Cursor showed us what it looks like to embed AI into tools we already use, and vibe coding loosened the interaction into something more human, then whats next? What happens when you dont even need to steer, when AI takes thewheel?Manus = the interface is the intelligenceWhen Manus hit the scene, it made waves. Headlines hailed it as the first true AI agent. Twitter lit up. Demos showed the tool writing code, running it, fixing bugs, and trying again, all without needing constant human input. But heres the quiet truth: Manus runs on Sonnet 3.7, a Claude model you can already access elsewhere. The model wasntnew.What was new, and genuinely exciting, was the interface and interactions.Instead of prompting line by line, Manus lets users speak in goals: Analyze this dataset and generate a chart, or Build a basic login page. Then it takes over. It writes the code. It runs the code. If something breaks, it investigates. It doesnt ask you to prompt again. It acts like it knows what youmeant.Demo of Manus creating icons that match TechCrunch websiteThis is delegation as design. Youre no longer babysitting the model. Youre handing off intent and expecting results. Thats not about intelligence. Thats about trust in the interface. You dont need to wrangle it like a chatbot or vibe with it like a coder who knows the right lingo. You just ask and Manus handles thehow.And thats the point. The Manus moment wasnt about a smarter model. It was about making an existing model feel smarter through better interaction. Its leap forward wasnt technical it was experiential. It didnt beat other tools by out-reasoning them. It beat them by understanding the humanbetter.This is the future of AI: tools that dont just process our input, but interface with ourintent.All three examplesCursor, vibe coding, and Manusprove the same point: its not the size of the model that changes the game, its the shape of the interaction. The moment AI starts understanding people, instead of people learning how to talk to AI, everything shifts.The missing link between power andpeopleIf we return to Johns struggle with <text> tags, its clear his predicament wasnt just a glitch. It revealed how these advanced models still force us to speak their language rather than meeting us halfway. Even though powerful LLMs can write code and pass exams, many AI interactions still feel like sending Morse code to a starship. The real shortfall isnt intelligence; its an outdated, command-line-like user experience that demands ritual rather than collaboration.Tools like Cursor, Manus, and vibe coding show glimpses of a new reality. By embedding AI directly into our workflows and allowing more natural, goal-oriented conversations, they move us closer to what Jakob Nielsen calls a complete reversal of control, where people state what they want and let the computer figure out how to get there. Weve watched technology make leaps like this before, from command lines to GUIs, from styluses to touchscreens, and each time, the real revolution was about removing friction between people and possibility.AI will be no different. The next major step isnt just bigger models or higher accuracy: its designing interfaces that make AI feel like a partner rather than a tool. When that happens, well stop talking about prompts and simply use AI the way we use search bars or touchscreens. Intuitively and effortlessly.This shift is not only the future of AI, but also the future of computing itself.Cursor, vibe coding, and Manus: the UX revolution that AI needs was originally published in UX Collective on Medium, where people are continuing the conversation by highlighting and responding to this story.
0 Comments ·0 Shares ·15 Views