We used to open apps. Now apps open us. That polite tap on the shoulder from a chatbot asking if it may book your flight, fix your deck, and rearrange your company is the sound of software dissolving into the interface that wins everything: the thread. And this week, the thread got teeth.
A leaked demo shows OpenAI testing a visual Agent Builder that wires actions, tools, files, search, and MCP connectors into real workflows — think Google and Microsoft services showing up as first-class nodes. If that wasn’t enough, the official keynote then rolled out Apps in ChatGPT (a full-blown marketplace inside the chat), Apps SDK, a refreshed Codex, ChatGPT 5 Pro in the API, and Sora 2 for video. In one swoop, the chat window stopped being a clever toy and started looking suspiciously like an operating system with superpowers.
Yes, Google is on the same road. The company’s “Opal” experiment auto-generates and deploys mini web-apps from a single prompt, sitting on top of Gemini for text/code, Imagen for images, SoundStorm for TTS, and Veo for video. Yandex is pushing a similar direction in AI Studio. The pattern is unsubtle: vertical integration. Pick an ecosystem, and you get the models, the compute, the builder, the data storage, the marketplace, the distribution. One bill. One vendor. One throat to choke.
And if you’re a generalist “wiring tool” like Zapier, n8n, Botpress? Press F. Your advantage — being neutral — still matters, especially with MCP bridging the silos. But the gravitational pull of a chat-native OS with an app store is ruthless.
The chat window ate the browser
Apps in ChatGPT flips the model-picker logic on its head. Perplexity lets you choose models; most people don’t. OpenAI’s play is simpler: bring Spotify, Figma, Canva, Booking (with Uber, Tripadvisor, DoorDash and friends on deck) into ChatGPT. You’ll bark a command; the app will execute; results will land back in the same thread. No tabs. No mental context switches. No “copy link, paste token, re-auth.”
For developers, this is “an offer you can’t refuse.” Publish inside the marketplace, inherit an audience the size of a continent, and let the assistant orchestrate the last mile. Stay outside, and a competitor who integrates will eat your lunch inside your users’ favorite conversation.
And then there’s AgentKit. A visual builder for multi-step agents you can inspect node-by-node, tune, and attach to external data. Pair it with ChatKit to drop those agents into your own product. It’s not a toy; it’s a flight deck. You won’t “figure it out over coffee”; you’ll need a course. But once you get it, you’ll stop stitching brittle one-offs and start shipping repeatable, observable, debuggable automations that survive contact with customers.
The unfunny future: agents that learn your company for a month
Here’s the uncomfortable bit. The next non-trivial jump isn’t just nicer node UIs. It’s RL agents that ingest a month of your Slack, email, tickets, docs, and call transcripts, then spit out a JSON flow akin to a mature n8n graph with hundreds of nodes — triage, routing, QA, billing exceptions, the lot. It will arrive wrapped in enterprise controls, and it will quietly replace the first wave of “AI integrators” who sold glue code as strategy. Russian roulette for workflows, but with a safety: you review, you approve, you observe. And you keep the margin.
Which brings me to the part most new builders skip.
Before you drag nodes: do the work
I’ve built more automations than I care to count. The ones that stuck had the same boring secret: we did the human work first.
Phase 1: Research the problem
Watch the job before you wire the job.
- Spend 2–3 days shadowing real people doing the task.
- Write their workflow in plain language.
- Identify the 20% of cases that cause 80% of the daily pain.
- State the ideal outcome in human terms.
Document before you open a builder
- Business problem: “Lyuba spends 45 minutes each morning classifying support emails and forwarding urgent ones.”
- Inputs/Outputs: Input = emails to support@; Output = categorized snippets to team chats + direct alert for urgent.
- Success metrics: Primary = Lyuba’s morning batch drops to ≤10 minutes; Secondary = no urgent ticket sits unacknowledged >30 minutes.
- Bare-bones logic: Fetch → Extract key fields → Classify urgency/category → Route → Log.
If your diagram can’t be explained to a mildly caffeinated intern, your workflow is theatre.
Phase 2: Don’t invent wheels
Go hunting before building.
- n8n community templates (search exact keywords).
- Reddit r/n8n and r/automation (sort by top this month).
- YouTube walkthroughs of near-by workflows.
- #n8n on X for living examples.
Find a template that nails 60–70% of your problem. Copy the structure. Adapt the logic. Only then earn your right to build the boring MVP.
In the new world — Agent Builder, Apps SDK, ChatKit — you’ll still need this discipline. The difference is the scaffolding stops fighting you, and the debugging surface is designed for humans.
“Why does my GPT feel dumber lately?”
Because, bluntly, it sometimes is. GPT Store is open to everyone, but only paid users can publish. Free users hammer those GPTs under hard limits; when they hit the ceiling, the system quietly detunes to a mini model. Same prompt, different brain, flatter answer. That’s half the “it used to be smarter” moaning you hear.
If you’re shipping a public GPT to reach the widest audience, you have two sane options:
- Optimize for the mini from day one. You’ll get predictable quality for free users even after the auto-downgrade.
- Or optimize for GPT-5 (or 5 Pro) and accept split quality: paying users smile; free users flip a coin.
Do your future self a favor and say it in the description: “Optimized for GPT-5; mini answers will be simpler.” It won’t stop the odd one-star review, but it will set expectations like an adult.
Also, words matter. A lot of people call their GPTs “agents.” Most aren’t. Custom GPTs still rely on you to click, copy, and care. Agents go click the web, fill the form, move the file, and return with receipts — under your control, but on their legs. If it can’t act, it’s an assistant, not an agent.
The platform war is vertical — and it’s already here
- OpenAI: chat marketplace, Agent Builder/AgentKit/Apps SDK/ChatKit, Sora 2, 5 Pro API.
- Google: Gemini + Imagen + SoundStorm + Veo + Opal to mint and host prompt-born apps.
- Yandex: AI Studio pulling more of the “last mile” inside.
This is vertical integration as strategy. Compute, models, tools, hosting, identity, distribution — one vendor. It’s tidy. It’s fast. It’s dangerously comfortable.
Neutral platforms like n8n and Zapier still have a trump card: portability. With MCP acting like a universal adapter, there’s a genuine shot at staying un-captured. But the default gravity pulls you into someone’s chat OS. If you care about optionality, design for it explicitly: keep your logic in portable graphs, demand MCP-style connectors, and escrow the boring bits — parsers, formatters, retries — outside anyone’s walled garden.
What to build now (and how to not get replaced by your own agent)
- Pick a real business pain, not a novelty. “Reply to LinkedIn comments” is cute; “close L2 tickets 30 minutes faster” is revenue.
- Prototype in the ecosystem your buyer already lives in. If their day is ChatGPT, use Agent Builder and ship a thread-native experience. If they’re deep in Google land, test Opal for disposable internal tools and push your logic to Gemini actions.
- Instrument everything. Node-level telemetry, retries, fallbacks, and human-in-the-loop for edge cases. Agents fail like people — quietly, and at the worst time.
- Design for “mini” first, then scale up. If your logic collapses on a smaller model, the problem isn’t the model.
- Own the brief, not the wiring. Your competitive moat isn’t the node-graph; it’s the domain insight, the data access, the KPI you moved.
If you follow that playbook and your agent later learns your org for a month and proposes a better flow? Congratulations. You’ve been promoted — from the person dragging nodes to the person setting the targets.
The browser is the new legacy app
In the demo I watched, a single ChatGPT thread juggled Coursera (watching a course), Canva (building a deck), and a property site (comparing listings). No tabs, no yak shaving, no “paste your API key here.” It looked less like “yet another chatbot” and more like an early cut of a conversation-first OS.
Will there be a device? Probably. If OpenAI ships something voice-centric before end of year, the question won’t be “is the mic good”; it will be “does the hand-off from device to cloud feel instant.” If it does, the browser becomes the thing you use to download the chat app you actually live in.
And yes, if you’re an n8n diehard (I am), this isn’t a eulogy — it’s a pivot. Use n8n for what it’s great at: neutral, auditable, portable automations you can host and version. Then bolt your flows into the ecosystems your users already occupy via MCP or actions. Be Switzerland, but learn the local languages.
TL;DR (that isn’t actually short)
- Chat is the OS. The interface that owns the thread will own the work.
- Vertical beats horizontal. Ecosystems offering models + builder + marketplace will win mindshare by default.
- Portability still matters. MCP and neutral graphs keep you from being captured.
- Do the human work. Research, measure, template, then build.
- Ship for mini, wow with Pro. Reach first, glory second.
- Call things what they are. Assistants answer; agents act.
If you’re building in this moment, it’s the best time in history to create — and the worst time to be precious about tools. The node editor is not your identity. The KPI is.
