Back to Blog

Why AI Coding Tools Still Feel Stuck on Localhost

7 min read Updated April 11, 2026

I like the models. I do not like the box they come in. Modern AI work is mostly CLI-driven, remote, and collaborative, but most AI tools still assume one laptop, one repo, and one small chat window.

Published April 11, 2026

I have spent real time with Codex, Claude, and other chat-based AI tools. They are useful. They can write code, explain errors, summarize diffs, and speed up local development.

The disappointment comes later, when you try to use them for serious engineering. Modern AI usage is not mostly about chatting in a browser. It is about terminals, logs, remote hosts, Kubernetes APIs, deployment state, and long-running sessions. Most tools still stop at localhost.

Browser AI workspace with split panes, terminal, and assistant context panel.
A stronger AI surface already looks less like chat and more like an operating console.

The model is not the problem. The surface is.

I am not disappointed by model quality. I am disappointed by product design. Most AI tools still feel like chat apps with a terminal attached, or terminals with a chat box glued on top. That is fine for a small script on your laptop. It is not enough for real operations.

What serious AI work needs What most chat-first tools offer
Multi-pane workflow with persistent context One main chat view and maybe one terminal tab
First-class remote execution over ssh Mostly local shell access or manual workarounds
Kubernetes-aware browsing, logs, and object state Paste kubectl output back into chat
Real session history and replay Prompt transcript without operator context
Team sharing, recording, and broadcast modes Personal workspace only

CLI is where the real work happens

For engineers, platform teams, and operators, the command line is still the center of gravity. That is where you run tests, tail logs, diff configs, inspect processes, restart services, and verify what changed. AI is now part of that flow, but the interface has not caught up.

This is a normal workflow. Yet most AI products still behave as if the real action is the prompt box. It is not. The prompt box is just one control in a much larger operating surface.

Animated demo of a browser workspace moving through shell, SSH, and Kubernetes-aware views without losing context.
A short loop of the workflow the article is arguing for: shell-first, remote-capable, and context-preserving.

Localhost is only the first five minutes

Current tools are mostly good for localhost development. That is useful, but it is only the beginning. The moment you need to connect to a remote machine, inspect a container, follow a rollout, or debug a cluster issue, the experience falls apart.

I want browser-native remote connection over ssh. I want direct Kubernetes API access in the same workspace, not a copy-paste loop. I want to see logs, pod status, events, files, and commands side by side. I want a tmux-style split screen inside the browser because that is how this work is actually done.

A real file explorer changes the workflow

One feature that would immediately make a tool feel more serious is a real file explorer. Not a tiny repo sidebar. A proper explorer that works across local files, remote hosts over ssh, and Kubernetes-backed runtime paths in the same workspace.

This matters because engineers do not work only from prompts and terminals. They jump between source trees, config files, manifests, logs, generated outputs, and temporary artifacts. A strong file explorer makes the tool feel operational, not cosmetic. It also makes the product unique because most AI tools still treat files as passive context instead of an active surface for navigation, preview, and action.

File explorer view with folders, filters, and file actions inside the browser workspace.
A file explorer becomes a differentiator when it is part of the working surface, not just a sidebar.

A built-in editor should finish the loop

The file explorer is only half the story. If the tool can browse remote machines, it should also edit remote files directly in the same browser workspace. Opening a config, changing it, reviewing the diff, and asking AI to explain, refactor, or fix a block should not require jumping out to another app.

This is where advanced AI features actually matter. Inline edits, patch suggestions, config validation, command-aware fixes, and context from the current host or cluster turn the editor into an operating surface, not just a text box. That is much more useful than copying file contents into chat and hoping nothing gets lost on the way back.

Built-in browser editor for a remote file, with sensitive values redacted.
A remote editor becomes much stronger when AI actions live next to the file instead of outside it.

Notes need to work like a notebook, not a chat log

Another missing piece is notes. Most AI tools still treat notes as a long thread of messages. I want a real notebook surface made of typed cells: a place where prompts, drafts, terminal output, research frames, and idea blocks can live as separate working units instead of one endless conversation.

That matters because different tasks have different contracts. A writer cell should rewrite or expand text. A terminal cell should preserve commands and outputs. An idea cell should force a thesis, risks, and validation steps. Once notes are structured this way, the workspace becomes much more useful for real thinking, execution, and review.

Typed note cells for prompt drafting and debate framing inside the browser workspace.
Notes get stronger when the document is split into typed working cells instead of one chat thread.
Idea cell showing a structured startup thesis with controls, validation, and snapshot history.
A good note surface should turn generation into something structured, reviewable, and reusable.

A transcript is not history

Another missing piece is recording. A chat transcript is not enough. It tells you what the assistant said, not what the engineer ran, what changed on screen, or how the session evolved. Serious AI tooling should record command history, screen state, context switches, and replayable timelines.

History view showing saved command sessions, timestamps, and terminal playback context.
Session history becomes useful when it captures commands, context, and replay, not just prompt text.

Profiles should be products, not prompt hacks

People also work in different modes. Sometimes I want a strict reviewer. Sometimes I want a fast debugger. Sometimes I want an infra operator that assumes Kubernetes, logs, and rollout safety. Today this is usually handled with prompt tricks and manual setup. That is weak.

A serious tool should have a real profile wizard for different AI personas: coding assistant, incident responder, release engineer, cluster troubleshooter, documentation editor. Each profile should carry the right tools, permissions, defaults, layout, and guardrails.

Spotlight should be a central feature

One feature I keep wanting is a real Spotlight-style command surface. Not just a search box, but one place to jump between sessions, hosts, docs, actions, tunnels, and layouts without hunting through sidebars.

In a tool like this, Spotlight is not decoration. It is the control plane. It should let me move fast by keyboard, trigger common workflows, switch context instantly, and expose power features that are otherwise hidden. When the product gets bigger, this becomes one of the most important ways to keep it usable.

Spotlight-style command palette for sessions, hosts, actions, docs, and layout presets.
A good Spotlight makes the whole workspace feel faster because navigation and action live in one place.

Broadcast and multi-screen are not edge features

AI work is often collaborative even when one person is typing. I want broadcasting over multiple screens, a wall mode for incident response, and read-only follow mode for teammates. The assistant view, terminal view, and cluster view should be shareable independently.

Collaboration overlay with targets, templates, recent actions, and shared command entry.
Collaboration needs its own surface, not just a chat transcript and a screen share.

What I actually want

The wishlist is not complicated. It is just closer to how engineers already work.

  • Tmux-style split screens inside the browser.
  • An integrated file explorer across local, ssh, and cluster contexts.
  • A built-in remote editor with inline AI actions for real files, not pasted snippets.
  • First-class remote connections over ssh.
  • Direct Kubernetes API views for pods, logs, events, and rollout state.
  • Persistent sessions that survive reloads and reconnect cleanly.
  • A notebook-style notes surface with typed cells for writing, execution, and idea work.
  • Command history, screen recording, and replayable session timelines.
  • A profile wizard for different AI personas and operator roles.
  • A real Spotlight for actions, sessions, hosts, docs, and layout switching.
  • Broadcast mode for second screens, war rooms, and review sessions.
  • Better audit trails around who did what, where, and why.

Final take

I am not disappointed because Codex, Claude, or other AI systems are useless. They are not. I am disappointed because the surrounding products are still too small for the job. They are good at helping one person code on one machine. They are still weak at helping a team operate real systems.

The next step for AI is not only better benchmarks. It is better runtime surfaces: browser-based, split-pane, remote-capable, history-aware, and built for real CLI work. Until that happens, most of these tools will stay impressive and incomplete at the same time.

The gap is no longer model intelligence. The gap is whether the interface is built for actual engineering work instead of demo-friendly chat.

See the rest of the ktl site at https://kubekattle.github.io/ktl/.

I’m open to job opportunities. If you’d like to talk, email me at antonvkrylov@gmail.com.