Back to Research

Building Anthropic Artifacts for Claude Desktop: An MCP Integration Experiment

The Promise and the Problem

Yesterday, Anthropic announced the ability to create custom Artifacts for Claude Desktop - self-contained apps that live alongside your conversations. We immediately jumped on this to test a complex use case: building a visual workflow orchestrator for our video generation MCP server.

First look at Anthropic Artifacts in Claude Desktop

First look at the Interactive Video Generation Canvas in Claude Desktop

The goal was ambitious yet straightforward: create an app that visualizes the multi-step video creation process, allowing users to see, modify, and control each stage of generation. When a user asks Claude to "create a 15-second video with three scenes about Alaska, including drone footage over a glacier," the app would display the entire workflow as an interactive pipeline.

What We Built

In just five minutes, Claude generated a surprisingly good interface:

  • Visual workflow representation: Create Project → Generate Script → Plan Scenes → Generate Images → Generate Videos → Download → Assemble
  • Interactive step execution: Each node could be triggered individually or run as a complete flow
  • Status tracking: Real-time updates showing which steps were running, completed, or failed
  • Queue management: Parallel execution of image and video generation tasks

The interface looked like what we envisioned - a clean, left-to-right flow with distinct stages and clear visual feedback.

Interactive Video Generation Workflow interface

The complete video generation workflow with all steps from Create Project to Assemble Video

The MCP Wall

But here's where we hit the wall: Artifacts cannot directly access MCP tools.

When the Artifact tried to call our video generation MCP server, we got: "I don't have access to create project function or any video creation tool. I am a text-based AI assistant."

The issue? Artifacts use window.claude.complete() to communicate with Claude, which spawns a sub-agent. These sub-agents don't inherit the parent's MCP connections. It's Claude-in-Claude, but without the tools.

MCP access error in Anthropic Artifacts

The error message when Artifacts try to access MCP tools - sub-agents don't inherit parent's tool connections

The Workaround

Claude suggested a manual workaround:

  1. The Artifact generates commands as text that includes tools calling
  2. User copies it to the main Claude conversation
  3. User executes it and copies back the response
  4. The Artifact continues with the next step

While functional, this breaks the seamless experience. What should have been a single click became a copy-paste dance between windows.

Split Screen Vindication

Despite the MCP limitation, we're thrilled to see Anthropic implementing the split-screen pattern we described in our AI Interface Patterns research. The ability to maintain conversation on the left while generating task-specific interfaces on the right is exactly the future we envisioned.

This experiment directly relates to our three-layer framework of Software 3.0. To understand what's happening here, let's recall the layers:

The Three Layers of Software 3.0

  • Top Layer: Constructive Thinking - World knowledge, principles, and instructions
  • Middle Layer: Communication - Where planning happens through Human-to-Agent (H2A) interaction
  • Ground Layer: Execution - Where plans become reality through tool orchestration

Anthropic Artifacts are fundamentally exploring the Communication Layer - the middle layer where human intent transforms into actionable plans. The split-screen interface is a concrete implementation of H2A communication, where:

  • The left panel maintains the conversational context (traditional chat interface)
  • The right panel dynamically generates task-specific interfaces (the Artifact)
  • Together, they create a richer communication channel than text alone

This is precisely what we mean by multimodal negotiation in the Communication Layer - the interface itself becomes part of the conversation, dynamically adapting to the task at hand. The Artifact isn't just displaying information; it's creating a shared visual language between human and agent.

This experiment proves the pattern works - the interface dynamically adapted to our video generation workflow without any predetermined templates. It "conversed itself into existence" based on what we were trying to achieve.

Current State & Future Potential

Today's limitations:

  • No MCP access in Artifacts
  • Manual intervention required for external tool calls
  • Sub-agents can't share state or configuration

But the potential is clear. Once Anthropic enables MCP pass-through to sub-agents, we'll have truly autonomous AI applications that can orchestrate complex workflows while maintaining visual clarity and user control.

The Code

For those interested in the technical details, our video generation MCP server is here: https://github.com/h2a-dev/video-gen-mcp-monolithic

Conclusion

Anthropic Artifacts represent a significant step toward the AI-powered app future. While current limitations prevent full MCP integration, the rapid prototyping capability and visual clarity make them invaluable for testing ideas and workflows.

We're particularly excited that the split-screen pattern is becoming reality. As we noted in our research: "The interface generates based on need, not predetermined templates." That's exactly what we experienced - from concept to visual workflow in minutes.

The MCP limitation will likely be addressed soon. When it is, we'll see an explosion of AI applications that combine conversational flexibility with visual sophistication.

Until then, we'll keep experimenting and pushing the boundaries of what's possible.