Direct Manipulation vs. Interface Agents: Revisiting the Classic HCI Debate Decades Later
More than two decades have come and gone, yet the timeless debate in human-computer interaction remains unresolved. Let’s take a fresh look at the concepts of direct manipulation and interface agents—exploring how they continue to resonate today and what design principles might guide the development of modern AI features.

In the spring of 1997, two distinguished scholars in the field of human-computer interaction (HCI), researchers Ben Shneiderman of University of Maryland and Pattie Maes of MIT engaged in a historic debate on user interface design at the CHI conference, the premier forum for human-computer interaction research. The debate helped define the two dominant themes in human-computer interaction: direct manipulation versus interface agents. Shneiderman championed user control via direct manipulation, while Maes advocated intelligent agents to assist users.
Shneiderman advocated for sophisticated visualizations, arguing that users desire control and predictability in their systems. In contrast, Maes contended that as computing environments grew increasingly complex, relying solely on user interface buttons would prove insufficient for users to confidently complete many tasks. She proposed implementing software agents as assistants to automate these complex tasks, thereby reducing the cognitive burden on users who would otherwise need to process overwhelming amounts of information. Their debate remains highly influential, with the original CHI proceedings receiving 584 citations by March 2025. This impact is especially notable within user interface design communities.
Should your computer listen to you, or make decisions for you?
Is there truly a definitive answer when choosing between these two approaches? Now, more than twenty-five years later, as we witness the emergence and proliferation of various AI agents—particularly those powered by advanced Large Language Models (LLMs) such as OpenAI’s Operator, Anthropic’s Computer Use, or Google's AgentSpace—does this indicate that the interface agent approach is beginning to dominate? Furthermore, how can we effectively determine the optimal balance between these distinct interaction paradigms?
What Is Direct Manipulation?
Direct Manipulation, often associated with the concept of What You See Is What You Get (WYSIWYG), represents an interactive design approach where users can manipulate digital objects directly and immediately observe the results of their actions. A classic example is drag and drop functionality. On a computer, users can use a mouse to move a folder or drag a file from one directory to another—mirroring how one might relocate a physical document between folders in a tangible filing system. This concept formed the foundation of the graphical user interface (GUI), which has profoundly transformed design and human-computer interaction. Direct manipulation metaphors utilize familiar physical objects, such as a desk, to represent and facilitate user interactions with computational systems.

This concept brought a massive shift to UI in the late 1970s, when Xerox's Palo Alto Research Center (PARC) pioneered the first modern GUI (e.g., the Alto computer), later inspiring the development of modern personal computers, including Apple’s Lisa and Macintosh.
From this, it becomes clear that direct manipulation offers several key benefits:
- Intuitiveness: Minimal learning curve, enabling users to quickly understand and effectively interact with the interface.
- Enhanced Sense of Control: Users experience a direct connection to their tasks and maintain a clear sense of authority over interactions.
- Reduced Error Rates: Interaction clarity and immediacy result in fewer user errors.
- Visual Engagement: Emphasizes visual interaction, providing users with a clearer, more immediate understanding of system functionality.
What Is Interface Agent?
In contrast to direct manipulation, the interface agent approach functions like having an exceptional assistant who maintains comprehensive knowledge of your information organization. With a simple command, this assistant can execute tasks on your behalf—whether moving a file between folders or handling more complex operations. This approach requires substantial knowledge infrastructure, which is why many companies are now developing verticalized agents—specialized assistants with expertise in specific domains. These purpose-built agents excel at particular tasks such as writing code, reviewing literature, or addressing travel-related inquiries. By automating these processes, verticalized agents reduce the burden of routine tasks through simple instruction-based interactions.

In the enterprise digital landscape, we encounter countless tasks nestled within environments lacking programmatic APIs for precise control. Even when such interfaces offer API access, they fundamentally embody human cognitive patterns—intuitive information hierarchies and organizational structures that programmatic approaches struggle to fully embrace. Many digital workflows demand multi-step sequences that become particularly challenging when scaled. Though an agent might not always outpace human hands in execution speed, it excels in shouldering those monotonous, repetitive responsibilities that humans naturally resist. Consider the illustration below: agents navigate interfaces by orchestrating carefully sequenced actions—a choreography of specific commands and actions such as navigating Flights.com, typing in arrival and destination, clicking search, and finally being able to search for flights from Boston to Washington.

Pienso’s Design Philosophy: Harmonizing AI with Human Experience
This scenario showcases an agent’s convenience, but it also highlights concerns raised by skeptics. As HCI pioneer Ben Shneiderman cautioned during the debate:
I think the intelligent agent notion limits the imagination of the designer, and it avoids dealing with interface issues. That’s my view of the agent literature—there is insufficient attention to the interface.
While it might seem tempting to rely on a super agent to handle all tasks, it raises important questions: How do users recover from errors if they can’t see the process behind the agent? How do we build trust so users understand what the agent is doing and know when to step in? At Pienso, we grapple with these questions daily to guide our design of AI features. We are a design-first company. Pienso’s no-code platform empowers domain experts (even without ML training) to build and refine AI models, always keeping the human-in-the-loop for critical decisions. For example, our PromptFactory offers a comprehensive suite of ready-to-use templates with customizable parameters, enabling users to efficiently compare summaries and modify prompts through an intuitive interface. Similarly, our NLQ enables users to ask a question through extensive document sets, accurately pinpointing the exact source of each insight. This ensures that responses can be directly referenced and verified by human oversight, maintaining both transparency and reliability in the decision-making process.
Finding a Balance: The Hybrid Future
As multimodal interaction gains momentum, these two approaches will likely share the stage with emerging paradigms. Focusing on the interface level, the future appears to be in hybrid interfaces that offer flexibility and contextual adaptability.
Consider a design tool as an example: designers could directly manipulate objects through dragging and positioning while simultaneously benefiting from interface agents that provide contextual recommendations—suggesting established UX conventions and design best practices. In medical settings, physicians might directly interact with patient records while interface agents automatically identify potential medication conflicts or concerning patterns.
This hybrid approach combines the intuitive control of direct manipulation with the proactive assistance of intelligent agents, creating interfaces that adapt to various contexts and user needs.
However, this integration introduces a significant challenge for developers and designers: establishing user trust in AI-generated suggestions, particularly in high-stakes domains like healthcare and finance. Users may question when to follow AI recommendations versus relying on their own judgment, especially when the consequences of decisions are significant. Building interfaces that clearly communicate the reasoning behind AI suggestions, establish appropriate confidence levels, and maintain user agency will be crucial for the successful implementation of these hybrid systems.
Key Principles for Designing AI Feature
At Pienso, one of our key insights is that an exceptional agent without thoughtful UI design is destined for the POC (point-of-production) graveyard. Currently, there's a disproportionate focus on AI/ML model capabilities while design considerations receive insufficient attention. This imbalance leads to user frustration as people don't feel confidently in the "driver's seat" during interactions. From a design perspective, without delving into excessive technical details, we maintain several guiding principles when integrating direct manipulation with AI agents:
- Transparency: Ensure users clearly understand system operations, especially with interface agents.
- Human-in-the-loop components: Allow users to override or modify system behavior at any point to prevent feelings of powerlessness.
- Progressive guidance: Introduce intelligent features gradually as users become familiar with the system, starting with simple direct manipulation before advancing to more sophisticated capabilities.
What's Next?
Of course, the intersection of these approaches isn't as straightforward as it might initially seem. Beneath the surface lies a rich tapestry of nuances and contextual considerations waiting to be unraveled. We'll explore these fascinating complexities in depth at our upcoming MIND Workshop.
We're thrilled to be organizing the MIND (Mixed-Initiative Next-gen Design) Workshop at the Intelligent User Interface (IUI) conference on March 24th 2025, exploring innovative ways to combine agent-based approaches with direct manipulation interfaces—two paradigms often viewed as separate but with incredible potential when unified.
Building upon our successful summit at MIT Media Lab last year, this workshop brings together a strong industry presence to brainstorm and prototype the next generation of AI interfaces. By bringing researchers and practitioners from HCI, IUI, and AI communities, we aim to establish a collaborative network focused on developing a deeper understanding of next-generation interactive AI systems. Join us and share your insights as we shape the future of human-AI collaboration.