Teaching Computers (And AI) How To Interact With Humans
Introducing The Human Conversational Model
We’re surrounded by systems that talk at us—but rarely with us. As AI and automation scale across every corner of business, the quality of our conversations (or lack of them) is quietly defining the quality of our relationships—with customers, employees, and even technology itself.
In this week’s feature article, I introduce the Human Conversational Model—a simple but powerful framework for designing systems that build trust through human-like interaction. It breaks down what makes great conversations work, and how we can infuse those elements into digital and AI-driven experiences.
The same theme shows up in my latest podcast episode with workplace strategist and bestselling author Erica Keswin. We explore what it takes to build truly human workplaces—where empathy, rituals, and connection aren’t nice-to-haves, but non-negotiables in a rapidly evolving, tech-driven world.
If we want to lead with trust and scale with integrity, we have to get serious about how we communicate—human to human, and human to machine.
I hope you enjoy this edition. If you do, I’d love for you to subscribe and share it with others who might benefit. Let’s dive in.
The Secret to Humanizing Digital Systems? Conversations.
Swipe, tap, click—modern life has collapsed into frictionless micro-transactions. We order dinner without speaking to a waiter and approve expenses without seeing a coworker. Efficiency is soaring, yet something vital leaks out of these seamless flows: the electric pulse of real conversation.
From the first coo between infant and caregiver, dialogue teaches us to trust, coordinate, and imagine together. Neuroscientists call it neural coupling; anthropologists call it collective sense-making. Whatever the label, the pattern is the same: back-and-forth talk synchronizes brains, lowers cortisol, and releases oxytocin—the chemistry of rapport.
Yet our daily digital exchanges rarely rise to that level. Most brand encounters feel like vending machines—select, wait, accept whatever drops. As AI automates more touchpoints, the drift toward faceless efficiency accelerates, stripping away the give-and-take that turns service into relationship. We finish tasks—but we walk away hungry for connection.
Introducing The Human Conversational Model
How do we restore the spark that vending-machine interactions drain away? We turn to humanity’s oldest technology: conversation. Every thriving tribe, team, and partnership relies on that rhythmic call-and-response to build trust and weave collective meaning.
To design more human systems, we studied what makes real conversations work—breaking down the underlying elements that consistently create connection, momentum, and mutual understanding. What emerged is a sturdy, repeatable architecture that can be grafted onto chatbots, apps, even leadership meetings. We call it The Human Conversational Model.
It is built on seven elements across these two foundational building blocks::
Co-operative Interface: The Visible Components. Cooperative Interface refers to the direct, observable elements of a conversation—the words people use, the timing of their responses, the emotional tone they convey, and the real-time signals that guide the flow of dialogue. In human interactions, it shows up through verbal exchanges, active listening cues like nodding or clarifying questions, and adjustments in tone or pace based on the other person’s reactions. In digital experiences, it takes the form of clearly sequenced steps, confirmation messages, intuitive prompts, and responsive feedback that helps users stay oriented. A strong Cooperative Interface creates a sense of alignment and momentum throughout the interaction. It ensures that participants—whether human or digital—stay connected, understand each other’s intent, and feel confident progressing toward their goals.
Background Mindfulness: The Hidden Foundation. Background Mindfulness reflects the deeper forces that silently guide interactions—your internal compass and your emotional intelligence. It’s how each participant’s self-awareness, sense of values, and commitment to learning shape not just this conversation, but the next one too. In human conversations, Background Mindfulness influences the interaction when a person chooses words carefully to stay true to who they are, notices the emotional ripple effects of what’s been said, and later reflects on how to show up better next time. In digital environments, it’s how a system embodies the organization’s brand values, protects trust by handling data respectfully, and uses experience data to evolve future interactions. Strong Background Mindfulness doesn’t just create one good exchange—it builds a foundation for lasting relationships over time.
Cooperative Interface — The Visible Components
When a Cooperative Interface works well, the dialogue demonstrates these five practices:
Intent Decoding: What do they want? In a good conversation, participants quickly pick up on each other’s underlying goals. A friend noticing your distracted glances might cut short a story, realizing you need to leave soon. Similarly, a digital experience decodes intent when, for example, a banking app surfaces "Bill Pay" automatically at the end of the month based on your usual patterns.
Contextual Framing: What do they already know? Effective communication fits information to the listener’s level of understanding. A teacher explains photosynthesis differently to a child than to a graduate student. Digital environments frame context when a travel site recognizes you’re an experienced user and skips basic onboarding.
Empathetic Agility: How are they feeling? Good conversationalists adjust in real time based on emotional cues. A manager noticing frustration might slow down and show more patience. Similarly, a chatbot that detects agitation through repeated sharp language adapts by offering simpler options or escalation to a human.
Supportive Feedback: Do they feel confident? Progress feedback keeps momentum alive. A friend giving directions points out landmarks; a digital system provides visible progress indicators and status updates, reassuring users they’re on the right path.
Basic Manners: Do they feel respected? Positive conversations are grounded in basic norms: letting others finish, staying on topic, acknowledging contributions. Digital systems show manners by not interrupting workflows, offering easy save-and-exit options, and using clear, respectful language.
Background Mindfulness — The Hidden Foundation
When Background Mindfulness is strong, the conversation is shaped by these two practices:
Self-Awareness: Are we living our values? Strong conversationalists stay anchored in who they are. A purpose-driven CEO reaffirms the company’s mission when pressured to compromise. Likewise, a digital experience shows self-awareness when, for instance, a sustainability-focused retailer avoids promoting fast-fashion products that contradict its commitments.
Emotional Reflection: Are we continuously learning? Skilled communicators reflect after critical interactions. A top salesperson dissects a tough negotiation to refine future approaches. In digital experiences, reflection happens when teams study chat logs, friction points, and emotional signals—and evolve the design accordingly.
Making Digital Interactions More Conversational
Designing digital systems through the Human Conversational Model means treating every user interaction as part of an evolving relationship—not just a task to complete. It’s about building experiences that recognize intent, adapt to emotional signals, reinforce trust, and reflect shared values over time.
Here are three examples of what that looks like in practice:
A mortgage chatbot that builds confidence, not just closes cases. When a user logs in to compare refinance options, the chatbot doesn't just present static loan rates—it adjusts its flow based on user behavior. If the user pauses on jargon-heavy terms like “escrow” or revisits an earlier step, the system slows down, offering a short video or a simple written explanation. That’s Cooperative Interface: sensing confusion, reading pacing, and dynamically adapting in the moment. Behind the scenes, Background Mindfulness shows up in how the chatbot reinforces the company’s brand values of transparency and empowerment—it avoids pressure language, aligns its tone with the human loan advisors, and flags unclear exchanges so the team can fine-tune the experience over time.
A health app that adapts to motivation and emotion. After a user misses three days of workouts, the app doesn't simply nudge with a “You’re falling behind!” alert. Instead, it sends a message that acknowledges the break—“We all have off weeks. Want to reset your plan or just talk about it?” That’s a smart Cooperative Interface: adjusting tone, offering options, and reading emotional cues from behavior. The Background Mindfulness shows up in how the app reflects the brand’s values of well-being and compassion—it suppresses overly competitive messaging, recommends mindfulness exercises when patterns of drop-off emerge, and learns from long-term behavior to keep adjusting support strategies that fit the user’s rhythm and emotional state.
An e-commerce journey that prioritizes partnership over pushiness. A shopper spends several minutes reading product descriptions in the “sustainable goods” section. Instead of prompting with a limited-time discount, the system surfaces a short article: “How We Source Our Materials.” That’s Cooperative Interface—responding in real time to user signals and offering context instead of pressure. The Background Mindfulness is embedded in the brand’s decision not to use dark patterns like countdown timers or surprise fees. The system is trained to uphold trust, not just drive conversion, and feedback from support chats and returns data is used to refine how product recommendations align with shopper values—not just purchase history.
Implications for More Humanistic AI
As AI becomes a primary bridge between organizations and the people they serve, it carries new responsibility—not just to complete tasks, but to build trust. In many ways, AI is stepping into the same space that human conversations have occupied for centuries: helping people feel seen, understood, and supported. Yet most AI systems today are still designed around efficiency, not relational depth. Without a stronger foundation, even technically impressive systems risk feeling mechanical, brittle, and disconnected.
The Human Conversational Model offers a blueprint for building AI that doesn’t just transact, but actually relates. Getting there requires two major shifts:
Strengthening the Cooperative Interface. For AI to behave like a true conversational partner, it must be able to interpret user intent even when inputs are messy, incomplete, or emotionally charged. It should recognize when a user is hesitating, confused, or frustrated—and adapt its pace, tone, and level of support accordingly. That means training models not just on "correct" outputs, but on how conversations evolve moment to moment. Techniques like sophisticated prompting strategies, memory structures that track the flow of interaction, and reinforcement systems that reward relational intelligence—not just task accuracy—will all be essential. The goal isn't just answering faster; it's staying aligned with human rhythms in a way that builds confidence and momentum naturally.
Embedding Background Mindfulness. Strong conversations aren’t just responsive—they're anchored in who we are. AI needs that same grounding. Systems can't invent values on their own; they must be trained with clear organizational principles around trust, ethics, and brand identity. Every decision—what to prioritize, what to recommend, how to handle ambiguity—should reflect those deeper commitments. And just like human communicators improve through reflection, AI systems must be designed to monitor emotional patterns, detect points of relational strain, and evolve through structured retraining—not just optimization for speed.
Sparking New Leadership Thinking
Leaders who want to create more human-centered organizations can take immediate action:
Uncover and define your core values and promises. Before you can expect digital or AI systems to reflect your brand authentically, you need to articulate what your organization stands for. Gather cross-functional leaders and frontline voices to identify your non-negotiable commitments, then pressure-test them: would your AI recommend differently than your people?
Audit a key digital journey for conversational health. Choose one major customer or employee journey and assess it through the lens of a conversation. Where does the system recognize intent? Where does it misread emotion? Where could trust be reinforced?
Set conversational design standards for new projects. Make conversation a formal requirement in every new initiative. Teams should define how systems will sense and respond to goals, adjust to emotional signals, and reinforce user confidence—with these standards codified early, not retrofitted later.
Define conversational integrity as an AI imperative. Make conversational quality a visible, enforceable priority for AI teams—not just a soft aspiration. Set explicit objectives that require systems to behave like skilled conversational partners, build them into project goals and technical reviews, and hold teams accountable for how well experiences support trust, understanding, and brand alignment across human-facing interactions.
Run conversation reviews after major launches. After releasing new digital experiences or AI touchpoints, hold focused debriefs that study interactions through a conversational lens. Instead of only tracking completion rates or bug counts, examine how well the system picked up on user goals, adapted to emotional shifts, and stayed anchored to brand values. Use those insights to drive your next iteration.
The Bottom Line
Conversations aren’t just how we connect—they’re the blueprint for designing systems that build trust at scale. The Human Conversational Model shows how to create experiences that feel responsive in the moment and anchored in deeper values over time. As AI reshapes every interaction, embedding these principles is how organizations will deepen loyalty, strengthen relationships, and lead in a human-centered future.
*Note: We first introduced the Human Conversational Model in a 2017 Temkin Group report, Humanizing Digital Interactions.
Additional Resources
Here’s some relevant content that you may find interesting:
Humanizing Digital Interactions. In this 2017 Temkin Group report which is now part of the XM Institute library, Isabelle Zdatny and I first introduced the Human Conversational Model.
Empathetic AI: Unlocking Trust Between Humans and Machines. In this informative TED Talk, Peter J. Scott discusses how empathy can be integrated into AI systems to foster trust between humans and machines.
Interactivity, Humanness, and Trust: A Psychological Approach to AI. This study by Yi Ding & Muzammil Najaf investigates how interactivity and perceived humanness in AI chatbots influence user trust in e-commerce settings.
🔥 Spark Of The Week
Here’s an idea to make you rethink your leadership:
Trust isn’t an empty feeling, it’s a willingness to take a risk with another party
Your Leadership Challenge: Consider whether you are worth the risk
Your team, your customers, your stakeholders—whether they say it or not, they’re constantly scanning for signals. And every small action sends a message.
Here are three trust signals leaders often overlook:
🔹 Do you act consistently—even when it’s inconvenient?
🔹 Do you explain your decisions when trust might be strained?
🔹 Do you repair trust when it’s been broken—or quietly move on?
Trust isn’t built by accident. And once broken, it’s not easily restored. This week, take one small step to make your trustworthiness more visible.
👉 Read more: It’s Time To (Re)Build Trust
Humanity At Scale: Redefining Leadership Podcast
Available on Apple Podcasts, Spotify, and YouTube.
Make sure to check out my podcast, where I reimagine leadership for today’s dynamic world—proving that true success begins with prioritizing people, including employees, customers, and the communities you serve. From candid conversations with executives to breakthrough insights from experts, Humanity at Scale: Redefining Leadership Podcast is your ultimate guide to leading with purpose and empathy.
Here are some recent episodes:
Empathy, AI, and the New Rules of the Human Workplace with Erica Keswin. In this episode, I sit down with WSJ bestselling author, human workplace expert, and keynote speaker, Erica Keswin and we explore the future of human-centric leadership in a tech-driven world.
Innovate or Stagnate: Mastering Leadership in a Dynamic World with Charlene Li. In this episode, I sit down with Charlene Li, Strategic Advisor and Keynote Speaker, Founder and CEO of Quantum Networks Group, and New York Times bestselling author, to explore how leaders can harness disruption as a catalyst for innovation.
From Conflict to Connection: Rethinking Moral Disagreements with Kurt Gray. In this episode, I speak with psychologist and neuroscientist Kurt Gray, Professor of Psychology and Neuroscience at UNC, Chapel Hill, to explore how leaders can navigate moral complexity and foster empathy in organizations.
Scaling with Heart: How Safelite Balances Growth, People, and Purpose with Renee Cacchillo. In this episode, Renee Cacchillo, CEO of Safelite, shares how the company became the dominant force in its industry by putting people and customers at the center of every decision.
Why Smart Leaders Ask More and Tell Less with Liz Wiseman. In this episode, leadership expert and bestselling author Liz Wiseman shares powerful insights from her groundbreaking work on how leaders can amplify the intelligence of those around them.
The podcast is available on Apple Podcasts, Spotify, and YouTube.
Humanity at Scale is a movement to inspire and empower leaders to create humanity-centric organizations