Psychological Safety and the AI Question
Érica November 18, 2025

Psychological Safety and the AI Question

13 min read

Amy Edmondson first described psychological safety in 1999, studying hospital nursing teams. She expected that teams with better interpersonal dynamics would make fewer medication errors. Instead, she found the opposite: teams with high psychological safety reported more errors. Not because they made more mistakes — because they were willing to surface them.

The finding reframed safety. Psychological safety is not the absence of mistakes. It is the belief that you won’t be punished for surfacing mistakes, asking questions, or admitting what you don’t know. High-performing teams don’t avoid errors. They catch errors faster because speaking up is safe.

Twenty-six years later, this framework collides with AI adoption in a way Edmondson did not anticipate — and that most companies deploying AI tools have not considered.

The AI Question Problem

Every AI tool works by answering questions. You ask, it responds. The interface is a question-and-answer loop, whether the tool is a chatbot, a search engine, a recommendation system, or a decision-support tool. The fundamental interaction is: the human asks, the machine answers.

Now consider what asking reveals.

When a procurement officer types “What does incoterm DDP mean?” into the company’s AI assistant, the question contains an admission: I don’t know what DDP means. In a private interaction — alone at a desk, nobody watching — this admission costs nothing. The machine does not judge. The machine does not remember that you didn’t know. The machine does not tell your manager.

But AI tools are increasingly deployed in shared environments. The query history is logged. The screen is visible to colleagues. The team lead asks, “How are you using the new tool?” and the answer reveals the questions you asked — which reveals what you didn’t know.

In a psychologically safe environment, this is fine. “I didn’t know what DDP meant, so I asked the tool — now I do.” Competence is demonstrated through learning, not through pretending.

In a psychologically unsafe environment — where admitting ignorance carries career risk, where “not knowing” is conflated with “not being qualified,” where knowledge is currency and spending it feels like depletion — asking the AI tool becomes a hazard. The tool is not the problem. The room is the problem. The organisational culture determines whether the tool is a resource or a threat.

Edmondson’s Framework Applied

Edmondson defined four dimensions of psychological safety behaviour:

Speaking up with concerns. In AI adoption, this translates to: Can a team member say, “This AI tool gave me an answer I think is wrong” without being dismissed as resistant to change? Can they flag an error in the system without being told they “just need more training”?

Asking questions. Can a team member ask how the tool works, what its limitations are, or why it gave a particular answer — without the question being interpreted as technophobia? Can a junior employee ask a question that a senior employee answered incorrectly, without social penalty?

Admitting mistakes. Can a team member say, “I used the tool’s recommendation and it was wrong” without being blamed for trusting the tool? Can they say, “I didn’t use the tool because I didn’t trust its output” without being blamed for not adopting?

Offering ideas. Can a team member suggest a better way to use the tool, modify the workflow, or adjust the integration — without the suggestion being perceived as a criticism of the person who designed the deployment?

In most organisations, the answer to at least two of these questions is no. The deployment proceeds anyway. The team interacts with the tool in ways that minimise self-exposure: they ask simple questions, they don’t flag errors, they don’t experiment, they don’t suggest improvements. They use the tool enough to avoid being marked as non-adopters. They don’t use it enough to derive real value.

This is the adoption plateau — the flat line between “deployed” and “embedded” where most AI tools live and die.

The Visibility Asymmetry

There is an asymmetry in AI tool interaction that amplifies the psychological safety problem.

When a team member asks a colleague a question — “What does DDP mean?” — the interaction is bilateral. The colleague knows you didn’t know. Nobody else does. The social cost is contained.

When a team member asks the AI tool the same question, the interaction is logged. It is visible to system administrators. It may be visible in usage dashboards. It is visible to anyone who walks past the screen. The interaction that was bilateral is now potentially multilateral.

This visibility asymmetry changes the calculus. The cost of asking the machine is higher than the cost of asking a colleague — not because the machine judges, but because the asking is more visible. And in environments where visibility equals vulnerability, increased visibility equals increased risk.

I have watched teams develop workarounds specifically to avoid visible AI interactions. Using the tool on a personal device. Typing questions into the tool and then clearing the history. Asking a colleague to ask the tool on their behalf. These are not irrational behaviours. They are rational responses to an environment where visible learning is penalised.

The workarounds reduce adoption metrics. Management interprets low adoption as evidence that the tool isn’t useful. The tool is discontinued or deprioritised. The actual cause — the environment, not the tool — is never addressed.

The Manager’s Paradox

The irony is that managers who champion AI adoption are often the ones who inadvertently undermine the psychological safety required for it.

The manager who says, “This tool is intuitive — you should be able to figure it out in an afternoon” has established an implicit benchmark: competence with this tool should be immediate. Anyone who struggles is not meeting the benchmark. The statement, intended to be encouraging, is received as a performance standard.

The manager who monitors adoption dashboards and asks team members why their usage is low has established another dynamic: tool usage is observed. Non-usage will be noticed. The observation is not neutral — it carries the weight of managerial attention, which in most organisations is associated with evaluation. The monitoring, intended to support adoption, is received as surveillance.

The manager who showcases the tool’s most impressive capabilities in the demo — the complex query, the clever analysis, the impressive output — has set expectations at the ceiling. The team’s first real interaction will be a simple query with a simple answer. The gap between the demo and the reality creates disappointment. The tool that seemed magical in the demo is merely functional in practice. This is not a failure — functionality is what matters. But the gap registers as a let-down.

In each case, the manager’s actions are well-intentioned. In each case, the effect is a reduction in psychological safety around the tool. The intention is “this tool will help you.” The reception is “this tool is another thing I’ll be evaluated on.”

The Question Behind the Question

Daniel Kahneman’s work on cognitive ease and cognitive strain adds a layer. Kahneman showed that when people encounter information that is easy to process — familiar, clearly presented, consistent with expectations — they experience cognitive ease. They feel confident. They are less likely to engage in deliberate, critical thinking.

When people encounter information that is difficult to process — unfamiliar, complex, inconsistent with expectations — they experience cognitive strain. They feel uncertain. They are more likely to engage in deliberate thinking, but they are also more likely to feel anxious, uncomfortable, and avoidant.

An AI tool is a source of cognitive strain. It is new. Its outputs are unpredictable. Its logic is opaque. The interface may be familiar (a chat window), but the interaction pattern (speaking to a machine that understands language) is deeply unfamiliar. The cognitive strain is inherent to the novelty.

In a psychologically safe environment, cognitive strain is tolerable. Uncertainty is allowed. Questions are welcome. The strain resolves into learning.

In a psychologically unsafe environment, cognitive strain is intolerable. Uncertainty must be hidden. Questions reveal weakness. The strain resolves into avoidance.

The question the team member asks the tool is the surface question. The question underneath is: “Is it safe for me to not know this?” The answer to the surface question comes from the AI model. The answer to the underlying question comes from the room.

Measuring Psychological Safety for AI Adoption

Edmondson developed a seven-item survey to measure psychological safety. For AI adoption contexts, I suggest adapting five of the items:

  1. “If I make a mistake using the AI tool, it will be held against me.” (Reverse-scored.)
  2. “Members of this team are able to bring up problems and tough issues about the AI tool.”
  3. “People on this team sometimes reject others for asking the AI tool basic questions.” (Reverse-scored.)
  4. “It is safe to take a risk with the AI tool on this team — such as trying a new use case.”
  5. “It is easy to ask other members of this team for help with the AI tool.”

Administer this survey before the AI deployment, two weeks after, and six weeks after. The trajectory tells you more about adoption outcomes than any usage dashboard.

If the scores decrease after deployment — if the introduction of the AI tool reduced psychological safety — the tool is not the problem. The deployment architecture is. And the fix is not more training. The fix is the environment.

Building Safety Before Building Capability

The practical sequence matters. Most deployments follow: build the tool, deploy the tool, train the team, measure adoption. The psychological safety architecture is absent from this sequence.

The alternative sequence: assess the team’s psychological safety, address the gaps, deploy the tool, support the adoption, measure both usage and safety.

Before deployment: Conduct the adapted Edmondson survey. If scores are below the threshold (Edmondson suggests a team average of 5.0 on a 7-point scale as a minimum for effective team learning), address the safety gaps first. This may mean having explicit conversations about the learning expectations — specifically, that the learning curve is normal, that questions are valued, and that the team’s query history is not a performance evaluation tool.

During deployment: Make three explicit commitments, in writing, communicated to the team and their managers. First: query history is private. No manager will review individual query logs. Second: the learning period is defined (two to four weeks) and during this period, reduced productivity is expected and protected. Third: error reporting is rewarded. If the tool gives a wrong answer and you flag it, that’s a contribution — not a complaint.

After deployment: Monitor both adoption metrics and safety metrics. If adoption rises and safety holds, the deployment is healthy. If adoption rises and safety drops, adoption is compliance-driven and will not sustain. If adoption drops and safety holds, the tool may need improvement. If both drop, the deployment is failing and the root cause is environmental.

The Team That Named the Tool

I want to return to the naming signal — the observation that when a team gives their AI tool a name, adoption has crossed a critical threshold.

The naming is a psychological safety indicator. To name the tool is to claim a relationship with it. It is to say, publicly, to colleagues: “I use this tool. I know it well enough to name it. My use of it is part of my professional identity, not a secret.” Naming requires the safety to be associated with the tool publicly.

In psychologically unsafe environments, tools are not named. They are referred to generically — “the system,” “the new thing,” “that AI tool.” The anonymity of the reference is a distance mechanism. The team member is not rejecting the tool. They are protecting themselves from being too closely associated with it, in case the association becomes a liability.

When I see a team that names its tool, I know the environment is safe enough for sustained adoption. When I see a team that keeps the tool at arm’s length linguistically, I know the environment needs work — regardless of what the usage metrics say.

The Remote Work Complication

There is a dimension of psychological safety and AI adoption that has become more relevant since 2020: the remote work context.

In a physical office, the social dynamics of AI tool usage are partially visible. You can see who uses the tool. You can see the screen. You can overhear the query. The visibility is a double-edged sword: in safe environments, it creates social proof (“she’s using it, I should try it”); in unsafe environments, it creates surveillance (“he’s asking basic questions, he must not know the domain”).

In a remote work environment, the visibility is mediated by digital tools. Screen sharing, activity dashboards, Slack messages — all create selective visibility. The user controls what is visible and what is hidden. This control can amplify psychological safety (“I can use the tool privately, nobody sees my learning curve”) or undermine it (“the usage dashboard shows I’ve asked 47 questions this week, management can see that”).

The hybrid environment — which is the reality for most EU companies — creates a third dynamic. The in-office team sees each other’s AI usage. The remote team does not. The psychological safety conditions differ between the two groups, even within the same team. The in-office members develop shared practices and social proof. The remote members do not.

The design implication: for hybrid and remote teams, AI tool adoption strategies must explicitly address the visibility asymmetry. Make tool usage visible by choice, not by default. Share tool successes (good outputs, time saved) through team channels — voluntarily, not mandatorily. Create peer learning opportunities where tool usage is explicitly social and exploratory, not evaluative.

The remote work context has made psychological safety both more important and harder to cultivate. The AI tool deployment must account for this — not as a special case, but as the standard operating environment of the modern European workplace.

The Integration

Psychological safety and AI adoption are not separate conversations. They are the same conversation, viewed from different angles.

The AI industry focuses on the tool: capabilities, accuracy, speed, integration. The organisational psychology field focuses on the environment: trust, safety, belonging, autonomy. Both are right. Neither is sufficient.

A perfect tool in an unsafe environment will not be adopted. An imperfect tool in a safe environment will be adopted, improved, and eventually named. The environment is not a modifier of the tool’s success. It is a precondition.

Edmondson’s work gives us the framework. Kahneman gives us the cognitive mechanism. Karasek gives us the demand-control structure. Together, they explain why the most capable AI tool on the market can sit unused on a team’s desktop while a mediocre workaround thrives — because the workaround doesn’t require anyone to reveal what they don’t know.

The machine is not the problem. The room is the problem. Fix the room, and the machine works.

The room is the architecture. The architecture is the set of conditions — trust, safety, control, social proof — that determine whether a tool is adopted or abandoned. The conditions are designable, measurable, and improvable. They are not mysterious. They are not soft. They are the infrastructure of adoption, and like all infrastructure, they must be built before the thing they support.

Written by
Érica
Organizational Psychologist

She knows why people resist tools — and how to design tools they’ll love. When Érica speaks, companies change direction. Not from persuasion. From understanding.

← All notes