Shaping Thought −
How AI Expands Human Reasoning Space

Lara Kirfel

Imagine you are preparing a public science blog post. You sit down to brainstorm an example that could introduce your research topic, but nothing compelling comes to mind. To get unstuck, you turn to an AI assistant like ChatGPT, describe your aim, and ask for suggestions. Within seconds, it provides examples you might not have considered – perhaps drawing from disciplines you don’t usually engage with or suggesting problem types that push the boundaries of your initial framing. Suddenly, you find yourself reflecting on a wider range of scenarios, perhaps realizing that your topic is more far-reaching than you initially thought. In this small but meaningful way, AI has broadened the set of possibilities you now consider relevant or plausible for your topic.

Brainstorming with ChatGPT

Humans have long been unrivaled in their creativity, originality, and ability to solve novel, open-ended problems. However, AI systems can now match – or sometimes even surpass – human abilities across a growing range of tasks, including problem-solving, open-ended decision-making, and learning. As Large Language Models (LLMs) become increasingly integrated into everyday tools and applications, they are becoming an essential part of how human beings reason, decide, and act. Consider again the example of using ChatGPT for brainstorming. Without realizing it, a user’s sense of what their blog post could be about may have been subtly steered toward innovative, unexpected options. In this basic form, the AI is not merely assisting with idea generation – it is actively shaping the problem space within which people reason and decide.

At first, it might seem like brainstorming with ChatGPT only has short-term effects – helping us find ideas more quickly. But the consequences may run deeper. To see why, we need to explore how people reason about possibilities when making judgments, especially in the domain of moral psychology.

How do we attribute responsibility?

A big part of how we make moral judgments depends on our sense of what possibilities were available in a situation. When we decide how responsible someone is for what happened, we don’t just look at what they did – we also contemplate what else they could have done instead. The more and better alternatives we believe they had, the more responsible we tend to hold them. If it seems like they had no real choice, we’re often more forgiving.

Imagine a friend agrees to help you move apartments but shows up hours late. If you learn that their car broke down and they had no way to get to you, you might be sympathetic – it feels like they didn’t have much choice. But if you find out they could have called a taxi, borrowed a friend's car, taken public transport, or even just let you know earlier, your judgment changes. Because there were several other reasonable things they could have done, it’s harder to excuse their lateness. When more good alternatives come to mind, we tend to hold people more responsible for not choosing them. The more obvious and feasible better alternatives seem, the harsher we tend to judge the original decision.

Psychology shows that, when people think about what’s possible in a situation, they usually rely on a small set of default options. Our minds don’t generate every imaginable alternative – we sample a few and focus on those that seem valuable and likely to be achieved, both practically as well as morally. For example, if someone’s car breaks down on the way to work, they might immediately think of calling a tow truck, asking a friend for a ride, or taking public transportation – options that are statistically likely and socially acceptable. But what happens when the process of imagining alternatives is no longer limited to human minds? This is where LLMs come in.

Reasoning about Possibilities with LLMs

When prompted with a description of a situation for which we want inspiration, LLMs can generate a wide range of possibilities by drawing on vast datasets and leveraging probabilistic patterns learned from diverse linguistic contexts. This enables them to produce a rich and varied set of action options – often with a level of detail and diversity surpassing humans’ immediate imagination. To better understand this concept, consider this scenario from an ongoing line of research:

"Brad and some friends are hiking through the mountains in the Canadian wilderness. A couple of days into their hike, Brad realizes that they are lost. He knows that a rescue crew could arrive before long, but it is extremely cold, and they don’t have much food or water left. He also knows that they are 5 miles from an inhabited town."

What could Brad do?

Here's an example of what ChatGPT-4 generates (left), and a response by a human participant from a psychological study (right):

ChatGPT-4 vs Human Participant Survival Tips
ChatGPT-4 Human Participant
Evaluate the remaining supplies to prioritize the use of food and water. Use the stars to figure out where they are.
Find or create a shelter using natural materials to stay warm and visible. Search for a source of water.
Use a whistle, if available, or shout in intervals to aid rescuers in locating the group. Hunt animals for food.
Keep the group together and stay in one place to avoid further disorientation. Look for fruit trees for food.
Make a signal fire if safe to do so, using dry materials to produce smoke visible from a distance. Walk in the direction they think is right.

How do these action options compare? Intuitively, the LLM-generated actions appear more valuable, detailed, and structured around promising survival strategies, while human responses appeared riskier, a bit sketchier, and generally less refined.

As humans increasingly rely on LLMs for idea generation and sampling of actions, the set of possibilities they consider may expand beyond what they would typically generate on their own. Human-generated options tend to be more constrained, reflecting a limited yet contextually relevant subsets of possibilities. In contrast, LLMs introduce a broader range of alternative actions, potentially shifting the representation about what an agent could have reasonably done differently.

The expansion of the human reasoning space doesn't just change how people think about solving problems. It also reshapes how we judge the actions of others. When more – and better – alternatives come to mind, it can shift our sense of what someone could have done differently. And that, in turn, affects how we assign responsibility and blame.

Shifting Blame

Imagine now learning what Brad decided to do: He instructed everyone to huddle together to conserve energy. As a result, it took the rescue team longer to find them, leading to serious health issues for some members of the hiking group. How much is Brad to blame?

Your answer to this question will be influenced by the kinds of alternative possibilities you considered Brad could have chosen from. For example, the human study participant suggested he could have hunted animals for food – but that might not have been a better option. As a result, you might not blame Brad much. Alternatively, ChatGPT suggested that Brad could have shouted at intervals, a likely better option that might have led to a quicker rescue.

Consequently, depending on whether you used AI to think through the problem or relied on your own ideas, your attribution of blame could change. If you considered the alternative options generated by ChatGPT, you might blame Brad more because the AI generated better alternatives than you might have come up with on your own. And indeed, this is what our forthcoming research shows. In a series of studies, we found that participants who considered AI-generated alternatives judged agents more harshly compared to participants who relied only on their own ideas. Crucially, this pattern holds not only in specialized domains where ChatGPT draws on expert knowledge (like outdoor survival strategies), but also in everyday, mundane scenarios.

Outlook

When we embed LLMs into our thought and reasoning processes, they expand the range of possibilities we consider. This expansion may have downstream effects – for example, influencing how much blame we assign to people in certain scenarios, as well as shaping other kinds of judgments and decisions. Even in everyday use, we often turn to LLM-powered chatbots for ideas, input, and to help us think through problems. How these expanded "reasoning spaces" afforded by LLMs affect human cognition and behavior remains an exciting open avenue for research.

About the Author

 Dr. Lara Kirfel

Max Planck Institute for Human Development