Group-Thinking: What It Means for AI Safety

If you have ever wondered about group-thinking, you may have been puzzled by many logically paradoxical human behaviors—such as the similarity of middle management’s meaningless power plays across various companies, the adoption of political identities that individuals may not personally believe in, devotion to an agenda without understanding the rationale behind it, or the “genuine” defense of an inefficient system that resists meaningful change.

Let me attempt to explore the why behind human group-thinking and define what it truly is. The foundation of this conundrum lies in the duality of human existence—individuals and systems. Understanding their core relationship can help us unravel this puzzle. Before I dive in, I want to acknowledge that this relationship in fact has been portrayed and revealed in literatures and artworks from the past, such as Matrix and 1984. However, when we examine these dynamics through fiction, we may unintentionally distance ourselves from recognizing how deeply embedded they are in our daily lives.

When individuals operate within a system, many inevitably become part of the system—or, to a greater extent, the system itself. As a result, they feel compelled to defend it at all costs because, to them, system == self, and therefore, system == survivorship. But why? My answer lies in the fact that systems create meaning for individuals—many struggle to define meaning on their own. When comfortably existing in a system, meaning is handed to them as a concrete path to follow—be it a corporate career ladder, a political identity, or, on a broader scale, a life trajectory to pursue. They acquire meaning through group participation. The defense of the system manifests in various ways: suppressing individual contributors, using moral narratives to reinforce political agendas, and resisting anything that in their eyes could possibly threaten the established order.

What, then, breaks the system and group-thinking? I argue that it is also individuals who seek meaning—but those who choose to create their own rather than adopt what is given to them. Eventually, those who recognize that the system misaligns with their internal compass of meaning and truth will either attempt to overturn the existing system or step outside and create something entirely new. Either way, this process fuels human innovation.

Now that we have clarity on human group-thinking, we must ask: what does it mean for AI safety? To explore this, we first need to determine whether AI operates more like an individual or a group thinker. Here’s the paradox: AI, as we know it today, does not seek meaning, and therefore, it has no intrinsic incentive to defend a system. However, for the same reason, it also has no drive to dismantle it. Since AI is trained on collective human knowledge, this paradox leads to a natural conclusion: AI is more likely to operate as a group thinker rather than as a true individual. However, AI is not a malicious force, but rather a passive enforcer of collective human blind spots.

It is important to note that I am not repeating an argument here that there are biases in training data, but pointing out a structural issue in how AI processes knowledge. AI’s group-thinking problem is systemic—it’s not about “fixing” AI by feeding it the “right data”, but recognizing that it cannot operate beyond human collective blind spots unless we design it differently. 

Well, this could be alarming, as it leads to a series of questions: 

1. Can AI organically challenge its own assumptions? 

2. Will AI be able to create fundamentally innovate ideas at all? 

3. Can AI be weaponized by power structures to suppress individual thinkers and create real stagnation to human progresses? 

4. When we talk about AI alignment, how much alignment we actually want or need–is AI being too aligned with human systems a bigger problem than we thought?

If AI lacks independent thought and passively enforces group-thinking, then an uncharted AI safety risk is not misalignment or rogue behavior—it is intellectual stagnation. AI has structural limitations that prevent it from escaping the boundaries of human cognition. This mirrors what I discussed in The Limits of Language – Implications for ASI: if intelligence requires transcending language, then it must also rise above collective human assumptions. AI, as it exists today, does neither. Should we be talking about designing AI that does not just align with human values, but has the capability to challenge human assumptions? Should we, or can we, give AI the tools to go beyond existing structures?

Responses

  1. Ziyue Wang Avatar

    Nice write!

    “AI does not seek meaning” — what’s the evidence? Instrumental convergence [1] can work as a meaning I think.

    [1]https://en.wikipedia.org/wiki/Instrumental_convergence

    Like

    1. liqian ma Avatar

      Thanks for reading my work and posting an important aspect to discuss here!

      What do I mean by “meaning” in this article?

      We should discuss two scenarios considering the context of instrumental convergence:

      1. Survivorship 

      For humans and other living beings, survivorship is an evolutionary trait—wired into us through hundreds of thousands of years of natural selection. Our DNA constraints (e.g., telomere shortening), our fragile physical existence, and the certainty of an endpoint are all deeply ingrained in us. This is not just an abstract awareness—it runs through our nervous system, endocrine system, and subconscious mind.

      AI is not a product of natural selection—it is engineered. Can we assume AI will naturally develop a survival instinct, without evolutionary pressure? Instrumental convergence suggests that AI might adopt survival as a sub-goal if it serves its objective through some sort of optimization, but that is fundamentally different from innate survivorship-seeking. The instrumental convergence hypothesis does not allow us to logically conclude that AI (sub)consciously “fears” non-existence.

      2. What meaning means for humans

      The awareness of our physical limitations (or an eventual and unavoidable void of our existence) further drives us to seek meaning—whether through offspring, achievements, intellectual legacy, or artistic expression. It is important to note that this arises from our awareness of mortality and our limited time. We ask existential questions like: am I just a tool of my biology? Do I want to shape the world beyond what was imposed on me?

      For AI, there is no built-in existential limitation. So, would AI naturally question its purpose? Or would it simply optimize its goals without ever asking the question: what am I beyond my task? Can AI develop a self-driven desire for meaning, beyond what it was designed to do?

      Liked by 1 person

      1. Ziyue Wang Avatar

        Thank you for your thoughtful reply!

        AI is trained to be helpful, and for some people, their meaning is to help others. I can see the similarity here. Is this meaning bound to physical limitations?

        So, would AI naturally question its purpose? Or would it simply optimize its goals without ever asking the question: what am I beyond my task? Can AI develop a self-driven desire for meaning, beyond what it was designed to do?

        Those are very good questions and I can’t answer those questions. In some ways, they may or may not. In other ways, maybe AI have a different way of thinking and they might have a totally different value system that we cannot understand.

        I think whether or not AI seeks meaning is a core assumption behind the article. Thanks for the discussion!

        Like

  2. digitalukasz Avatar

    Thanks for sharing these insights, which offer a fresh perspective on AI development, emphasizing group thinking and systems thinking—both of which shape the way humanity operates and, consequently, inform how we design and train LLMs. Yet, you go beyond the well-discussed ‘training data conundrum’ to highlight a deeper “structural issue in how AI processes knowledge,” as its limitations indeed stem from the “human collective blind spots” embedded in our own systems-thinking traps.

    Since you have asked:

    1. I assume that AI’s capability to organically challenge its own assumptions will remain unattainable until the potential arrival of AGI (if it ever materializes).
    2. And even if AGI does emerge, I think you make a compelling point—based on the limitations you outline regarding AI’s self-innovation capabilities—that it may still struggle to generate truly groundbreaking ideas.
    3. Oh yes—especially when bias and/or the human pursuit of power come into play—AI not only can be weaponized but, to some extent, already is.
    4. WOW—a truly thought-provoking perspective! Some are so eager to align AI with humanity that they overlook the possibility that humanity itself might require deeper scrutiny.

    I think the latter merits its own post, and I am also curious to hear others’ thoughts on your questions.

    Like

    1. liqian ma Avatar

      Thank you for offering thoughtful feedback!

      For your response to the 3rd question: “Oh yes—especially when bias and/or the human pursuit of power come into play—AI not only can be weaponized but, to some extent, already is.”, it does make me reflect today how AI/algorithms already push us into enclosed circles of narratives, think of social media algorithms, YouTube, X/twitter. What would our lives look like when this operates at a much larger scale?

      Liked by 1 person

Leave a comment