[dropcap]T[/dropcap]he term “hive mind” refers to the apparent intelligence that emerges at the group level in some social species, particularly insects like honeybees and ants. An individual honeybee might not be very bright (although that’s debatable), but the honeybee colony as a collective might be very intelligent.
Science fiction is replete with depictions of higher-intelligence organisms forming a hive mind. One canonical example is the Borg from Star Trek. The Borg assimilate other intelligent organisms like humans into the “Collective,” but in so doing obliterate the individuality of the organisms they are assimilating.
Clearly the assimilation of humans into a hive mind is supposed to be a bad thing, according to the writers of Star Trek. The same goes for the writers of Doctor Who: the Cybermen pose a very similar threat as the Borg, capturing and “upgrading” humans so that they lose emotion and the ability to think individually (in fact there’s a comic in which the Borg and Cybermen team up).
These depictions of a potential future for humanity raise a number of questions. The first is whether humans are even capable of forming a hive mind, and if so whether it is likely that they will. Second, must the individual be erased as it is subsumed into a collective consciousness? Third, if individuality must be lost, is that necessarily a bad thing?
Last year, io9 ran an article about the potential for humans to form a hive mind. [Daniel’s post originally appeared Oct. 28, 2013 on danieltoker.com – Ed.]. Several scientists weighed in on the question and concluded that in order to get a hive mind, humans would have to evolve greater cooperation and division of labor. That is, after all, how we think honeybees did it evolutionarily. But I want to put this into more fundamental terms.
“Why is it that greater cooperation and division of labor would lead to the development of a hive mind?”
One possible answer is because that’s how large neural networks work. One popular (though largely untested) theory of consciousness is that of neuroscientist Giulio Tononi, whose information integration theory proposes that consciousness emerges wherever you have a system that is simultaneously functionally specialized and integrated, so that specific modules process specific kinds of information, but the output of those modules gets integrated across the entire network. According to Tononi, consciousness is the information that gets integrated across the system. This is what you see in the thalamocortical network: various modules are specialized to process certain types of stimuli like faces or speech, but the “output” of those modules gets integrated across the entire cortex and thalamus. And, in some sense, that’s what honeybees do: individual honeybees have very specific tasks like brood rearing, hive maintenance, foraging, or defense, but they share information with one another in a highly integrated fashion whenever there’s something the entire colony needs to be aware of (like where to find food).
“Now the real question is, what happens to our individuality if we give rise to a hive mind?”
So, in order to form a hive mind, humans may have to act more like an integrated neural network. There are many forms that can take. In a sense, it’s already happening thanks to cell phones and the Internet. Careers are becoming increasingly specialized and we share almost all our information on the web or email, phone, text, conferences, etc. So humans do exhibit simultaneous specialization and integration. One limiting factor in forming a mind, however, is time: it takes a while for us to figure out anything that is relevant to all of humanity (or just a large group of people), and it takes a long time to communicate what we’ve found to everyone. One example is climate change: how long did it take us to figure out it’s happening, and how much longer before the information about climate change really sinks in across the board and we do something about it as a collective? One possible way to circumvent this limitation is through wireless brain-to-brain communication, which isn’t that far off in terms of available technology but might be quite far off in terms of practical applicability.
Now the real question is, what happens to our individuality if we give rise to a hive mind? Are our individual minds obliterated? Or is it just business as usual while a collective intelligence supervenes on our brains, unbeknownst to us? Giulio Tononi (of the information integration theory) actually thinks that the individual consciousness winks out when the group forms a truly integrated whole. At a lecture last year Tononi said that if the United States ever came to a point where it, as a nation, integrated more information than any of the individuals that make up the nation, those individuals would cease to be conscious. This is because, as has been shown mathematically in artificial networks, the amount of information integrated inside the parts of the network (e.g. U.S. citizens) winks out to 0 bits as soon the information integrated across a whole network (e.g. the United States) exceeds the amount previously integrated inside its parts (its citizens). But, the theory remains untested in real brains, so it’s not entirely clear how plausible this assertion is.
But, say that we do lose our individual consciousness if a hive mind emerges. Is that necessarily a bad thing? Clearly our gut reaction is yes, because we value our own mentality. But suppose it were the case that the collective human mind could receive more utility than could all individual humans combined. I’m getting this idea from the notion of a “utility monster,” proposed by philosopher Robert Nozick as a critique of utilitarianism. According to utilitarianism, we would be morally obligated to sacrifice our own utility in the service of the utility monster, whose utility in being served outweighs our suffering in serving it. Some philosophers see this as a great example why utility is not the metric to be thinking about when it comes to morality. Utilitarians, on the other hand, bite the bullet and say that we should in fact suffer for this being if the utility it gains is greater than our suffering. In terms of the hive mind, utilitarians might say that we should allow ourselves to be subsumed into a collective consciousness if that consciousness can reap greater utility than all of us as individuals combined.
Personally, I hope this won’t be a problem. Nor do I really worry that it will be a problem. It may be possible for a higher collective consciousness to emerge without turning us into the Borg. Maybe a hive mind will only use a small portion of each person’s brain, which would allow people to carry out their lives as usual. And maybe that collective mind would be far more intelligent than any single human and could call us into action when all of humanity is threatened – again, without destroying our individuality in the process.
This post originally appeared on Daniel Toker’s website. Minor changes were made to this text on February 3, 2018.