Home » Posts tagged 'Theory of Mind'
Tag Archives: Theory of Mind
Want yet more email?
Recent blog posts
John Wyatt is a doctor, author and research scientist. His concern is the ethical challenges that arise with technologies like artificial intelligence and robotics. On Tuesday this week (11th March 2019) he gave a talk called ‘What does it mean to be human?’ at the Wesley Methodist Church in Cambridge.
To a packed audience, he pointed out how interactions with artificial intelligence and robots will never be the same as the type of ‘I – you’ relationships that occur between people. He emphasised the important distinction between ‘beings that are born’ and ‘beings that are made’ and how this distinction will become increasingly blurred as our interactions with artificial intelligence become commonplace. We must be ever vigilant against the use of technology to dehumanise and manipulate.
I can see where this is going. The tendency for people to anthropomorphise is remarkably strong - ‘the computer won’t let me do that’, ‘the car has decided not to start this morning’. Research shows that we can even attribute intentions to animated geometrical shapes ‘chasing’ each other around a computer screen, let alone cartoons. Just how difficult is it going to be to not attribute the ‘human condition’ to a chatbot with an indistinguishably human voice or a realistically human robot. Children are already being taught to say ‘please’ and ‘thank you’ to devices like Alexa, Siri and Google Home – maybe a good thing in some ways, but …
One message I took away from this talk was a suggestion for a number of new human rights in this technological age. These are: (1) The right to cognitive liberty (to think whatever you want), (2) The right to mental privacy (without others knowing) (3) The right to mental integrity and (4) The right to psychological continuity - the last two concerning the preservation of ‘self’ and ‘identity’.
A second message was to consider which country was most likely to make advances in the ethics of artificial intelligence and robotics. His conclusion – the UK. That reassures me that I’m in the right place.
See more of John’s work, such as his essay ‘God, neuroscience and human identity’ at his website johnwyatt.com
How do we know what we know?
This article considers:
(1) the ways we come to believe what we think we know
(2) the many issues with the validation of our beliefs
(3) the implications for building artificial intelligence and robots based on the human operating system.
I recently came across a video (on the site http://www.theoryofknowledge.net) that identified the following ‘ways of knowing’:
- Sensory perception
This list is mainly about mechanisms or processes by which an individual acquires knowledge. It could be supplemented by other processes, for example, ‘meditation’, ‘science’ or ‘history’, each of which provides its own set of approaches to generating new knowledge for both the individual and society as a whole. There are many difference ways in which we come to formulate beliefs and understand the world.
Youtube Video, TOK Ways of Knowing EXPLAINED | Theory of Knowledge Advice, Ivy Lilia, October 2018, 6:16 minutes
In the spirit of working towards a description of the ‘human operating system’, it is interesting to consider how a robot or other Artificial Intelligence (AI), that was ‘running’ the human operating system, would draw on its knowledge and beliefs in order to solve a problem (e.g. resolve some inconsistency in its beliefs). This forces us to operationalize the process and define the control mechanism more precisely. I will work through the above list of ‘ways of knowing’ and illustrate how each might be used.
Let’s say that the robot is about to go and do some work outside and, for a variety of reasons, needs to know what the weather is like (e.g. in deciding whether to wear protective clothing, or how suitable the ground is for sowing seeds or digging up for some construction work etc.) .
First it might consult its senses. It might attend to its visual input and note the patterns of light and dark, comparing this to known states and conclude that it was sunny. The absence of the familiar sound patterns (and smell) of rain might provide confirmation. The whole process of matching the pattern of data it is receiving through its multiple senses, with its store of known patterns, can be regarded as ‘intuitive’ because it is not a reasoning process as such. In the Khanemman sense of ‘system 1’ thinking, the robot just knows without having to perform any reasoning task.
Youtube Video, System 1 and System 2, Stoic Academy, February 2017, 1:26 minutes
The knowledge obtained from matching perception to memory can nevertheless be supplemented by reasoning, or other forms of knowledge that confirm or question the intuitively-reached conclusion. If we introduce some conflicting knowledge, e.g. that the robot thinks it’s the middle of the night in it’s current location, we then create a circumstance in which there is dissonance between two sources of knowledge – the perception of sunlight and the time of day. This assumes the robot has elaborated knowledge about where and when the sun is above the horizon and can potentially shine (e.g. through language – see below).
In people the dissonance triggers the emotional state of ‘surprise’ and the accompanying motivation to account for the contradiction.
Youtube Video, Cognitive Dissonance, B2Bwhiteboard, February 2012, 1:37 minutes
Likewise, we might label the process that causes the search for an explanation in the robot as ‘surprise’. An attempt may be made to resolve this dissonance through Kahneman’s slower, more reasoned, system 2 thinking. Either the perception is somehow faulty, or the knowledge about the time of day is inaccurate. Maybe the robot has mistaken the visual and audio input as coming from its local senses when in fact the input has originated from the other side of the world. (Fortunately, people do not have to confront the contradictions caused by having distributed sensory systems).
Probably in the course of reasoning about how to reconcile the conflicting inputs, the robot will have had to run through some alternative possible scenarios that could account for the discrepancy. These may have been generated by working through other memories associated with either the perceptual inputs or other factors that have frequently led to mis-interpretations in the past. Sometimes it may be necessary to construct unique possible explanations out of component part explanations. Sometimes an explanation may emerge through the effect of numerous ideas being ‘primed’ through the spreading activation of associated memories. Under these circumstances, you might easily say that the robot was using it’s imagination in searching for a solution that had not previously been encountered.
Youtube Video, TEDxCarletonU 2010 – Jim Davies – The Science of Imagination, TEDx Talks, September 2010, 12:56 minutes
Lastly, to faith and language as sources of knowledge. Faith is different because, unlike all the other sources, it does not rely on evidence or proof. If the robot believed, on faith, that the sun was shining, any contradictory evidence would be discounted, perhaps either as being in error or as being irrelevant. Faith is often maintained by others, and this could be regarded as a form of evidence, but in general if you have faith in or trust something, it is at least filling the gap between the belief and the direct evidence for it.
Here is a religious account of faith that identifies it with trust in the reliability of God to deliver, where the main delivery is eternal life.
Youtube video, What is Faith – Matt Morton – The Essence of Faith – Grace 360 conference 2015,Grace Bible Church, September 2015, 12:15 minutes
Language as a source of evidence is a catch-all for the knowledge that comes second hand from the teachings and reports of others. This is indirect knowledge, much of which we take on trust (i.e. faith), and some of which is validated by direct evidence or other indirect evidence. Most of us take on trust that the solar system exists, that the sun is at the centre, and that earth is in the third orbit. We have gained this knowledge through teachers, friends, family, tv, radio, books and other sources that in their turn may have relied on astronomers and other scientist who have arrived at these conclusions through observation and reason. Few of us have made the necessary direct observations and reasoned inferences to have arrived at the conclusion directly. If our robot were to consult databases of known ‘facts’, put together by people and other robots, then it would be relying on knowledge through this source.
People like to think that their own beliefs are ‘true’ and that these beliefs provide a solid basis for their behaviour. However, the more we find out about the psychology of human belief systems the more we discover the difficulties in constructing consistent and coherent beliefs, and the shortcomings in our abilities to construct accurate models of ‘reality’. This creates all kinds of difficulties amongst people in their agreements about what beliefs are true and therefore how we should relate to each other in peaceful and productive ways.
If we are now going on to construct artificial intelligences and robots that we interact with and have behaviours that impact the world, we want to be pretty sure that the beliefs a robot develops still provide a basis for understanding their behaviour.
Unfortunately, every one of the ‘ways of knowing’ is subject to error. We can again go through them one by one and look at the pitfalls.
Sensory perception: We only have to look at the vast body of research on visual illusion (e.g. see ‘Representations of Reality – Part 1’) to appreciate that our senses are often fooled. Here are some examples related to colour vision:
Youtube Video, Optical illusions show how we see | Beau Lotto,TED, October 2009, 18:59 minutes
Furthermore, our perceptions are heavily guided by what we pay attention to, meaning that we can miss all sorts of significant and even life-threatening information in our environment. Would a robot be similarly misled by its sensory inputs? It’s difficult to predict whether a robot would be subject to sensory illusions, and this might depend on the precise engineering of the input devices, but almost certainly a robot would have to be selective in what input it attended to. Like people, there could be a massive volume of raw sensory input and every stage of processing from there on would contain an element of selection and interpretation. Even differences in what input devices are available (for vision, sound, touch or even super-human senses like perception of non-visual parts of the electromagnetic spectrum), will create a sensory environment (referred to as the ‘umwelt’ or ‘merkwelt’in ethology) that could be quite at variance with human perceptions of the world.
YouTube Video, What is MERKWELT? What does MERKWELT mean? MERKWELT meaning, definition & explanation, The Audiopedia, July 2017, 1:38 minutes
Memory: The fallibility of human memory is well documented. See, for example, ‘The Story of Your Life’, especially the work done by Elizabeth Loftus on the reliability of memory. A robot, however, could in principle, given sufficient storage capacity, maintain a perfect and stable record of all its inputs. This is at variance with the human experience but could potentially mean that memory per se was more accurate, albeit that it would be subject to variance in what input was stored and the mechanisms of retrieval and processing.
Intuition and reason: This is the area where some of the greatest gains (and surprises) in understanding have been made in recent years. Much of this progress is reported in the work of Daniel Kahneman that is cited many times in these writings. Errors and biases in both intuition (system 1 thinking) and reason (system 2 thinking) are now very well documented. A long list of cognitive biases can be found at:
Would a robot be subject to the same type of biases? It is already established that many algorithms, used in business and political campaigning, routinely build in the biases, either deliberately or inadvertently. If a robot’s processes of recognition and pattern matching are based on machine learning algorithms that have been trained on large historical datasets, then bias is virtually guaranteed to be built into its most basic operations. We need to treat with great caution any decision-making based on machine learning and pattern matching.
Youtube Vide, Cathy O’Neil | Weapons of Math Destruction, PdF YouTube, June 2015, 12:15 minutes
As for reasoning, there is some hope that the robustness of proofs that can be achieved computationally may save the artificial intelligence or robot from at least some of the biases of system 2 thinking.
Emotion: Biases in people due to emotional reactions are commonplace. See, for example:
Youtube Video, Unconscious Emotional Influences on Decision Making, The Rational Channel, February 2017, 8:56 minutes
However, it is also the case that emotions are crucial in decision–making. Emotions often provide the criteria and motivation on which decisions are made and without them, people can be severely impaired in effective decision-making. Also, emotions provide at least one mechanism for approaching the subject of ethics in decision-making.
Youtube Video, When Emotions Make Better Decisions – Antonio Damasio, FORA.tv, August 2009, 3:22 minutes
Can robots have emotions? Will robots need emotions to make effective decisions? Will emotions bias or impair a robot’s decision-making. These are big questions and are only touched on here, but briefly, there is no reason why emotions cannot be simulated computationally although we can never know if an artificial computational device will have the subjective experience of emotion (or thought). Probably some simulation of emotion will be necessary for robot decision-making to align with human values (e.g. empathy) and, yes, a side-effect of this may well be to introduce bias into decision-making.
For a selection of BBC programmes on emotions see:
Imagination: While it doesn’t make much sense to talk about ‘error’ when it comes to imagination, we might easily make value-judgments about what types of imagination might be encouraged and what might be discouraged. Leaving aside debates about how, say excessive experience of violent video games, might effect imagination in people, we can at least speculate as to what might or should go on in the imagination of a robot as it searches through or creates new models to help predict the impacts of its own and others behaviours.
A big issue has arisen as to how an artificial intelligence can explain its decision-making to people. While AI based on symbolic reasoning can potentially offer a trace describing the steps it took to arrice at a conclusion, AIs based on machine learning would be able to say little more than ‘I recognized the pattern as corresponding to so and so’, which to a person is not very explanatory. It turns out that even human experts are often unable to provide coherent accounts of their decision-making, even when they are accurate.
Having an AI or robot account for its decision-making in a way understandable to people is a problem that I will address in later analysis of the human operating system and, I hope, provide a mechanism that bridges between machine learning and more symbolic approaches.
Faith: It is often said that discussing faith and religion is one of the easiest ways to lose friends. Any belief based on faith is regarded as true by definition, and any attempt to bring evidence to refute it, stands a good chance of being regarded as an insult. Yet people have different beliefs based on faith and they cannot all be right. This not only creates a problem for people, who will fight wars over it, but it is also a significant problem for the design of AIs and robots. Do we plug in the Muslim or the Christian ethics module, or leave it out altogether? How do we build values and ethical principles into robots anyway, or will they be an emergent property of its deep learning algorithms. Whatever the answer, it is apparent that quite a lot can go badly wrong if we do not understand how to endow computational devices with this ‘way of knowing’.
Language: As observed above, this is a catch-all for all indirect ‘ways of knowing’ communicated to people through media, teaching, books or any other form of communication. We only have to consider world wars and other genocides to appreciate that not everything communicated by other people is believable or ethical. People (and organizations) communicate erroneous information and can deliberately lie, mislead and deceive.
We strongly tend to believe information that comes from the people around us, our friends and associates, those people that form part of our sub-culture or in-group. We trust these sources for no other reason than we are familiar with them. These social systems often form a mutually supporting belief system, whether or not it is grounded in any direct evidence.
Youtube Video, The Psychology of Facts: How Do Humans (mis)Trust Information?, YaleCampus, January 2017
Taking on trust the beliefs of others that form part of our mutually supporting social bubble is a ‘way of knowing’ that is highly error prone. This is especially the case when combined with other ‘ways of knowing’, such as faith, that in their nature cannot be validated. Will robot communities develop, who can talk to each other instantaneously and ‘telepathically’ over wireless connections, also be prone to the bias of groupthink?
The validation of beliefs
So, there are multiple ways in which we come to know or believe things. As Descartes argued, no knowledge is certain (see ‘It’s Like This’). There are only beliefs, albeit that we can be more sure of some that others, normally by virtue of their consistency with other beliefs. Also, we note that our beliefs are highly vulnerable to error. Any robot operating system that mimics humans will also need to draw on the many different ‘ways of knowing’ including a basic set of assumptions that it takes to be true without necessarily any supporting evidence (it’s ‘faith’ if you like). There will also need to be many precautions against AIs and robots developing erroneous or otherwise unacceptable beliefs and basing their behaviours on these.
There is a mechanism by which we try to reconcile differences between knowledge coming from different sources, or contradictory knowledge coming from the same source. Most people seem to be able to tolerate a fair degree of contradiction or ambiguity about all sorts of things, including the fundamental questions of life.
Youtube Video, Defining Ambiguity, Corey Anton, October 2009, 9:52 minutes
We can hold and work with knowledge that is inconsistent for long periods of time, but nevertheless there is a drive to seek consistency.
In the description of the human operating system, it would seem that there are many ways in which we establish what we believe and what beliefs we will recruit to the solving of any particular problem. Also, the many sources of knowledge may be inconsistent or contradictory. When we see inconsistencies in others we take this as evidence that we should doubt them and trust them less.
Youtube Video, Why Everyone (Else) is a Hypocrite, The RSA, April 2011, 17:13 minutes
However, there is, at least, a strong tendency in most people, to establish consistency between beliefs (or between beliefs and behaviours), and to account for inconsistencies. The only problem is that we are often prone to achieve consistency by changing sound evidence-based beliefs in preference to the strongly held beliefs based on faith or our need to protect our sense of self-worth.
Youtube Video, Cognitive dissonance (Dissonant & Justified), Brad Wray, April 2011. 4:31 minutes
From this analysis we can see that building AIs and robots is fraught with problems. The human operating system has evolved to survive, not to be rational or hold high ethical values. If we just blunder into building AIs and robots based on the human operating system we can potentially make all sorts of mistakes and give artificial agents power and autonomy without understanding how their beliefs will develop and the consequences that might have for people.
Fortunately there are some precautions we can take. There are ways of thinking that have been developed to counter the many biases that people have by default. Science is one method that aims to establish the best explanations based on current knowledge and the principle of simplicity. Also, critical thinking has been taught since Aristotle and fortunately many courses have been developed to spread knowledge about how to assess claims and their supporting arguments.
Youtube Video, Critical Thinking: Issues, Claims, Arguments, fayettevillestatenc, January 2011
Sensory perception – The robot’s ‘umwelt’ (what it can sense) may well differ from that of people, even to the extent that the robot can have super-human senses such as infra-red / x-ray vision, super-sensitive hearing and smell etc. We may not even know what it’s perceptual world is like. It may perceive things we cannot and miss things we find obvious.
Memory – human memory is remarkably fallible. It is not so much a recording, as a reconstruction based on clues, and influenced by previously encountered patterns and current intentions. Given sufficient storage capacity, robots may be able to maintain memories as accurate recording of the states of their sensory inputs. However, they may be subject to similar constraints and biases as people in the way that memories are retrieved and used to drive decision-making and behaviour.
Intuition – if the robot’s pattern-matching capabilities are based on the machine learning of historical training sets then bias will be built into its basic processes. Alternatively, if the robot is left to develop from it’s own experience then, as with people, great care has to be taken to ensure it’s early experience will not lead to maladaptive behaviours (i.e. behaviours not acceptable to the people around it).
Reason – through the use of mathematical and logical proofs, robots may well have the capacity to reason with far greater ability than people. They can potentially spot (and resolve) inconsistencies arising out of different ‘ways of knowing’ with far greater adeptness than people. This may create a quite different balance between how robots make decisions and how people do using emotion and reason in tandem.
Emotion – human emotion are general states that arise in response to both internal and external events and provide both the motivation and the criteria on which decisions are made. In a robot, emerging global states could also potentially act to control decision-making. Both people, and potentially robots, can develop the capacity to explicitly recognize and control these global states (e.g. as when suppressing anger). This ability to reflect, and to cause changes in perspective and behaviour, is a kind of feedback loop that is inherently unpredictable. Not having sufficient understanding to predict how either people or robots will react under particular circumstances, creates significant uncertainty.
Imagination – much the same argument about predictability can be made about imagination. Who knows where either a person’s or a robot’s imagination may take them? Chess computers out-performed human players because of their capacity to reason in depth about the outcomes of every move, not because they used pattern-matching based on machine learning (although it seems likely that this approach will have been tried and succeeded by now). Robots can far exceed human capacities to reason through and model future states. A combination of brute force computing and heuristics to guide search, may have far-reaching consequences for a robot’s ability to model the world and predict future outcomes, and may far exceed that of people.
Faith – faith is axiomatic for people and might also be for robots. People can change their faith (especially in a religious, political or ethical sense) but more likely, when confronted with contradictory evidence or sufficient need (i.e. to align with a partner’s faith) people with either ignore the evidence or find reasons to discount it. This way can lead to multiple interpretations of the same basic axioms, in the same way as there are many religious denominations and many interpretations of key texts within these. In robots, Asimov’s three laws of robotics would equate to their faith. However, if robots used similar mechanisms as people (e.g. cognitive dissonance) to resolve conflicting beliefs, then in the same way as God’s will can be used to justify any behaviour, a robot may be able to construct a rationale for any behaviour whatever its axioms. There would be no guarantee that a robot would obey its own axiomatic laws.
Communication – The term language is better labeled ‘communication’ in order to make it more apparent that it extends to all methods by which we ‘come to know’ from sources outside ourselves. Since communication of knowledge from others is not direct experience, it is effectively taken on trust. In one sense it is a matter of faith. However, the degree of consistency across external sources and between what is communicated (i.e. that a teacher or TV will re-enforce what a parent has said etc.) and between what is communicated and what is directly observed (for example, that a person does what he says he will do) will reveal some sources as more believable than others. Also we appeal to motive as a method of assessing degree of trust. People are notoriously influenced by the norms, opinions and behaviours of their own reference groups. Robots with their potential for high bandwidth communication could, in principle, behave with the same psychology of the crowd as humans, only much more rapidly and ‘single-mindedly’. It is not difficult to see how the Dr Who image of the Borg, acting a one consciousness, could come about.
Other Ways of Knowing
It is worth considering just a few of the many other ‘ways’ of knowing’ not considered above, partly because some of these might help mitigate some of the risks of human ‘ways of knowing’ .
Science – Science has evolved methods that are deliberately designed to create impartial, robust and consistent models and explanations of the world. If we want robots to create accurate models, then an appeal to scientific method is one approach. In science, patterns are observed, hypotheses are formulated to account for these patterns, and the hypotheses are then tested as impartially as possible. Science also seeks consistency by reconciling disparate findings into coherent overall theories. While we may want robots to use scientific methods in their reasoning, we may want to ensure that robots do not perform experiments in the real world simply for the sake of making their own discoveries. An image of concentration camp scientists comes to mind. Nevertheless, in many small ways robots will need to be empirical rather than theoretical in order to operate at all.
Argument – Just like people, robots of any complexity will encounter ambiguity and inconsistencies. These will be inconsistencies between expectation and actuality, between data from one way of knowing and another (e.g. between reason and faith, or between perception and imagination etc.), or between a current state and a goal state. The mechanisms by which these inconsistencies are resolved will be crucial. The formulation of claims; the identification, gathering and marshalling of evidence; the assessment of the relevance of evidence; and the weighing of the evidence, are all processes akin to science but can cut across many ‘ways of knowing’ as an aid to decision making. Also, this approach may help provide explanations of a robot’s behaviour that would be understandable to people and thereby help bridge the gap between opaque mechanisms, such as pattern matching, and what people will accept as valid explanations.
Meditation – Meditation is a place-holder for the many ways in which altered states of consciousness can lead to new knowledge. Dreaming, for example, is another altered state that may lead to new hypotheses and models based on novel combination of elements that would not otherwise have been brought together. People certainly have these altered states of consciousness. Could there be an equivalent in the robot, and would we want robots to indulge in such extreme imaginative states where we would have no idea what they might consist of? This is not to necessarily attribute consciousness to robots, which is a separate, and probably meta-physical question.
Theory of mind – For any autonomous agent with its own beliefs and intentions, including a robot, it is crucial to its survival to have some notion of the intentions of other autonomous agents, especially when they might be a direct threat to survival. People have sophisticated but highly biased and error-prone mechanisms for modelling the intentions of others. These mechanisms are particularly alert for any sign of threat and, as a proven mechanism, tend to assume threat even when none is present. The people that did not do this, died out. Work in robotics already recognizes that, to be useful, robots have to cooperate with people and this requires some modelling of their intentions. As this last video illustrates, the modelling of others intentions is inherently complex because it is recursive.
YouTube Video, Comprehending Orders of Intentionality (for R. D. Laing), Corey Anton, September 2014, 31:31 minutes
If there is a conclusion to this analysis of ‘ways of knowing’ it is that creating intelligent, autonomous mechanisms, such as robots and AIs, will have inherently unpredictable consequences, and that, because the human operating system is so highly error-prone and subject to bias, we should not necessarily build them in our own image.
Executive Function in the Individual and the Organisation
Successful organisations like Google and Facebook allow their employees an opportunity to experiment and pursue their own projects. Many public sector organisations also allow their employees opportunity for personal development. Why does this work and what does it say about how organisations need to be run in a world of increasingly rapid change? What kind of executive control is appropriate for organisations in the 21st Century?
To answer this we could look at all kinds of management, organisational and accounting theory. But there is another perspective. This is to look at what psychology has revealed about the executive function (REFs 1, 2, 3) in the individual and then to map that back onto what it means in terms of the organisation. This perspective can be revealing. It highlights why organisations behave in certain ways, it can help distinguish useful and healthy behaviours from those that are ineffective, aberrant and perhaps eventually self-defeating, and it can give us a way of looking at the executive function that is grounded in an increasingly sophisticated understanding of the human condition. It can point the way to making organisations more resilient.
YouTube Video, 2012 Burnett Lecture Part 2 ADHD, Self-Regulation and Executive Functioning Theory, UNCCHLearningCentre, November 2012, 58:44
YouTube Video, InBrief: Executive Function: Skills for Life and Learning, Center on the Developing Child at Harvard University, June 2012, 5:35 minutes
Youtube Video, Executive Function and the Developing Brain: Implications for Education, AMSDMN’s channel, November 2013, 58:22 minutes
There are several distinct components to executive function in the individual. These develop from infancy to adulthood more or less in order. This article looks at the overall architecture of control within an organisation then goes through seven executive functions one by one, first looking at what it means in psychological terms, then mapping it onto what it might mean in terms of organisational behaviour and the functions of an executive board.
In both the individual and the organisation, executive function is self-regulation. It is ‘actions on oneself’ or, in the organisational context, the executive actions in relation to the organisation itself. When fully developed the several aspects of executive function go together to provide the capacity for self-control in a complex and changing world.
There are numerous accounts of what makes for success in both the individual (and in the organisation). Many of these emphasise one or other aspect of executive function such as self-awareness, self-direction or emotional intelligence. However, all aspects of the executive function have a part to play, and understanding executive function helps demonstrate how all these parts develop and integrate to provide the many capabilities needed for success.
The Architecture of Control
The overall architecture of control in both the individual and the organisation can be seen as a two-part system with executive function residing in the second part.
Part 1 – An Automatic System
Much of what happens in both individuals and organisations goes on without much thought or reflection.
Individuals follow their routines and habits. When everything is predictable, actions like cooking or driving can be carried out largely on ‘autopilot’, often while thinking about something entirely different. In this manner, we can operate adequately on the basis of responding to cues in the immediate environment with little conscious control or effort. This is what Kahneman in his book ‘Thinking fast and Slow’ (REF 4) calls ‘System 1’ or intuitive thinking. It deals with the here and now when everything is familiar and reasonably certain.
YouTube Video, Kahneman: “Thinking, Fast and Slow” | Talks at Google, November 2011, 1:02 hours
Similarly, in the organisation many activities can be carried out according to it’s established procedures and practices and require no executive intervention. They may have needed executive intervention to set them up but once bedded-in they can run without further executive input unless something unexpected or out of the ordinary happens.
Part 2 – A System for Exception Handling and Taking Proactive Control
This system is engaged when the automatic system needs help. In terms of Khaneman’s theory, it is ‘System 2’ thinking. It is engaged when encountering difficulty, novelty and in matters that are not in the here and now. The executive level allows ‘action at a distance’ from the here and now, and deals with circumstances that are less certain and predictable.
In the individual, when something unpredictable happens, this system seems to pop items into consciousness and then relies on a somewhat slow and labourious form of conscious processing to effect a resolution. This takes effort and resource. It takes willpower and can use up cognitive capacity. What can be done is limited by the available capacity, and focus of attention on one thing will limit the capacity to pay attention to another.
In the organisation the executive may be called in to deal with some problem or may step in when it sees something going off-track (like profits, sales, production, costs, staff-turnover etc). As with the individual, this system, is slow and labourious by contrast to the ‘business as usual’ operation. It also requires the consumption of precious resources and dealing with one situation can detract from dealing with another, perhaps causing the organisation to take it’s eye off the ball.
In practice, in a healthy individual or organisation, the automatic system and the exception handling system work together in a highly interleaved manner. They also develop together. Functions that start out as requiring exception handling, become automated over time as they become embedded.
In the individual, the pro-active element of executive function emerges in childhood. The extent of the ability to inhibit certain behaviours correlates highly with many factors in later life including academic and social competence, wealth, health and (negatively with) criminality. It seems that the developing child moves gradually from reactive control to pro-active control. (REF 5).
YouTube Video, Integrative Science Symposium: Lifespan Development of Executive Control, July 2015, 2:10:08 hours
This is more than just being able to stop or inhibit certain behaviours. There is an extra step. This is to re-construe the world in a slightly different way and become alert to other things going on in the environment. The extra step leaves open the option to continue the behaviour, stop it or do something more subtle.
Similarly, in the organisation, spotting an undesirable behaviour, trend or process does not generally result in immediately shutting it down. There is a period of reflection, where attention may be re-focused and alternatives considered. In both organisations and individuals, this may take time. An individual may think through the potential consequences of taking particular courses of action. The organisation may do the same by embarking on investigations or sophisticated modelling and simulations to help clarify the consequences of running with different options.
A simplified model of the control and executive function is:
- No problems – continue on autopilot
- Problem – re-focus attention, generate and evaluate options
Also, it is notable that problem solving can go on recursively. So, if a problem is encountered with any of the sub-processes of problem solving, then that’s a new problem that is subject to the same processes in achieving resolution. Meanwhile the whole process is being recursively and externally evaluated such that if it’s not going anywhere useful, it itself can be re-considered.
7 Stages of Development of Executive Function
Seven stages of development of the executive function are described in terms of what’s going on for both the individual and the organisation. There are striking parallels.
Many of these stages have a time dimension. An infant lives in the here and now, a teenager in perhaps weeks. An adult, like an organisation, may have a time horizon of months or years. The development of executive function enables longer time horizons.
In Early Development (0 years to 5 years)
Stage 1 – Self-Awareness
In the individual, self-awareness develops in infancy (from 3 months and continues to develop for a further 10 years). The capacity to turn one’s attention away from the environment and towards ones own actions and thoughts, grows. The self-monitoring function redirects attention back on the self. An executive has developed that watches the self.
An organisation might be self-aware from the start if it has been set up with management information systems. The executives can inspect the reports and consider actions designed to affect trends they see in the data. The executives normally act within a stable framework of parameters albeit that the values on the parameters are changing. Also an organisation may grow in self-awareness by introducing new systems to provide feedback on what are deemed to be key parameters.
When (informal or formal) management information systems come into play, they can become the basis of a prevailing viewpoint on the direction of the organisation, both past and future. This is the backdrop against which executive decision-making takes place. The organisation has become to some degree self-aware and able to turn attention onto itself.
However, both the individual and the organisation operate in what Simon refers to as ‘bounded rationality’ (REF 6). Organisational self-awareness is subjective in the sense that the organisation is only aware of what it is aware of. It may not be aware of all manner of things and what it is aware of may not be representative or accurate. In this sense, the organisation mimics the individual and may be subject to the same misconceptions about the self.
YouTube Video, Herbert Simon, rationalLeft, July 2013, 3:42 mins
For example, Baring Bank may have been blind to the extent of the damage a rogue trader could do. Kodak may have deluded itself into thinking that there would always be a market for film (as opposed to digital). In 2008, the banking industry may have been unaware of the damage it might do to itself by not containing risk. Just as likely, these organisations were aware but didn’t care or know how to react.
These are just examples of self-awareness of organisations and begs the question of who holds this awareness. Is it distributed throughout the organisation or is it held by the executive? Just as in the brain, any one neuron is ‘aware’ of the activity of other neurons it is connected to, whether they be near neighbours or in some more remote location, awareness is distributed throughout the individuals and departments in an organisation. Each department and organisational role is tasked with being informed about particular things – production, human resources, suppliers and so on. In collecting and reporting both quantitative and qualitative data about local activity, the executive is fed information from across the organisation and can build a broader awareness. However, just as the brain can be deceived by it’s senses or selective and biased in its interpretation, the uncritical executive can also be led astray.
Stage 2 – Self-Restraint
Self-restraint develops In the individual between the ages of 3 and 5 years. This is the ability for a child to stop him or herself doing something that they would otherwise do automatically (like taking a sweet). It is inhibition of action. Children between 3 and 5 years will put their hands over their mouths to stop themselves saying something. In time this ‘executive inhibition’ can be done internally but it takes effort. It draws down on a limited resource.
In the organisation there are several mechanisms for self-restraint. Budgets act as inhibitors by containing costs in parts of the organisation and overall. Also, organisational policies, procedures, standards and guidelines are often designed to inhibit behaviours other than those already approved. The development of procedures is often motivated by ‘error’ – something has gone wrong and the organisation tries to make sure it doesn’t happen again. Procedures, new or just changing, often meet some resistance and it takes effort or resource to overcome it. Once bedded in, however, they can be executed more cheaply. The operation of restraint has become automatic. It has moved from the proactive and exception handling control system to the routine and automatic control system.
Stage 3 – Imagery and rapid brain development
In the individual imagery is ‘the mind’s eye’ or ‘a theatre in your mind’. It is the ability to resurrect (visual and other) images from the past (together with accompanying emotions) to deal with the present. In the child, imagery develops between 3 and 5 years. This is mainly imagery of situations represented in all the salient senses – visual, auditory, tactile, taste and smell – whatever was relevant at the time. The imagery may be stored along with how you feel about it, good or bad (to some degree). Pattern-matching triggers memories and resurrects relevant imagery from the past to act as a guide or a map that can be used in the here and now.
The developing brain at this stage consumes 60% of the glucose consumed in food and is creating new connections between neurones at a rapid rate. Although the visual systems in the brain can be fully wired up from the age of one, other sensory modalities take longer. At some stage a tipping point is reached when infrequently used connections are purged. This developmental progression is thought to facilitate innovation and hypothesis testing about the environment up to the point where consolidation on viable interpretations set in. The young mind mimics the progression of science in exploration before consolidation into useful knowledge that can drive applications.
Youtube Video, Alison Gopnik Lecture at CFI – When and why children are more intelligent than adults are, Future of Intelligence, September 2017, 1:31:50 hours starting at 15:35 minutes
In the organisation, the memories of staff and the management of files, file sharing, databases and information systems is its ‘working memory’. These systems are largely designed with retrieval in mind. Although, the data captured is not inherently what you would call image-evoking, it does perform the same function of retrieving memories (records, documents, anecdotes) that help guide action in the here and now. Envisioning activities, prototyping, simulation and modelling activities in the organisation, parallel the ‘theatre of the mind’. They are part of the organisations imaginative activity, where ideas can be tried out before they are fully implemented (and incur the full costs and consequences of acting on the world outside the organisation).
Stage 3a – Theory of Mind
There is another important stage that appears to develop between the ages of 3 and 5 years, that tends not to be emphasised in the mainstream literature on executive function. This is the so-called ‘theory of mind’ (REF6a) – the ability of a person to model what another person is thinking and feeling. Experiments with children show that a three year old expects everybody else to know what they themselves know, while by 5 years a child understands that other people can have different beliefs from themselves. If, for example, the content of a chocolate box is replaced with, say crayons, in front of a three year old, they will think that somebody later coming into the room will expect to find crayons in the box rather than chocolates. They are unable to differentiate between their own knowledge and the knowledge of others.
YouTube Video, Robert Seyfarth: Theory of Mind, Richard Dawkins Foundation for Reason & Science, May 2010, 3:36 minutes
Theory of mind may not be addressed in mainstream accounts of executive function because it is thought of as a social skill rather than a fundamental information processing capability, but I think it should be thought of as a key part of executive function because it is a base on which later executive functions develop. The multiple voices that develop in private speech, for example, are akin to the playing out in the mind of multiple belief systems and theory of mind must also impact on management of ones own emotions and motivations. In fact, the implications of theory of mind are so significant that it has generated its own large literature. Autism and Asperger’s Spectrum disorders increasingly reference both deficiencies in executive function and in theory of mind, adding further support to the argument that theory of mind should be seen as an aspect of executive function.
Research at the Max Planck Institute suggests that the maturation of fibres of a brain structure called the arcuate fascicle, between the ages of three and four years, establishes a connection between (1) a region at the back of the temporal lobe that supports adults thinking about others and their thoughts and (2) a region in the frontal lobe that is involved in keeping things at different levels of abstraction.
Article, Brain Structues that help us understand What Others Think Revealed, Neuroscience News, March 2017
In the organisation, ‘theory of mind’ is akin to understanding your competitors and your markets. If an organisation’s theory is accurate then it will be better able to anticipate the consequences of events, both those that it control and those that are external (e.g. government legislation and changes in market conditions). For example, if an organisation changes the price of one of its products, it would be useful to be able to predict what its customers and competitors would think about this and how they are likely to respond. A good businessman, like a good car salesman, may have an instinct about how customers will respond and may be able to construct a more or less complicated strategy that will drive the behaviours of others in particular ways.
From 5 Years
Stage 4 – Private Speech
In the individual, at 3 years old everything is public. Children talk to themselves about the world. Listening to their own speech is a mechanism facilitating reflection and self-control. Between 3 and 5 years vocal actions and accompanying facial expressions become suppressed and the voice becomes internalized as a silent mechanisms of self-control.
Artificial intelligence is now being recruited to re-create the kind of dialogue we have with our inner voices.
BBC Radio 4, The Digital Human – Series 11, Echo, May 2017, 5:27 minutes
However, even as adults, the nuances of facial expression leak information about what is going on in the mind, but most adults learn to distinguish between situations in which this is useful and those in which it presents some danger. Also, they can learn how to dissociate what is going on in the mind from what leaks out in the face and body language, thereby conferring the ability to deceive. (REF 7)
BBC Radio 4, Where do voices inside our heads come from?, April 2016, 5:27 minutes
Youtube Video, What Causes The Voice In Your Head?, Thoughty2, August 2015, 6:57 minutes
In the organisation, executives are only too aware that they cannot air all of their thoughts in public. Executives are ultra-careful about what they communicate, to whom and how, or they soon learn. Board meetings are often closed and communications can be deliberately targeted, sometime with the help of a communications or PR department. Wise executives rarely blurt out the first thing that comes into their heads. They inhibit that tendency and use their own thoughts to first control their own behaviour. They exercise self-control. Some private speech ends up in the boardroom, especially the closed sessions, while the public speech is crafted by the PR department. Just as in the individual this can be crafted to deceive or mislead, but also, just as in the individual, the real danger comes when the executive starts to believe its own deceptions.
Stage 5 – Management of own emotions
The individual, by resurrecting images of the past, can ‘control’ his or her own emotional states in order to be able to socialize more effectively and not drive other people away. An individual can act to put themselves in a better frame of mind, not make important decision while angry, and otherwise act to exert some control over their own emotional state.
Daniel Goleman (REF 8) in his book ‘Emotional Intelligence’ identifies 4 aspects of emotional intelligence – (1) Self Awareness (2) Self-Management (3) Empathy and (4) Relationship Management. The first two of these are regarded as key executive functions whilst empathy and relationship management extend executive function into the social sphere that do not fully develop until adulthood.
YouTube Video, Daniel Goleman Introduces Emotional Intelligence, April 2012, 5:31 minutes
Can organisations be said to have emotions? The answer is ‘yes’. Announcing profits, losses, redundancies, being given awards or a bad press can have emotional repercussions throughout the organisation. Sustained ‘moods’ can have implications for the organisation culture. Some organisations have enthusiasm and optimism while others have low morale and become depressed and dysfunctional. Some organisations feel threatened, get anxious and show some of the common human defence mechanisms such as denial, over-compensation, projection and compartmentalisation (REF 9).
Article, 15 Common Defense Mechanisms, John M. Grohol, Psy.D.
The organisational memory of emotional events in the past (both traumatic and elating) can help to manage a current situation but often relies on there still being people engaged that remember the past. Most executives are only too aware of the relationship between the mood and culture of an organisation and its performance, and act to manage the mood. Even when times are hard they convey a positive message and vision that helps take the organisation forward. However, an unrealistic representation, or a glossing over of current circumstances, risks losing the trust of the people that are the key to future success.
Stage 6 – Management of own motivations
In the individual, this is self-motivation or self-determination. Management of your own motivations frees the individual from thinking and acting in ways that have been learnt, either through practice in response to circumstances or by copying others. It opens new doors. You no longer have to be driven by habits or others expectations. You can think for yourself, determine your own goals, prioritise them as you think fit and work towards them in any way that you like.
Imagery has already developed to allow external consequences to be substituted by mental representation. Motivations can thereby be created in relation to events that are distant in space and time, and these can be reasoned about and managed without recourse to acting on the outside world.
In the organisation: Organisations specialise in the management of motivation. In particular they manage motivations with respect to profit (or at least self sustainability) but this is achieved with reference to the organisation’s mission. Many companies will have a list of strategic goals or intentions, and although profit often comes high on the list, there are others such as customer and staff satisfaction. Often separate departments take charge of these different motivations but eventually it is the executive that must coordinate them. At worst it must suppress conflicts. At best it provides orchestration, aligning motivations so that parts of the organisation support each other.
Daniel Pink in his book ‘Drive’ (REF 10) describes recent studies on how organisational incentives affect employee motivation. For routine tasks that can be performed on ‘auto-pilot’ (system 1 thinking) monitory incentives seem to work well as a means of keeping people on task and increasing performance. However, and running counter to previous views, it seems that financial incentives either do not work or actually impair performance when the work involves higher level executive functions such as reasoning and problem solving. The more effective incentives for these types of tasks are: autonomy, mastery and purpose. Autonomy means allowing employees to have the freedom to achieve goals in a manner of their own choosing, rather than having the method defined and prescribed. Mastery means giving the employee the freedom and resources to develop their own skills to a high standard, helping engender a greater degree of self-worth. Purpose refers to a socially useful purpose beyond that of the individual. It means joining with others to achieve something great, that the individual could not have achieved alone. Pink argues that the 21st Century worker must be incentivised in this way or they will not be sufficiently agile , resourceful, flexible and resilient to cope with the rapidly changing demands of a modern global economy.
YouTube Video, The puzzle of motivation | Dan Pink, TED, August 2009, 18:36 minutes
Stage 7 – Internalised Play
In the individual, the last manifestation of executive function is internalised play (REF 11). Internal play involves self-awareness and analysis, imagery, synthesis, planning, emotional and motivational control, and problem solving. It builds on all the other executive functions to allow us to take apart any object of our thoughts and re-construct them in the mind, in novel ways, to meet the needs of the moment.
YouTube Video, Learning Through Play: Developing Children’s Executive Function, Center on the Developing Child at Harvard University, September 2015, 27 seconds
In the organisation: Organisations that are big and profitable enough, make room for a lot of internal play and experimentation, only some of which will lead anywhere. Play itself allows the organisation to exercise its muscles, fine tune its processes and see where ideas might lead without heavy financial commitment.
Several references are provided below to elaborate on this and to show how play is a necessary ingredient in the development of the highest levels of executive function.
Implications for Organisational Development
Self-Awareness: Without self-awareness there is no self-control, but equally damaging is an inaccurate or biased self-awareness. Key Process Indicators (KPIs) and other management information systems can provide self-awareness but in the same way that an individual can become pre-occupied with their own inaccurate perception of themselves, an organisation can become equally distracted by KPIs that are easy to measure but are not closely aligned to its mission and strategy. It is only too easy to be deceived by the apparent objectivity of KPIs, especially when it is in the interests of different parts of the organisation to supply data that it knows the executive wants to see. In the same way that individuals tend to select the information that confirms their prejudices, an organisation can be similarly ‘blind’ to information it feels uncomfortable with.
Self-Restraint: An individual without self-restraint is often impulsive, easily distracted, lacks focus and fails to finish tasks. Too much restraint, by contrast, makes the individual inflexible and fixated. Organisations without self-restraint often pursue short-term goals at the expense of longer-term profitability. They are unable to defer gratification. By contrast, some organisations the have a tendency to over-control and tie themselves up in their own bureaucracy. Over time more and more procedure is put into place, often to correct errors of the past, until it is so rigid that it cannot respond to change. This is one reason that organisations often continuously re-structure and why some organisations seem to pulsate as control alternates between being drawn into the centre and distributed to autonomous operating units. Each process seems to run away with itself then needs to be reigned in again. The best organisations, rather than control, simply provide services to their management making it easy to carry out the functions that are central to its strategy, and more difficult to do anything else.
Imagery: The capacity to create, store and retrieve information is key to the individual in their personal development. Without the capability to learn from the past and retrieve that information when relevant an individual would act like an amnesic. An organisation without a memory of the past is similarly disoriented, and will stumble about without an understanding of what works and what doesn’t. Furthermore, without memory it is impossible to imagine what could be. Images of the past are the building blocks on which futures are built, often combining elements of the past in new ways to create novel solutions.
Theory of Mind: Understanding how customers and markets will react to events, including those events an organisation has control over, is critical in navigating and organisation through a constantly changing world. An organisation, say a government, that fails to anticipate a negative reaction to a new policy, law or budget change may find itself having to backtrack and even apologise. This is perceived as a weakness, precisely because it demonstrates that the organisation has an inadequate theory about other players. In politics in particular, it shows incompetence because politics is all about the anticipation and management of others’ reactions.
Private Speech: Private speech is more than just suppressing what is shown in public. It is the capacity to create internal dialogue and debate, to model and speculate on possible consequences of actions that have not yet been performed. The evaluation of actions before they are performed is essential to good decision making in both the individual and the organisation. Investment decisions benefit greatly from hearing a range of voices, from both within and outside the organisation, before they are acted on.
Management of Emotions: Emotions are at the route of most decision-making because they impact both an individual’s and an organisation’s priorities. The prevailing ethos of an organisation and how it affects the way staff feel, can be critical to the smooth functioning of the organisation. Some organisations have a ‘blame’ culture and, all factors being equal, any spontaneous activity on behalf of employees is suppressed. Others encourage free-thinking and innovation. How many organisations monitor these cultural and emotional factors and manage them as standing agenda items? Even when managed, most organisations are ineffective in the control of the prevailing emotional ethos and their interventions to control can easily backfire, especially if they look manipulative.
Management of Motivations: Like other functions ‘management of motivations’ can be done over-zealously or in too relaxed a manner. Self-determination is an asset so long as it does not fly in the face of circumstances. Motivations and intentions have to compromise with circumstance. The market has to be ready for your brilliant idea. Having said that, switching motivations has a cost. Even introducing a new service is expensive, let alone an entire change of strategy. Balancing the benefits of sticking to your guns with the cost of being flexible is a necessary skill. Organisations that manage to successfully grow organically achieve this balance.
Play: An individual cannot be fully functional without the opportunity to integrate all its mechanisms of executive control around the activity of play. Play allows experimentation and innovation in a safe environment, away from the dangers of the real world. Similarly, the organisation cannot be said to be fully functional without some room to play. The organisational practices of accounting for everything, monitoring every key performance indicator or extorting every last drop of employee or shareholder value, leads to organisations that are essentially reactive and immature. They are unpractised at thinking deeply at all organisational levels, and therefore lack resilience and the ability to adapt smoothly to changing circumstances. Like the individual that has failed to develop a wide range of coping strategies, they may lurch from crisis to crisis. Essentially, any change in circumstances can result in them becoming ‘out of control’.