Search company, investor...

Predict your next investment

Public-Private Partnership
BUSINESS PRODUCTS & SERVICES | Consulting & Outsourcing / Market, Industry & Competitive Research
bi.team

About Behavioural Insights Team

The Behavioural Insights Team (BIT) is a social purpose company, jointly owned by the UK Government; Nesta (the innovation charity); and its employees. The group aims to use behavioral insights to influence government policy.

Headquarters Location

4 Matthew Parker Street

London, England, SW1H 9NP,

United Kingdom

Want to inform investors similar to Behavioural Insights Team about your company?

Submit your Analyst Briefing to get in front of investors, customers, and partners on CB Insights’ platform.

Latest Behavioural Insights Team News

A manifesto for applying behavioural science

Mar 20, 2023

Abstract Recent years have seen a rapid increase in the use of behavioural science to address the priorities of public and private sector actors. There is now a vibrant ecosystem of practitioners, teams and academics building on each other’s findings across the globe. Their focus on robust evaluation means we know that this work has had an impact on important issues such as antimicrobial resistance, educational attainment and climate change. However, several critiques have also emerged; taken together, they suggest that applied behavioural science needs to evolve further over its next decade. This manifesto for the future of applied behavioural science looks at the challenges facing the field and sets out ten proposals to address them. Meeting these challenges will mean that behavioural science is better equipped to help to build policies, products and services on stronger empirical foundations—and thereby address the world’s crucial challenges. Main There has been “a remarkable increase in behavioural studies and interventions in public policy on a global scale” over the past 15 years 1 . This growth has been built on developments taking place over many preceding decades. One was the increasing empirical evidence of the importance of non-conscious drivers of behaviour. While psychologists have studied these drivers since at least as far back as the work of William James and Wilhelm Wundt in the nineteenth century, they received renewed attention from the research agenda that showed how “heuristics and biases” influence judgement and decision-making 2 . These and other studies led many psychologists to converge on dual-process theories of behaviour that proposed that rapid, intuitive and non-conscious cognitive processes sit alongside deliberative, reflective and self-aware ones 3 . These theories challenged explanations that foregrounded the role of conscious attitudes, motivations and intentions in determining actions 4 . One result was the creation of the field of behavioural economics, which developed new explanations for why observed behaviour diverged from existing economic models 5 . For example, the concept of “mental accounting” showed how people assign money to certain purposes and—contrary to standard economic theory—are reluctant to repurpose those sums, even when they might benefit from doing so 6 . Behavioural economics may represent only one strand of applied behavioural science, but it has attracted substantial attention. By the mid-2000s, these advances had an increasingly receptive audience among some governments and policymakers 7 . The publication of the book Nudge in 2008 responded to this demand by using the evidence mentioned earlier to create practical policy solutions (Box 1 ) 8 . Then, in 2010, the UK government set up its Behavioural Insights Team 9 . The creation of the Behavioural Insights Team is notable because it became “a paradigmatic example for the translation of behavioural insights into public policy” that acted as “a blueprint for the establishment of similar units elsewhere” 10 , 11 , 12 . Similar initiatives were adopted by many public sector bodies at the local, national and supra-national levels and by private companies large and small 1 , 11 , 13 , 14 . The Organisation for Economic Development and Cooperation has labelled this creation of more than 200 dedicated public entities a “paradigm shift” 15 that shows that applied behavioural science has “taken root in many ways across many countries around the world and across a wide range of sectors and policy areas” 16 . This history is necessarily selective; it does not attempt to cover the full range of work in the behavioural sciences. Rather, my focus is on the main ways that approaches often grouped under the term ‘behavioural insights’ have been applied to practical issues in the public and private sectors over the past 15 years 17 (see Box 1 for definitions of these and other terms). These approaches have been adopted in both developed and developing economies, and their precise forms of implementation have varied from context to context 18 . However, a crucial point to emphasize is that they have gone far beyond the self-imposed limits of nudges, even if that label is still used (often unhelpfully) as a blanket term. Instead, a broader agenda has emerged that explores how behavioural science can be integrated into core public and private sector activities such as regulation, taxation, strategy and operations. This broader agenda is reflected in the creation of research programmes on “behavioural public policy” 19 or “behavioural public administration” 20 . Proponents of these approaches can point to improved outcomes in many areas, including health 21 , education 22 , sustainability 23 and criminal justice 24 . Yet criticisms have emerged alongside these successes. For example, there is an ongoing debate about how publication bias may have inflated the published effect sizes of nudge interventions 25 , 26 . Other criticisms target the goals, assumptions and techniques associated with recent applications of behavioural science (Box 2 ). This Perspective attempts to respond to these criticisms by setting out an agenda to ensure that applied behavioural science can fulfil its potential in the coming decades. It does so by offering ten proposals, as summarized in Table 1 . These proposals fall into three categories: scope (the range and scale of issues to which behavioural science is applied), methods (the techniques and resources that behavioural science deploys) and values (the principles, ideals and standards of conduct that behavioural scientists adopt). These proposals are the product of a non-systematic review of relevant literature and my experience of applying behavioural science. They are not an attempt to represent expert consensus; they aim to provoke debate as well as agreement. Table 1 Summary of proposed actions to improve future applications of behavioural science Figure 1 shows how each proposal aims to address one or more of the criticisms set out in Box 2 . Figure 1 also indicates how responsibilities for implementing the proposals are allocated among four major groups in the behavioural science ecosystem: practitioners (individuals or teams who apply behavioural science findings in practical settings), the clients who commission these practitioners (for example, public or private sector organizations), academics working in the behavioural sciences (including disciplines such as anthropology, economics and sociology) and funders who support the work of these academics. These groups constitute the ‘we’ referred to in the rest of the paper, which summarizes a full-length, in-depth report available at www.bi.team . Fig. 1: Manifesto overview. The left side shows common criticisms made of the behavioural insights approach. The middle column presents ten proposals to improve the way behavioural science is applied. These proposals are organized into three categories (scope, methods and values), which are represented by red, blue and yellow, respectively. The arrows from the criticisms to the proposals show which of the latter attempt to address the former. The matrix on the right shows the four main groups involved with implementing the proposals: practitioners, clients, academics and funders. The dots in each column indicate that the relevant group will need to make a substantive contribution to achieving the goals of the proposal in the corresponding row. Box 1 Glossary of main terms Behavioural science. In its broadest sense, a discipline that uses scientific methods to generate and test theories that explain and predict the behaviour of individuals, groups and populations. This piece focuses particularly on the implications of dual-process theories of behaviour. Behavioural science is different from ‘the behavioural sciences’, which refers to a broader group of any scientific disciplines that study behaviour. Behavioural insights. The application of findings from behavioural science to analyse and address practical issues in real-world settings, usually coupled with a rigorous evaluation of the effects of any interventions. In the current piece, this term is used interchangeably with ‘applied behavioural science’. Behavioural economics. The application of findings from behavioural science to the field of economics to create explanations for economic behaviour that often diverge from the principles of neoclassical economic theory. Nudge. The design of choices so that non-conscious cognitive processes lead individuals to select the option that leaves them better off, as judged by themselves. Nudges do not involve coercion or any substantial change to economic incentives, leaving people with a meaningful ability to choose a different option from the one that the choice architect intends. Box 2 Criticisms of the behavioural insights approach Limited impact. The approach has focused on more tractable and easy-to-measure changes at the expense of bigger impacts; it has just been tinkering around the edges of fundamental problems 29 , 50 , 172 . Failure to reach scale. The approach promotes a model of experimentation followed by scaling, but it has not paid enough attention to how successful scaling happens—and the fact that it often does not happen 18 . Mechanistic thinking. The approach has promoted a simple, linear and mechanistic approach to understanding behaviour that ignores second-order effects and spillovers (and employs evaluation methods that assume a move from A to B against a static background) 29 , 62 , 173 . Flawed evidence base. The replication crisis has challenged the evidence base underpinning the behavioural insights approach, adding to existing concerns such as the duration of its interventions’ effects 79 , 174 . Lack of precision. The approach lacks the ability to construct precise interventions and establish what works for whom, and when. Instead, it relies either on overgeneral frameworks or on disconnected lists of biases 80 , 92 , 94 . Overconfidence. The approach can encourage overconfidence and overextrapolation from its evidence base, particularly when testing is not an option 175 . Control paradigm. The approach is elitist and pays insufficient attention to people’s own goals and strategies; it uses concepts such as irrationality to justify attempts to control the behaviour of individuals, since they lack the means to do so themselves 176 , 177 . Neglect of the social context. The approach has a limited, overly cognitive and individualistic view of behaviour that neglects the reality that humans are embedded in established societies and practices 125 , 178 , 179 . Ethical concerns. The behavioural insights approach will face more ethics, transparency and privacy conundrums as it attempts more ambitious and innovative work 143 , 145 , 154 . Homogeneity of participants and perspectives. The range of participants in behavioural science research has been narrow and unrepresentative 164 ; homogeneity in the locations and personal characteristics of behavioural scientists influences their viewpoints, practices and theories 124 , 166 . Scope Use behavioural science as a lens The early phase of the behavioural insights movement was marked by scepticism about whether effects obtained in laboratories would translate to real-world settings 27 . In response, practitioners developed standard approaches that could demonstrate a clear causal link between an intervention and an outcome 28 . In practice, these approaches directed attention towards how the design of specific aspects of a policy, product or service influences discrete behaviours by actors who are considered mostly in isolation 29 . These standard approaches are strong and have produced valuable results in many contexts around the world 20 , 30 . However, in the aggregate, they have also fostered a perspective centred on the metaphor of behavioural science as a specialist tool. This view mostly limits behavioural science to the role of fixing concrete aspects of predetermined interventions rather than aiding the consideration of broader policy goals 31 . Over time, this view has created a self-reinforcing perception that only certain kinds of tasks are suitable for behavioural scientists 29 . Opportunities, skills and ambitions have been constricted as a result; a rebalancing is needed. Behavioural science also has much to say about pressing societal issues such as discrimination, pollution and economic mobility and the structures that produce them 32 , 33 . These ambitions have always been present in the behavioural insights movement 34 , but the factors just outlined acted against their being realized more fully 35 . The first step towards achieving these ambitions is to replace the dominant metaphor of behavioural science as a tool. Instead, behavioural science should be understood as a lens that can be applied to any public or private issue. This change offers several advantages: A lens metaphor shows that behavioural science can enhance the use of standard policy options (for example, revealing new ways of structuring taxes) rather than just acting as an alternative to them. A lens metaphor conveys that the uses of behavioural science are not limited to creating new interventions. A behavioural science lens can, for example, help to reassess existing actions and understand how they may have unintended effects. It emphasizes the behavioural diagnosis of a situation or issue rather than pushing too soon to define a precise target outcome and intervention 31 . Specifying that this lens can be applied to any action conveys the error of separating ‘behavioural’ and ‘non-behavioural’ issues: most of the goals of private and public action depend on certain behaviours happening (or not). Behavioural science should therefore be integrated into an organization’s core activities rather than acting as an optional specialist tool 36 . It may seem odd to start with a change of metaphor, but the primary problem here is one of perception. Behavioural science itself shows us the power of framing: the metaphors we use shape the way we behave and therefore can be agents of change 37 . Metaphors are particularly important in this case because the task of broadening the use of behavioural science requires making a compelling case to decision makers 38 . The metaphor of behavioural science as a tool has established credibility and acceptance in a defined area; expanding beyond that area is the task for the next decade. Build behavioural science into organizations The second proposal is to broaden the scope of how behavioural science is used in organizations. Given that many dedicated behavioural science teams exist worldwide, it is understandable that much attention has been paid to the question of how they should be set up successfully. However, this focus has diverted attention from considering how to use behavioural science to shape organizations themselves 39 . We need to talk less about how to set up a dedicated behavioural science team and more about how behavioural science can be integrated into an organization’s standard processes. For example, as well as trying to ensure that a departmental budget includes provisions for behavioural science, why not use behavioural science to improve the way this budget is created (for example, are managers anchored to outdated spending assumptions) 40 ? The overriding message here is for greater focus on the organizational changes that indirectly apply or support behavioural science principles, rather than just thinking through how the direct and overt use of behavioural science can be promoted in an organization. One advantage to this approach is that it can help organizations to address problems with scaling interventions 36 . If some of the barriers to scaling concern cognitive biases in organizations, these changes could minimize the effect of such biases 41 . Rather than starting with a behavioural science project and then trying to scale it, we could start by looking at operations at scale and understanding how they can be influenced. It is useful to understand how this approach maps onto existing debates about how to set up a behavioural function in organizations. Doing so reveals six main scenarios, as shown in Table 2 . In the ‘baseline’ scenario, there is limited awareness of behavioural science in the organization, and its principles are not incorporated into processes. In the ‘nudged organization’, behavioural science awareness is still low, but its principles have been used to redesign processes to create better outcomes for staff or service users. In ‘proactive consultancy’, leaders may have set up a dedicated behavioural team without grafting it onto the organization’s standard processes. This lack of institutional grounding puts the team in a less resilient position, meaning that it must always search for new work. In ‘call for the experts’, an organization has concentrated behavioural expertise, but there are also prompts and resources that allow this expertise to be integrated into business as usual. Expertise is not widespread, but access to it is. Processes stimulate demand for behavioural expertise that the central team can fulfil. In ‘behavioural entrepreneurs’, there is behavioural science capacity distributed throughout the organization, through either direct capacity building or recruitment. The problem is that organizational processes do not support these individual pockets of knowledge. Finally, a ‘behaviourally enabled organization’ is one where there is knowledge of behavioural science diffused throughout the organization, which also has processes that reflect this knowledge and support its deployment. Table 2 Options for building behavioural science into organizations Most discussions make it seem like the meaningful choice is between the different columns in Table 2 —how to organize dedicated behavioural science resources. Instead, the more important move is from the top row to the bottom row: moving from projects to processes, from commissions to culture. A useful way of thinking about this task is about building or upgrading the “choice infrastructure” of the organization 42 . In other words, we should place greater focus on the institutional conditions and connections that support the direct and indirect ways that behavioural science can infuse organizations. Working out how best to build the choice infrastructure in organizations should be a major priority for applied behavioural science. Already we can see that some features will be crucial: reducing the costs of experimentation, creating a system that can learn from its actions, and developing new and better ways of using behavioural science principles to analyse the behavioural effects of organizational processes, rules, incentives, metrics and guidelines 36 . See the system Many important policy challenges emerge from complex adaptive systems, where change often does not happen in a linear or easily predictable way, and where coherent behaviour can emerge from interactions without top-down direction 43 . There are many examples of such systems in human societies, including cities, markets and political movements 44 . These systems can create “wicked problems”—such as the COVID-19 pandemic—where ideas of success are contested, changes are nonlinear and difficult to model, and policies have unintended consequences 45 . This reality challenges the dominant behavioural science approach, which usually assumes stability over time, keeps a tight focus on predefined target behaviours and predicts linear effects on the basis of a predetermined theory of change 46 . The result, some argue, is a failure to understand how actors are acting and reacting in a complex system that leads policymakers to conclude they are being irrational—and then actually disrupt the system in misguided attempts to correct perceived biases or inefficiencies 47 , 48 , 49 . These criticisms may overstate the case, but they point to a way forward. Behavioural science can be improved by using aspects of complexity thinking to offer new, credible and practical ways of addressing major policy issues. The first step is to reject crude distinctions of ‘upstream’ versus ‘downstream’ or the ‘individual frame’ versus the ‘system frame’ 50 . Instead, complex adaptive systems show that higher-level features of a system can actually emerge from the lower-level interactions of actors participating in the system 44 . When they become the governing features of the system, they then shape the lower-level behaviour until some other aspect emerges, and the fluctuations continue. An example might be the way that new coronavirus variants emerged in particular settings and then went on to change the course of the whole pandemic, requiring new overall strategic responses. In other words, we are dealing with “cross-scale behaviours” 49 . For example, norms, rules, practices and culture itself can emerge from aggregated social interactions; these features then shape cognition and behavioural patterns in turn 51 . Recognizing cross-scale behaviours means that behavioural science could: Identify “leverage points” where a specific shift in behaviour will produce wider system effects 52 . One option is to identify when and where tipping points are likely to occur in a system and then either nudge them to occur or not, depending on the policy goal 53 . For example, if even a subset of consumers decides to switch to a healthier version of a food product, this can have broader effects on a population’s health through the way the food system responds by restocking and product reformulation 54 . Model the collective implications of individuals using simple heuristics to navigate a system. For example, new models show how small changes to simple heuristics that guide savings (in this case, how quickly households copy the savings behaviours of neighbours) can lead to the sudden emergence of inequalities in wealth 55 . Find targeted changes to features of a system that create the conditions for wide-ranging shifts in behaviour to occur. For example, a core driver of social media behaviours is the ease with which information can be shared 46 . Even minor changes to this parameter can drive widespread changes—some have argued that such a change is what created the conditions leading to the Arab Spring, for example 56 . This approach also suggests that a broader change in perspective is needed. We need to realize the flaws in launching interventions in isolation and then moving on when a narrowly defined goal has been achieved. Instead, we need to see the longer-term impact on a system of a collection of different policies with varying goals 57 . The best approach may be “system stewardship”, which focuses on creating the conditions for behaviours and indirectly steering adaptation towards overall goals 58 . Of course, not every problem will involve a complex adaptive system; for simple issues, standard approaches to applying behavioural science work well. Behavioural scientists should therefore develop the skills to recognize the type of system that they are facing (see the system) and then choose their approach accordingly. These skills can be developed through agent-based simulations 59 , immersive technologies 60 or just basic checklists 61 . Methods Put randomized controlled trials in their place Randomized controlled trials (RCTs) have been a core part of applied behavioural science, and they work well in relatively simple and stable contexts. But they can fare worse in complex adaptive systems, whose many shifting connections can make it difficult to keep a control group isolated and where a narrow focus on predetermined outcomes may neglect others that are important but difficult to predict 43 , 62 . We can strengthen RCTs to deal better with complexity. We can try to gain a better understanding of the system interactions and anticipate how they may play out, perhaps through “dark logic” exercises that try to trace potential harms rather than just benefits 63 . For example, we might anticipate that sending parents text messages encouraging them to talk to their children about the school science curriculum may achieve this outcome at the expense of other school-supporting behaviours—as turned out to be the case 64 . Engaging the people who will implement and participate in an intervention will be a key part of this effort. Another option is to set up RCTs to measure diffusion and contagion in networks, either by creating separate online environments or by randomizing real-world clusters, such as separate villages 65 , 66 . Finally, we can build feedback and adaptation into the design of the RCT and the intervention, allowing adjustments to changing conditions 67 , 68 . Options include using two-stage trial protocols 69 , evolutionary RCTs 70 , sequential multiple assignment randomized trials 71 and ‘bandit’ algorithms that identify high-performing interventions and allocate more people to them 72 . Behavioural science can also be used to enhance alternative ways of measuring impacts—in particular, agent-based modelling, which tries to simulate the interactions between the different actors in a system 73 . The agents in these models are mostly assumed to be operating on rational choice principles 74 , 75 . There is therefore an opportunity to build in more evidence about the drivers of behaviour—for example, habits and social comparisons 49 . Replication, variation and adaptation The ‘replication crisis’ of the past decade has seen intense debate and concern about the reliability of behavioural science findings. Poor research practices were a major cause of the replication crisis; the good news is that many have improved as a result 76 , 77 . Now there are sharper incentives to preregister analysis plans, greater expectations that data and code will be freely shared, and wider acceptance of post-publication review of findings 78 . Behavioural scientists need to secure and build on these advances to move towards a future where appropriately scoped meta-analyses of high-quality studies (including deliberate replications) are used to identify the most reliable interventions, develop an accurate sense of the likely size of their effects and avoid the weaker options. We have a responsibility to discard ideas if solid evidence now shows that they are shaky, and to offer a realistic view of what behavioural science can accomplish 18 . That responsibility also requires us to have a hard conversation about heterogeneity in results: the complexity of human behaviour creates so much statistical noise that it is often hard to detect consistent signals and patterns 79 . The main drivers of heterogeneity are that contexts influence results and that the effect of an intervention may vary greatly between groups within a population 80 , 81 . For example, choices of how to set up experiments vary greatly between studies and researchers, in ways that often go unnoticed 82 . A recent study ran an experiment to measure the impact of these contextual factors. Participants were randomly allocated to studies designed by different research teams to test the same hypothesis. For four of the five research questions, studies actually produced effects in opposing directions. These “radically dispersed” results indicate that “idiosyncratic choices in stimulus design have a very large effect on observed results” 83 . These factors complicate the idea of replication itself: a ‘failed’ replication may not show that a finding was false but rather show how it exists under some conditions and not others 84 . These challenges mean that applied behavioural scientists need to set a much higher bar for claiming that an effect holds true across many unspecified settings 85 . There is a growing sense that interventions should be talked about as hypotheses that were true in one place and that may need adapting to be true elsewhere 18 , 86 . Narrative changes need to be complemented by specific proposals. The first concerns data collection: behavioural scientists should expand studies to include (and thus examine) a wider range of contexts and participants and gather richer data about them. To date, only a small minority of behavioural studies have provided enough information to see how effects vary 87 . Moreover, the gaps in data coverage may result from and create systemic issues in society: certain groups may be excluded or may have their data recorded differently from others 88 . Coordinated multi-site studies will be needed to collect enough data to explore heterogeneity systematically; crowdsourced studies offer particular promise for testing context and methods 83 . Realistically, this work is going to require a major investment in research infrastructure to set up standing panels of participants, coordinate between institutions, and reduce barriers to data collection and transfer 80 . These efforts cannot be limited to just a few countries. Behavioural scientists also need to get better at judging how strongly an intervention’s results were linked to its context and therefore how much adaptation it needs 81 . We should use and modify frameworks from implementation science to develop such judgement 89 . Finally, we need to codify and cultivate the practical skills that successfully adapt interventions to new contexts; expertise in behavioural science should not be seen as simply knowing about concepts and findings in the abstract. It is therefore particularly valuable to learn from practitioners how they adapted specific interventions to new contexts. These accounts are starting to emerge, but they are still rare 18 , since researchers are incentivized to claim universality for their results rather than report and value contextual details 82 . Beyond lists of biases The heterogeneity in behavioural science findings also means that our underlying theories need to improve: we are lacking good explanations for why findings vary so much 84 . This need for better theories can be seen as part of a wider ‘theory crisis’ in psychology, which has thrown up two big concerns for behavioural science 90 , 91 . The first stems from the fact that theories of behaviour often try to explain phenomena that are complex and wide-ranging 92 . If you are trying to show how emotion and cognition interact (for example), this involves many causes and interactions. Trying to cover this variability can produce descriptions of relationships and definitions of constructs that are abstract and imprecise 85 . The result is theories that are vague and weak, since they can be used to generate many different hypotheses—some of which may actually contradict each other 90 . That makes theories hard to disprove, and so weak theories stumble on, unimproved 93 . The other concern is that theories can make specific predictions, but they are disconnected from each other—and from a deeper, general framework that can provide broader explanations (such as evolutionary theory) 94 . The main way this issue affects behavioural science is through heuristics and biases. Examples of individual biases are accessible, popular and how many people first encounter behavioural science. These ideas are incredibly useful, but they have often been presented as lists of standalone curiosities in a way that is incoherent, reductive and deadening. Presenting lists of biases does not help us to distinguish or organize them 95 , 96 , 97 . Such lists can also create overconfident thinking that targeting a specific bias (in isolation) will achieve a certain outcome 98 . Perhaps most importantly, focusing on lists of biases distracts us from answering core underlying questions. When does one or another bias apply? Which are widely applicable, and which are highly specific? How does culture or life experience affect whether a bias influences behaviour or not 99 , 100 ? These are highly practical questions when one is faced with tasks such as taking an intervention to new places. The concern for behavioural science is that it uses both these high-level frameworks (such as dual-process theories) and jumbled collections of heuristics and biases, with little in the middle to draw both levels together 94 . Recent years have seen valuable advances in connecting and systematizing theories 101 , 102 . At the same time, there are various ongoing attempts to create strong theories: “coherent and useful conceptual frameworks into which existing knowledge can be integrated” 93 (see also refs. 91 , 103 , 104 ). Naturally, such work should continue, but I think that applied behavioural science will benefit particularly from theories that are practical. By this I mean: They fill the gap between day-to-day working hypotheses and comprehensive and systematic attempts to find universal underlying explanations. They are based on data rather than being derived from pure theorizing 105 . They can generate testable hypotheses, so they can be disproved 106 . They specify the conditions under which a prediction applies or does not 85 . They are geared towards realistic adaptation by practitioners and offer “actionable steps toward solving a problem that currently exists in a particular context in the real world” 107 . Resource rationality may be a good example of a practical theory. It starts from the basis that people make rational use of their limited cognitive resources 108 . Given that there is a cost to thinking, people will look for solutions that balance choice quality with effort. Resource rationality can offer a “unifying framework for a wide range of successful models of seemingly unrelated phenomena and cognitive biases” that can be used to build models for how people act 108 . A recent study has shown how these models not only can predict how people will respond to different kinds of nudges in certain contexts but also can be integrated with machine learning to create an automated method for constructing “optimal nudges” 109 . Such an approach could reveal new kinds of nudges and make creating them much more efficient. More reliable ways of developing personalized nudges are also possible. These are all highly practical benefits coming from applying a particular theory. Predict and adjust Hindsight bias is what happens when we feel ‘I knew it all along’, even if we did not 110 . When the results of an experiment come in, hindsight bias may mean that behavioural scientists are more likely to think that they had predicted them or quickly find ways of explaining why they occurred. Hindsight bias is a big problem because it breeds overconfidence, impedes learning, dissuades innovation and prevents us from understanding what is truly unexpected 111 , 112 . In response, behavioural scientists should establish a standard practice of predicting the results of experiments and then receiving feedback on how their predictions performed. Hindsight bias can flourish if we do not systematically capture expectations or priors about what the results of a study will be 113 . Making predictions provides regular, clear feedback of the kind that is more likely to trigger surprise and reassessment rather than hindsight bias 114 . Establishing the average expert prediction—which may be different from the null hypothesis in an experiment—clearly reveals when results challenge the consensus 115 . There are existing practices to build on here, such as the practice of preregistering hypotheses and trial protocols and the use of a Bayesian approach to make priors explicit. Indeed, more and more studies are explicitly integrating predictions 116 , 117 . However, barriers lie in the way of further progress. People may not welcome the ensuing challenge to their self-image, predicting may seem like one thing too many on the to-do list, and the benefits lie in the future. Some responses to these challenges are to make predicting easy by incorporating it into standard processes; minimize threats to predictors’ self-image (for example, by making and feeding back predictions anonymously) 118 ; give concrete prompts for learning and reflection, to disrupt the move from surprise to hindsight bias 119 ; and build learning from prediction within and between institutions. Values Be humble, explore and enable This proposal is made up of three connected ideas. First, behavioural scientists need to become more aware of the limits of their knowledge and to avoid fitting behaviours into pre-existing ideas around biases or irrationality. Second, they should broaden the exploratory work they conduct, in terms of gaining new types of qualitative data and recognizing how experiences vary by group and geography. Finally, they should develop new approaches to enable people to apply behavioural science themselves—and adopt new criteria for judging when these approaches are appropriate. Humility is important because behavioural scientists (like other experts) may overconfidently rely on decontextualized principles that do not match the real-world setting for a behaviour 29 . Deeper inquiry can reveal reasonable explanations for what seem to be behavioural biases 120 . In response, those applying behavioural science should avoid using the term ‘irrationality’, which can limit attempts to understand actions in context; acknowledge that diagnoses of behaviour are provisional and incomplete (epistemic humility) 121 ; and design processes and institutions to counteract overconfidence 122 . How do we conduct these deeper inquiries? Three areas demand particular focus in the future. First, pay greater attention to people’s goals and strategies and their own interpretations of their beliefs, feelings and behaviours 123 . Second, reach a wider range of experiences, including marginalized voices and communities, understanding how structural inequalities can lead to expectations and experiences varying greatly by group and geography 124 . Third, recognize how apparently universal cognitive processes are shaped by specific contexts, thereby unlocking new ways for behavioural science to engage with values and culture 125 , 126 . For example, one influential view of culture is that it influences action “not by providing the ultimate values toward which action is oriented but by shaping a repertoire or ‘toolkit’ of habits, skills, and styles” 127 . There are similarities here to the heuristics-and-biases toolkit perspective on behaviour: behavioural scientists could start explaining how and when certain parts of the toolkit become more or less salient. More can and should be done to broaden ownership of behavioural science approaches. Many (but far from all) behavioural science applications have been top-down, with a choice architect enabling certain outcomes 8 , 128 . One route is to enable people to become more involved in designing interventions that affect them—and “nudge plus” 129 , “self-nudges” 130 and “boosts” 131 have been proposed as ways of doing this. Reliable criteria are needed to decide when enabling approaches may be appropriate, including whether the opportunity to use an enabling approach exists; ability and motivation; preferences; learning and setup costs; equity impacts; and effectiveness, recognizing that evidence on this point is still emerging 132 , 133 . But these new approaches should not be seen simplistically as enabling alternatives to disempowering nudges 134 . Instead, we need to consider how far the person performing the behaviour is involved in shaping the initiative itself, as well as the level and nature of any capacity created by the intervention. People may be heavily engaged in selecting and developing a nudge intervention that nonetheless does not trigger any reflection or build any skills 135 . Alternatively, a policymaker may have paternalistically assumed that people want to build up their capacity to perform an action, when in fact they do not. This is the real choice to be made. A final piece missing from current thinking is that enabling people can lead to a major decentring of the use of behavioural science. If more people are enabled to use behavioural science, they may decide to introduce interventions that influence others 136 . Rather than just creating self-nudges through altering their immediate environments, they may decide that wider system changes are needed instead. A range of people could be enabled to create nudges that generate positive societal change (with no central actors involved). This points towards a future where policy or product designers act less like (choice) architects and more like facilitators, brokers and partnership builders 137 . Data science for equity Recent years have seen growing interest in using new data science techniques to reliably analyse the heterogeneity of large datasets 138 , 139 . Machine learning is claimed to offer more sophisticated, reliable and data-driven ways of detecting meaningful patterns in datasets 140 , 141 . For example, a machine learning approach has been shown to be more effective than conventional segmentation approaches at analysing patterns of US household energy usage to reduce peak consumption 142 . A popular idea is to use such techniques to better understand what works best for certain groups and thereby tailor an offering to them 143 . Scaling an intervention stops being about a uniform roll-out and instead becomes about presenting recipients with the aspects that are most effective for them 144 . This vision is often presented as straightforward and obviously desirable, but it runs almost immediately into ethical quandaries and value judgements. People are unlikely to know what data have been used to target them and how; the specificity of the data involved may make manipulation more likely, since it may exploit sensitive personal vulnerabilities; and expectations of universality and non-discrimination in public services may be violated 143 , 145 . Closely related to manipulation concerns is the fear that data science will open up new opportunities to exploit, rather than to help, the vulnerable 146 . One aspect is algorithmic bias. Models using data that reflect historical patterns of discrimination can produce results that reinforce these outcomes 147 . Since disadvantaged groups are more likely to be subject to the decisions of algorithms, there is a particular risk that inequalities will be perpetuated—although some studies argue that algorithms are actually less likely to be biased than human judgement 148 , 149 . There is also emerging evidence that people often object to personalization. While they support some personalized services, they consistently oppose advertising that is customized on the basis of sensitive information—and they are generally against the collection of the information that personalization relies on 150 . To navigate this landscape, behavioural scientists need to examine four factors: Who does the personalization target, and using what criteria? Many places have laws or norms to ensure equal treatment based on personal characteristics. When does personalization violate those principles? How is the intervention constructed? To what extent do the recipients have awareness of the personalization, choice over whether it occurs, control over its level or nature, and the opportunity to give feedback on it 151 ? When is it directed? Is it at a time when the participant is vulnerable? Would they probably regret it later, if they had time to reflect? Why is personalization happening? Does it aim to exploit and harm or to support and protect, recognizing that those terms are often contested? Taking these factors into account, I propose that the main opportunity is for data science to identify the ways in which an intervention or situation appears to increase inequalities, and reduce them 152 . For example, groups that are particularly likely to miss a filing requirement could be offered pre-emptive help. Algorithms can be used to better explain the causes of increased knee pain experienced in disadvantaged communities, thereby giving physicians better information to act on 153 . I call this idea data science for equity. It addresses the ‘why’ factor by using data science to support, not exploit. ‘Data science for equity’ may seem like a platitude, but it is a very real choice: the combination of behavioural and data science is powerful and has been used to create harm in the past. Moreover, it needs to be complemented by attempts to increase agency (the ‘how’ factors), as in a recent study that showed how boosts can be used to help people to detect micro-targeting of advertising 154 , and studies that obtain more data on which uses of personalization people find acceptable. No “view from nowhere” The final proposal is one of the most wide-ranging, challenging and important. For the philosopher Thomas Nagel, the “view from nowhere” was an objective stance that allowed us to “transcend our particular viewpoint” 155 . Taking such a stance may not be possible for behavioural scientists. We bring certain assumptions and ways of seeing to what we do; we are always situated in, embedded in and entangled with ideas and situations 124 . We cannot assume that there is some set-aside position from which to observe the behaviour of others; no objective observation deck outside society exists 156 . Behavioural scientists are defined by having knowledge, skills and education; many of them can use these resources to shape public and private actions. They are therefore in a privileged position, but they may not see the extent to which they hold elite positions that stop them from understanding people who think differently (for example, those who are sceptical of education) 157 . The danger is that elites place their group values and preferences on others, while thinking that they are adopting a view from nowhere 158 , 159 . This does not mean that they can never act or opine, but rather that they need to carefully understand their own positionality and those of others before doing so. There have been repeated concerns that the field is still highly homogeneous in other ways as well. Gender, race, physical abilities, sexuality and geography also influence the viewpoints, practices and theories of behavioural scientists 160 , 161 . Only a quarter of the behavioural insights teams catalogued in a 2020 survey were based in the Global South 162 . An over-reliance on using English in cognitive science has led to the impact of language on thought being underestimated 163 . The past decade has shown how behaviours can vary greatly from culture to culture, even as psychology has tended to generalize from relatively small and unrepresentative samples 164 . Behavioural science studies often present data from Western, educated, industrialized, rich and democratic samples as more generalizable to humans as a whole 165 . So, rather than claiming that science is value-free, we need to find realistic ways of acknowledging and improving this reality 166 . A starting point is for behavioural scientists to cultivate self-scrutiny by querying how their identities and experiences contribute to their stance on a topic. Hypothesis generation could particularly benefit from this exercise, since arguably it is closely informed by the researcher’s personal priorities and preferences 167 . Behavioural scientists could be actively reflecting on interventions in progress, including what factors are contributing to power dynamics 168 . Self-scrutiny may not be enough. We should also find more ways for people to judge researchers and decide whether they want to participate in research—going beyond consent forms. If they do participate, there are many opportunities to combine behavioural science with co-design 128 . Finally, we should take actions to increase diversity (of several kinds) among behavioural scientists, teams, collaborations and institutions. Doing this requires addressing barriers such as the lack of professional networks connecting the Global North and Global South, and the time needed to build understanding of the tactics required to write successful grant applications from funders 169 . In many countries, much more could be done to increase the ethnic and racial diversity of the behavioural science field—for example, through support for starting and completing PhDs or through reducing the substantial racial gaps present in much public funding of research 170 , 171 . Conclusion Applied behavioural science has seen rapid growth and meaningful achievements over the past decade. Although the popularity of nudging provided its initial impetus, an ambition soon formed to apply a broader range of techniques to a wider range of goals. However, a set of credible critiques have emerged as levels of activity have grown. As Fig. 1 indicates, there are proposals that can address these critiques (and progress is already being made on some of them). When considered together, these proposals present a coherent vision for the scope, methods and values of applied behavioural science. This vision is not limited to technical enhancements for the field; it also covers questions of epistemology, identity, politics and praxis. A common theme throughout the ten proposals is the need for self-reflective practice that is aware of how its knowledge and approaches have originated and how they are situated. In other words, a main priority for behavioural scientists is to recognize the various ways that their own behaviour is being shaped by structural, institutional, environmental and cognitive factors. Realizing these proposals will require sustained work and experiencing the discomfort of disrupting what may have become familiar and comfortable practices. That is a particular problem because incentives for change are often weak or absent. Improving applied behavioural science has some characteristics of a social dilemma: the benefits are diffused across the field as a whole, while the costs fall on any individual party who chooses to act (or act first). Practitioners are often in competition. Academics often want to establish a distinctive research agenda. Commissioners are often rewarded for risk aversion. Impaired coordination is particularly problematic because coordination forms the basis for several necessary actions, such as the multi-site studies to measure heterogeneity. Solving these problems will be hard. Funders need to find mechanisms that adequately reward coordination and collaboration by recognizing the true costs involved. Practitioners need to perceive the competitive advantages of adopting new practices and be able to communicate them to clients. Clients themselves need to have a realistic sense of what can be achieved but still be motivated to commit resources. Stepping back, the starting point for addressing these barriers needs to be a change in the narrative about what the field does and could do—a new set of ambitions to aim for. This manifesto aims to help to shape such a narrative. References Straßheim, H. The rise and spread of behavioral public policy: an opportunity for critical research and self-reflection. Int. Rev. Public Policy 2, 115–128 (2020). Gilovich, T., Griffin, D. & Kahneman, D. (eds) Heuristics and Biases: The Psychology of Intuitive Judgment (Cambridge Univ. Press, 2002). Lieberman, M. D. Social cognitive neuroscience: a review of core processes. Annu. Rev. Psychol. 58, 259–289 (2007). Thaler, R. Mental accounting and consumer choice. Mark. Sci. 4, 199–214 (1985). Thaler, R. & Sunstein, S. Nudge: Improving Decisions about Health, Wealth, and Happiness (Yale Univ. Press, 2008). Hallsworth, M. & Kirkman, E. Behavioral Insights (MIT Press, 2020). Oliver, A. The Origins of Behavioural Public Policy (Cambridge Univ. Press, 2017). Straßheim, H. & Beck, S. in Handbook of Behavioural Change and Public Policy (eds Beck, S. & Straßheim, H.) 1–22 (Edward Elgar, 2019). Ball, S. & Feitsma, J. The boundaries of behavioural insights: observations from two ethnographic studies. Evid. Policy 16, 559–577 (2020). Competing interests The author is the managing director, Americas, at the Behavioural Insights Team, which provides consultancy services in behavioural science. Peer review Peer review information Nature Human Behaviour thanks Peter John and the other, anonymous, reviewer(s) for their contribution to the peer review of this work. Additional information Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations. Rights and permissions Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Behavioural Insights Team Team

1 Team Member

Behavioural Insights Team has 1 team member, including current Chief Executive Officer, David Halpern.

Name

Work History

Title

Status

David Halpern

Chief Executive Officer

Current

Name

David Halpern

Work History

Title

Chief Executive Officer

Status

Current

Discover the right solution for your team

The CB Insights tech market intelligence platform analyzes millions of data points on vendors, products, partnerships, and patents to help your team find their next technology solution.

Request a demo

CBI websites generally use certain cookies to enable better interactions with our sites and services. Use of these cookies, which may be stored on your device, permits us to improve and customize your experience. You can read more about your cookie choices at our privacy policy here. By continuing to use this site you are consenting to these choices.