Unconscious Rules of Thumb

Unconscious Rules of Thumb

Some of the decisions that I make are based on thorough calculations, analysis, evaluation of available options, and deliberate considerations of costs and benefits. When I am planning my workout routine, I think hard about how my legs have been feeling and what distance, elevation, and pace is reasonable for my upcoming workouts. I think about how early I need to be out the door for a certain distance, and whether I can run someplace new to mix things up. I’ll map out routes, look at my training log for the last few weeks, and try putting together a plan that maximizes my enjoyment, physical health, and fitness given time constraints.

 

However, outside of running, most of my decisions are generally based on rules of thumb and don’t receive the same level of attention as my running plans. I budget every two weeks around payday, but even when budgeting, I mostly rely on rules of thumb. There is a certain amount I like to keep in my checking account just in case I forgot a bill or have something pop-up last minute. Its not a deliberate calculation, it is more of a gut feeling. The same goes for how much money I set aside for free spending or if I feel that it is finally time to get that thing I have had my eye on for a while. My budget is probably more important than my running routine, but I actually spend more time rationally developing a running plan than I spend budgeting. The same goes for house and vehicle maintenance, spending time with friends and family, and choosing what to eat on the days we plan to do take-out.

 

The budget example is interesting because I am consciously and deliberately using rules of thumb to determine how my wife and I will use our money. I set aside a certain amount for gas without going to each vehicle and checking whether we are going to need to fill up soon. I am aware of the rules of thumb, and they are literally built into my spreadsheet where I sometimes ask if I should deviate, but usually decide to stick to them.

 

I also recognize that I have many unconscious rules of thumb. In his book Risky Savvy, Gerd Gigerenzer writes the following about unconscious rules of thumb:

 

“Every rule of thumb I am aware of can be used consciously and unconsciously. If it is used unconsciously, the resulting judgment is called intuitive. An intuition, or gut feeling, is a judgment:
  1. that appears quickly in consciousness,
  2. whose underlying reasons we are not fully aware of, yet
  3. is strong enough to act upon.”
I have lots of intuitive judgements that I often don’t think about in the moment, but only realize when I reflect back on how I do something. When I am driving down the freeway, cooking, or writing a blog post, many of my decisions flow naturally and quickly. In the moment the decisions seem obvious, and I don’t have to think too deliberately about my action and why I am making a specific decision. But if I were asked to explain why I made a decision, I would have a hard time finding exact reasons for my choices. I don’t know exactly how I know to change lanes at a certain point on the freeway, but I know I can often anticipate points where traffic will slow down, and where I might be better off in another lane. I can’t tell you why I chose to add the marsala wine to the mushrooms at the precise moment that I did. I also couldn’t explain why I chose to present a certain quote right at the beginning of a post rather than in the middle. My answer for all of these situations would simply be that it felt right.

 

We use unconscious rules of thumb like these all the time, but we don’t often notice when we do. When we are budgeting we might recognize our rules of thumb and be able to explain them, but our unconscious rules of thumb are harder to identify and explain. Nevertheless, they still have major impacts in our lives. Simply because we don’t notice them and can’t explain them doesn’t mean they don’t shape a lot of our decisions and don’t matter. The intuitions we have can be powerful and helpful, but they could also be wrong (maybe all this time I’ve been overcooking the mushrooms and should add the wine sooner!). Because these intuitions are unconscious, we don’t deliberately question them, unless something calls them up to the conscious level. The feedback we get is probably indirect, meaning that we won’t consciously tie our outcomes the to the unconscious rules of thumb that got us to them.

 

I am fascinated by things like unconscious rules of thumb because they reveal how little we actually control in our lives. We are the ones who act on these unconscious rules of thumb, but in a sense, we are not really doing anything at all. We are making decisions based on factors we don’t understand and might not be aware of. We have agency by being the one with the intuition, but we also lack agency by not being fully conscious of the how and why behind our own decisions. This should make us question ourselves and choices more than we typically do.
Probability is Multifaceted

Probability is Multifaceted

For five years my wife and I lived in a house that was at the base of the lee side of a small mountain range in Northern Nevada. When a storm would come through the area it would have to make it over a couple of small mountain ranges and valleys before getting to our house, and as a result we experienced less precipitation at our house than most people in the Reno/Sparks area. Now my wife and I live in a house higher up on a different mountain that is more in the direct path of storms coming from the west. We receive snow at our house while my parents and family lower in the valley barely get any wind. At both houses we have learned to adjust our expectations for precipitation relative to the probabilities reported by weather stations which reference the airport at the valley floor. Our experiences with rain and snow at our two places is a useful demonstration that probability (in this case the probability of precipitation) is multifaceted – that multiple factors  play a role in the probability of a given event at a given place and time.

 

In his book Risk Savvy, Gerd Gigerenzer writes, “Probability is not one of a kind; it was born with three faces: frequency, physical design, and degrees of belief.” Gigerenzer explains that frequency is about counting. To me, this is the most clearly understandable aspect of probability, and what we usually refer to when we discuss probability. On how many days does it usually rain in Reno each year? How frequently does a high school team from Northern Nevada win a state championship and how frequently does a team from Southern Nevada win a state championship? These types of questions simply require counting to give us a general probability of an event happening.

 

But probability is not just about counting and tallying events. Physical design plays a role as well. Our house on the lee side of a small mountain range was shielded from precipitation, so while it may have rained in the valley half a mile away, we didn’t get any precipitation. Conversely, our current home is in a position to get more precipitation than the rest of the region. In high school sports, fewer kids live in Reno/Sparks compared to the Las Vegas region, so in terms of physical design, state championships are likely to be more common for high schools in Southern Nevada. Additionally, there may be differences in the density of students at each school, meaning the North could have more schools per students than the south, also influencing the probability of a north or south school winning. Probability, Gigerenzer explains, can be impacted by the physical design of systems, potentially making the statistics and chance more complicated to understand.

 

Finally, degrees of belief play a role in how we comprehend probability. Gigerenzer states that degrees of belief include experience and personal impression which are very subjective. Trusting two eye witnesses, Gigerenzer explains, rather than two people who heard about an event from someone else can increase our perception that the probability of an unlikely story is accurate. Degrees of belief can also be seen in my experiences with rain and our two houses. I learned to discount the probability of rain at our first house and to increase my expectation of rain at our new house. If the meteorologist said there was a low chance of rain when we lived on the sheltered side of a hill, then I didn’t worry much about storm forecasts. At our new house, however, if there is a chance of precipitation and storm coming from the west, I will certainly go remove anything from the yard that I don’t want to get wet, because I believe the chance that our specific neighborhood will see rain is higher than what the meteorologist predicted.

 

Probability and how we understand it and consequentially make decisions  is complex, and Gigerenzer’s explanation of the multiple facets of probability helps us better understand the complexity. Simply tallying outcomes and predicting into the future often isn’t enough for us to truly have a good sense of the probability of a given outcome. We have to think about physical design, and we have to think about the personal experiences and subjective opinions that form the probabilities that people develop and express. Understanding probability requires that we hold a lot of information in our head at one time, something humans are not great at doing, but that we can do better when we have better strategies for understanding complexity.
Navigating Uncertainty with Nudges

Navigating Uncertainty with Nudges

In Risk Savvy Gerd Gigerenzer makes a distinction between known risks and uncertainty. In a foot note for a figure, he writes, “In everyday language, we make a distinction between certainty and risk, but the terms risk and uncertainty are used mostly as synonyms. They aren’t. In a world of known risks, everything, including the probabilities, is known for certain. Here statistical thinking and logic are sufficient to make good decisions. In an uncertain world, not everything is known, and one cannot calculate the best option. Here, good rules of thumb and intuition are also required.” Gigerenzer’s distinction between risk and uncertainty is important. He demonstrates that people can manage decision-making when making risk based decisions, but that people need to rely on intuition and good judgement when dealing with uncertainty. One solution to improved judgement and intuition is to use nudges.

 

In the book Nudge, Cass Sunstein and Richard Thaler encourage choice architects to design systems and structures that will help individuals make the best decision in a given situation as defined by the chooser. Much of their argument is supported by research presented by Daniel Kahneman in Thinking Fast and Slow, where Kahneman demonstrates how predictable biases and cognitive errors can lead people to making decisions that they likely wouldn’t make if they had more clear information, had the ability to free themselves from irrelevant biases, and could improve their statistical thinking. Gigerenzer’s quote supports Sunstein and Thaler’s nudges by building on the research from Kahneman. Distinguishing between risk and uncertainty helps us understand when to use nudges, and how aggressive our nudges may need to be.

 

Gigerenzer uses casino slot machines as an example of risk and for examples of uncertainty uses stocks, romance, earthquakes, business, and health. When we are gambling, we can know the statistical chances that our bets will pay off and calculate optimal strategies (there is a reason the casino dealer stays on 17). We won’t know what the outcome will be ahead of time, but we can precisely define the risk. The same cannot be said for picking the right stocks, the right romantic partner, or when creating business, earthquake preparedness, or health plans. We may know the five year rate of return for a company’s stocks, the divorce rate in our state, the average frequency and strength of earthquakes in our region, and how old our grandfather lived to be, but we cannot use this information alone to calculate risk. We don’t know exactly what business trends will arise in the future, we don’t know for sure whether we have a genetic disease that will strike us (or our romantic partner) down sooner than expected, and we can’t say for sure that a 7.0 earthquake is or is not possible next month.

 

But nudges can help us in these decisions. We can use statistical information for business development and international stock returns to identify general rules of thumb when investing. We can listen to parents and elders and learn from their advice and mistakes when selecting a romantic partner, intuiting the traits that make a good (or bad) spouse. We can overengineer our bridges and skyscrapers by 10% to give us a little more assurance that they can survive a major and unexpected earthquake. Nudges are helpful because they can augment our gut instincts and help bring visualizations to the rules of thumb that we might utilize.

 

Expecting everyone’s individual intuition and heuristics to be up to the task of navigating uncertainty is likely to lead to many poor choices. But, if we help pool the statistical information available, provide guides, communicate rules of thumb that have panned out for many people, and structure choices in ways that help present this information, then people can likely make marginally better decisions. My suggestion in this post, is a nudge to use more nudges in moments of uncertainty. When certainty exists, or even when calculable risks exist, nudges may not be needed. However, once we get beyond calculable risk, where we must rely on judgement and intuition, nudges are important tools to help people navigate uncertainty and improve their decision making.
Inventing Excuses - Joe Abittan

Inventing Excuses

With the start of the new year and the inauguration of a new president of the United States, many individuals and organizations are turning their eyes toward the future. Individuals are working on resolutions to make positive changes in their lives. Companies are making plans and strategy adjustments to fit with economic and regulatory predictions. Political entities are adjusting a new course in anticipation of political goals, agendas, and actions of the new administration and the new distribution of political power in the country. However, almost all of the predictions and forecasts of individuals, companies, and political parties will end up being wrong, or at least not completely correct.

 

Humans are not great forecasters. We rarely do better than just assuming that what happened today will continue to happen tomorrow. We might be able to predict a regression to the mean, but usually we are not great at predicting when a new trend will come along, when a current trend will end, or when some new event will shake everything up. But this doesn’t mean that we don’t try, and it doesn’t mean that we throw in the towel or shrug our shoulders when we get things wrong.

 

In Risk Savvy Gerd Gigerenzer writes, “an analysis of thousands of forecasts by political and economic experts revealed that they rarely did better than dilettantes or dart-throwing chimps. But what the experts were extremely talented at was inventing excuses for their errors.” It is remarkable how poor our forecasting can be, and even more remarkable how much attention we still pay to forecasts. At the start of the year we all want to know whether the economy will improve, what a political organization is going to focus on, and whether a company will finally produce a great new product. We tune in as experts give us their predictions, running through all the forces and pressures that will shape the economy, political future, and performance of companies. And even when the experts are wrong, we listen to them as they explain why their initial forecast made sense, and why they should still be listened to in the future.

 

A human who threw darts, flipped a coin, or picked options out of a hat before making a big decision is likely to be just as wright or just as wrong as the experts who suggest a certain decision over another. However, the coin flipper will have no excuse when they make a poor decision. The expert on the other hand, will have no problem inventing excuses to explain away their culpability in poor decision-making. The smarter we are the better we are at rationalizing our choices and inventing excuses, even those that don’t go over so well.
The Human Need for Certainty - Joe Abittan

The Human Need for Certainty

Throughout the book Risk Savvy, Gerd Gigerenzer discusses the challenges that people face with thinking statistically, assessing different probable outcomes, and understanding risk. Gigerenzer also discusses how important it is that people become risk literate, and how the future of humanity will require that people better understand risk and uncertainty. What this future requires, he explains, is fighting against aspects of human psychology that are common to all of us and form part of our core nature. One aspect in particular that Gigerenzer highlights as a problem for humans moving forward, is our need for certainty.

 

“Humans appear to have a need for certainty, a motivation to hold onto something rather than to question it,” he writes. Whether it is our religion, our plans for retirement, or the brand of shoes we prefer, we have a need for certainty. We don’t want to question whether our religious, political, or social beliefs are correct. It is more comforting for us to adopt beliefs and be certain that we are correct. We don’t want to continuously re-evaluate our savings plans and open ourselves to the possibility that we are not doing enough to save for retirement. And we like to believe that we purchased the best running shoes, that we bough the most sustainable shoes for the planet, and that our shoe choices are the most popular. In all of these areas, ambiguity makes our decisions harder whereas a feeling of certainty gives us confidence and allows us to move through the world. In many ways, our need for certainty is simply a practicality. There are unlimited possibilities and decisions for us to make every day. Adopting certainty eliminates many possibilities and choices, simplifying our life and allowing us to move through the world without having to question every action of every second of every day.

 

But in the modern world, humans have to be more comfortable living with ambiguity and have to be able to give up certainty in some areas. “For the mature adult,” Gigerenzer writes, “a high need for certainty can be a dangerous thing.”  We live with risk and need to be able to adjust as we face new risks and uncertainties in our lives. We like to hold onto our beliefs and we are not comfortable questioning our decisions, but it can be necessary for us to do so in order to move forward and live in harmony in a changing world with new technologies, different demographics, and new uncertainties. A need for certainty can lead people to become dogmatic, to embrace apologetics when discounting science that demonstrates errors in thinking, and to ignore the realities of a changing world. One way or another, we have to find ways to be flexible and adjust our choices and plans according to risk, otherwise we are likely to make poor choices and be crushed when the world does not align itself with our beliefs and wishes.
Dread Risks - Joe Abittan

Dread Risks

Over the course of 2020 we watched COVID-19 shift from a dread risk to a less alarming risk. To some extent, COVID-19 became a mundane risk that we adjusted to and learned to live with. Our initial reactions to COVID-19, and our later discontent but general acceptance reveal interesting ways in which the mind works. Sudden and unexplained deaths and risks are terrifying, while continual risk is to some extent ignored, even if we face greater risk from dangers we ignore.

 

In Risk Savvy Gerd Gigerenzer describes dread risks and our psychological reactions by writing, “low-probability events in which many people are suddenly killed trigger an unconscious psychological principle: If many people die at one point in time, react with fear and avoid that situation.” Dread risks are instances like terrorist attacks, sudden bridge collapses, and commercial food contamination events. A risk that we did not consider is thrust into our minds, and we react strongly by avoiding something we previously thought to be safe.

 

An unfortunate reality of dread risks is that they distract us and pull our energy and attention away from ongoing and more mundane risks. This has been a challenge as we try to keep people focused on limiting COVID-19 and not simply accepting deaths from the disease the way we accept deaths from car crashes, gun violence, and second hand smoke exposure. Gigerenzer continues, “But when as many or more die distributed over time, such as in car and motorbike accidents, we are less likely to be afraid.” Dread risks trigger fears and responses that distributed risks don’t.

 

This psychological bias drove the United States into wars in Iraq and Afghanistan in the early 2000s and we are still paying the prices for those wars. The shift of COVID-19 in our collective consciousnesses from a dread risk to a distributed risk lead to mass political rallies, unwise indoor gatherings, and other social and economic events where people contracted the disease and died even though they should have known to be more cautious. Reacting appropriately to a dread risk is difficult, and giving distributed risks the attention and resources they deserve is also difficult. The end result is poor public policy, poor individual decision-making, and potentially the loss of life as we fail to use resources in a way that saves the most lives.
Stats and Messaging

Stats and Messaging

In the past, I have encouraged attaching probabilities and statistical chances to the things we believe or to events we think may (or may not) occur. For example, say Steph Curry’s three point shooting percentage is about 43%, and I am two Steph Currys confident that my running regiment will help me qualify for the Boston Marathon. One might also be two Steph Currys confident that leaving now will guarantee they are at the theater in time for the movie, or that most COVID-19 restrictions will be rescinded by August 2021 allowing people to go to movies again. However, the specific percentages that I am attaching in these examples may be meaningless, and may not really convey an important message for most people (Myself included!). It turns out, that modern day statistics and the messaging attached to it is not well understood.

 

In his book Risk Savvy, Gerd Gigerenzer discusses the disconnect between stats and messaging, and the mistake most people make. The main problem with using statistics is that people don’t really know what the statistics mean in terms of actual outcomes. This was seen in the 2016 US presidential election when sources like FiveThirtyEight gave trump a 28.6% chance of winning and again in 2020 when the election was closer than many predicted, but was still well within the forecasted range.  In both instances, a Trump win was considered such a low probability event that people dismissed it as a real possibility, only to be shocked when Trump did win in 2016 and performed better than many expected in 2020. People failed to fully appreciate that FiveThirtyEight’s prediction meant that in 28.6% of election simulations, Trump was predicted to win in 2016, and in 2020 many of their models predicted races both closer than and wider than the result we actually observed.

 

Regarding weather forecasting and statistical confusion, Gigerenzer writes, “New forecasting technology has enabled meteorologists to replace mere verbal statements of certainty (it will rain tomorrow) or chance (it is likely) with numerical precision. But greater precision has not led to greater understanding of what the message really is.” Gigerenzer explains that in the context of weather forecasts, people often misunderstand that a 30% chance of rain means that on 30% of days when when the observed weather factors (temperature, humidity, wind speeds, etc…) match the predicted weather for that day, rain occurs. Or that models taking weather factors into account simulated 100 days of weather with those conditions and included rain for 30 of those days.  What is missing, Gigerenzer explains, is the reference class. Telling people there is a 30% chance of rain could lead them to think that it will rain for 30% of the day, that 30% of the city they live in will be rained on, or perhaps they will misunderstand the forecast in a completely unpredictable way.

 

Probabilities are hard for people to understand, especially when they are busy, have other things on their mind, and don’t know the reference class. Providing probabilities that don’t actually connect to a real reference class can be misleading and unhelpful. This is why my suggestion of tying beliefs and possible outcomes to a statistic might not actually be meaningful. If we don’t have a reasonable reference class and a way to understand it, then it doesn’t matter how many Steph Currys likely I think something is. I think we should take statistics into consideration with important decision-making, and I think Gigerenzer would agree, but if we are going to communicate our decisions in terms of statistics, we need to ensure we do so while clearly stating and explaining the reference classes and with the appropriate tools to help people understand the stats and messaging.
Risk Literacy - Joe Abittan

Risk Literacy

In February of 2020 I finished a book called Risk Savvy by Gerd Gigerenzer. At the time I read the book, I could not predict that thinking about risk would come to dominate the remainder of the year. Throughout 2020 and into the start of 2021, humanity across the globe has demonstrated how poorly we think about and handle risk. The United States has clearly been worse than most countries, as we have failed to understand the risk of COVID-19, failed to grapple with the risk of crowds and appropriate uses of force, and failed to adequately assess the risk of a President living in a state of denial and delusion.  As Gigerenzer writes on page 6 of his book, Risk Literacy is the basic knowledge required to deal with a modern technological society,” and in many ways, the United States and the rest of humanity have shown that risk literacy is deeply lacking.

 

Gigerenzer believes that we are smart, that we are resourceful, and that with proper aids and education, we can become risk literate. Whether we recognize it or not, we already calculate risk and make decisions based on risk. Understanding risk can lead to us packing an umbrella and wearing a waterproof windbreaker when the weather station forecasts rain. We can make sound investments without understanding every aspect of an investment thanks to savings vehicles that help us better understand and calibrate risk. And we can decide to go to a movie or skip it based on aggregated reviews and ratings scores on Rotten Tomatoes.

 

At the same time, we have had trouble understanding our individual risks related to COVID-19, we have had trouble understanding the risks and benefits of wearing masks, and we have dismissed what seem like impossible possibilities until they happen to us personally, or happen in a dramatic way on tv. We are capable of making good decisions based on perceptions and understandings of risk, but at the same time, we have still shown ourselves to be risk-illiterate.

 

It is clear that moving forward societies will have to do better to become risk literate. We will have to improve our ability to communicate risk, estimate risk, and take appropriate precautions or actions. We cannot live in a world free from risk, and new technologies, ecological pressures, and sociopolitical realities will change the risk calculations that everyone will have to make. Improving our risk literacy might mean that we don’t have over 400,000 people die during future respiratory pandemics. It might mean we have robust economic systems that don’t damage the planet. And it might mean we are able to live together peacefully with global superpowers competing economically. Failure to address risk and failure to improve risk literacy could lead to disaster in any one of those areas.
Asymmetric Paternalism

Asymmetric Paternalism

While writing about the book Nudge by Richard Thaler and Cass Sunstein, I have primarily focused on an idea that the authors call Libertarian Paternalism. The idea is to structure choices and use nudges (slight incentives and structural approaches) to guide people toward making the best possible decision as judged by themselves. Maintaining free choice and the option to investigate or chose alternatives is an important piece of the concept, as is the belief that we will influence people’s decisions no matter what, so we should use that influence in a responsible way to help foster good decision-making.

 

But the authors also ask if it is reasonable to go a step beyond Libertarian Paternalism. Is it reasonable for choice architects, governments, and employers to go further than gentle nudges in decision situations? Are there situations where decision-making is too important to be left to the people, where paternalistic decision-making is actually best? Sunstein and Thaler present an introduction to Asymmetric Paternalism as one possible step beyond Libertarian Paternalism.

 

“A good approach to thinking about these problems has been proposed by a collection of behavioral economists and lawyers under the rubric of Asymmetric Paternalism. Their guiding principle is that we should design policies that help the least sophisticated people in society while imposing the smallest possible costs on the most sophisticated.”

 

This approach is appealing in many ways, but also walks the line between elitism, the marginalization of entire segments of society, and maximizing good decision-making. I hate having to make lots of decisions regarding appropriate tax filings, I don’t want to have to make decisions on lots of household appliances, and I don’t really want to have to spend too much time figuring out exactly what maintenance schedule is the best for all of my cars. However, I do want to get into the weeds of my healthcare plan, I want to micromanage my exercise routine, and I want to select all the raw ingredients that go into the dinners and lunches that I cook. On some decisions that I make, I want to outsource my decision-making and I would often be happy with having someone else make a decision so that I don’t have to. But in other areas, I feel very sophisticated in my decision-making approach, and I want to have maximum choice and freedom. Asymmetric Paternalism seems like a good system for those of us who care deeply about some issues, are experts in some areas, and want to maintain full decision-making in the areas we care about, while exporting decision-making in other areas to other people.

 

Of course, prejudices, biases, and people’s self-interest can ruin this approach. What would happen if we allowed ourselves to deem entire groups of people as unworthy of making decisions for themselves by default? Could they ever recover and be able to exercise their freedom to chose in important areas like housing, retirement, and investment spaces? Would we be able to operate for long periods of time under a system of Asymmetric Paternalism without the system devolving due to our biases and prejudices? These are real fears, and while we might like to selectively trade off decision-making when it is convenient for us, we also have to fear that someone else will be making decisions for us that are self-serving for someone other than ourselves.

 

The point, according to Sunstein and Thaler, would be to maintain the freedom of decision-making for everyone, but to structure choices in a way where those with less interest and less ability to make the best decisions are guided more strongly toward what is likely the best option for them. However, we can see how this system of asymmetric paternalism would get out of control. How do we decide where the appropriate level is to draw the line between strong guidance and outright choosing for people? Would people voluntarily give up their ability to chose and overtime hand over too many decisions without an ability to get their decision authority back? Transparency in the process may help, but it might not be enough to make sure the system works.
Influence is Unavoidable - Joe Abittan

Influence is Unavoidable

The last several years has seen the rise of social media influencers across the globe. Regular people have been able to amass large numbers of followers on social media, gaining status, earning contracts with brands and advertisers, and of course, influencing people. Their influence is often times out in the open, like when they endorse a brand or product, but is often hidden, in messages about how life should be, about how one should look, and about the things that people should focus on in life. These hidden and secondary influences may also be intentional, but may also be subconscious, and ultimately, they are unavoidable. To go even deeper, these types of influences are not just limited to bona fide social media influencers, they are unavoidable aspects of all our lives and actions.

 

Whether we want to or not, we influence people’s choices. Some of us are in spaces where we are explicitly asked and expected to influence people’s choices, but many of us are not. Nevertheless, our conversations, decisions, and actions can still influence how people make choices. If we deliberately shape choice environments via a position of authority, we are a choice architect. Human resource managers, club leaders, and parents are all choice architects, with the direct mandate to shape the choices of employees, club members, and children respectively. Nevertheless, fellow employees, adjacent club members, and other kids also influence the decision spaces that choice architects create, regardless as to whether they ever intended to.

 

Cass Sunstein and Richard Thaler consider the implications of unavoidable influence in their book Nudge. The authors believe that choice architects have the responsibility to nudge individuals toward specific decisions and choices that the individuals themselves would deem to be the best outcome. That responsibility, according to Sunstein and Thaler cannot be shirked, because influence cannot be avoided. “It is not possible to avoid choice architecture, and in that sense it is not possible to avoid influencing people.”

Sunstein and Thaler are specifically writing about the impossibility of choice architects to try not to influence people. Most of this post has been about inadvertent influence from people who are not trying to influence others. The two ideas are separate, but important to directly connect and discuss. Choice architects still influence people and their decisions even if they try to pull back completely. Sunstein and Thaler would argue that in doing so they often end up creating decision situations where people make worse choices than if the choice architect had deliberately intervened in an attempt to help people make a better choice. One reason for this is that unintentional influence will play a greater role in the absence of choice architects actively trying to improve people’s decision-making. Imagine that your HR representative had gone through the effort of trying to map out possible health benefits scenarios for employees based on age, family dynamics, and potential health concerns. The way that employees select health insurance plans could be shaped by these mapping scenarios, deliberately trying to help people make better choices. Or, the rep could decide not to try to influence decisions and step back. By doing so, employees might hear from colleagues about a great option that worked for someone else, and select that option even if it doesn’t fit their family and health needs. Other employees may not have been encouraging everyone to select an option, but they still have the ability to indirectly and unintentionally influence the choices of others, potentially for the worse. The choice architect who tried not to influence the decision created a situation where outside uneducated voices were louder – their actions still influenced the decision space.

Choice architects don’t exist in a vacuum. They are part of a larger ecosystem and the context for each choice matters. Choosing not to influence someone doesn’t mean that you don’t still influence them or their choices. Neither does being unaware that you influence others. Influence is unavoidable, and Sunstein and Thaler would argue that it is important that we therefore think about how we intentionally and unintentionally influence others, and for choice architects to be active in helping people understand the best option for themselves in a given choice.