Monday, June 13, 2011

Can Big (Companies) Be Beautiful?

As companies grow, they slow down. Once they get past a certain size, they usually experience what’s called organizational sclerosis--a culture change that makes effective adaptation almost impossible. They cling to old business models, become less responsive to their customers, and make big costly mistakes. Their shares drop.

People have various theories as to why this happens. Most of them relate to organizational complexity or conflict from within. The strategies for fixing it usually involve hiring expensive consultants or laying off hundreds of employees. These strategies are painful and often simply don’t fix the core problem. But what if there was a way to make large companies as dynamic as small ones? I think I might have found one.

Part of the Tokenomics research that I’ve been doing of late has involved building computer simulations of organizational cultures. (See previous post.) The simulations are still very simple and there’s lots of work still to be done before any of the results find their way into a science paper. However, some of the results I've discovered are pretty compelling. This is one of them. 

I’ve been exploring the idea that people look for predictable, repeatable self-validation in their interactions with others, or ‘token collection’--an idea that you can find out more about on the SF Behavior Lab website, or in the in-depth documents stored at Almost all the token collection that we do is driven by non-conscious pattern-seeking. This is very useful because it means that we don’t have to model complete human minds in order to get a rough picture of how cultures work. We can simply create populations of agents that interact by seeking out repeatable patterns of dialog, and modifying their behavior if their conversations don’t run as expected.

In the examples I described in my earlier post, the agents building cultures had a pretty limited set of options available to them and so the communities they formed didn’t show much variation. However, for the experiments I want to share with you this time, the interactions are rather different. Agents speak in turns, rather than at the same time, and have a choice of 128 different words they can say. (128 openings x 128 responses). This means that the agents have a much harder time of it finding familiar subjects to talk about with their peers, and so have to be a little creative (random) in their choices.

Most of the interactions that the agents try out don’t meet with much success. They produce fairly short-lived patterns shared only by a few robotic acquaintances. However, some interaction patterns manage to spread and dominate, just like fads in real life. Once these patterns get established, they become incredibly stable. They become ‘cultural interactions’ and end up being common to all the agents in the simulation. Removing or adding agents incrementally doesn’t change the culture one bit.

Also, interestingly, changing the size of the population of agents has a serious effect on how often new patterns get incorporated into the culture. Bigger populations have a lot more trouble accepting new ideas than small ones. At some level, though, this shouldn’t surprise us. In order for an idea to develop critical mass and spread rapidly through the group, a certain percentage of the agents have to have already heard it. That’s going to get steadily less likely as the system scales up.

For these simple agents, a critical drop-off happens before you get to a population size of forty.  So far as I know, there’s no equivalent change in performance for human groups in this size range. The literature suggests that real organizations start having problems at around a hundred and fifty people. However, in the context of agent simulations, this makes sense. The agents have much smaller memories than people and thus less intrinsic flexibility. So, while the drop off in effectiveness may not be exactly the same phenomenon as organizational sclerosis, there seems to be a close enough fit here to make exploring the effect worthwhile.

The question I asked was ‘what do I have to do to a set of agents in order to keep their rate of cultural adaptation high while the population increases in size’. Specifically, what did I need to do to make a population of forty agents as imaginative as one of size twenty.

The answer, for agents, at least, is simple. You divide them into two groups of size twenty and let them run independently. Then after a while, you reshuffle them into two new groups, and leave them alone again. Every time the two groups are shuffled, learning from each community gets transferred into the total population. However, because the groups stay small, the rate of creativity stays at what you’d expect for a much smaller population. 

This process is rather like the genetic recombination that happens in sexual selection. We keep the rate of change high by randomly mixing our ingredients and then making sure that any useful results get shared back into the population.

Also, it turns out that this isn’t the only method that will work. Any solution that involves keeping group sizes small most of the time, combined with random mixing, will do the job. And there appears to be an optimum group size for a given setting of the simulation parameters at which creativity can spread fastest.

The lesson seems clear: turning big organizations into a cross-polinating swarms of smaller ones might make them a lot more agile. While this approach probably isn’t a fit for all kinds of business, I can think of plenty where it might work: finance and software being the first two to spring to mind.

Once I’d managed to keep a big culture as adaptable as a small one, I wanted to know if I could do better. After all, large organizations have more people and therefore more ideas per day than small ones. In an ideal world, shouldn’t they be able to capitalize on that?

The answer is yes if you bend the rules a little. I couldn’t find a way of reorganizing the teams that led to further improvements in creativity beyond what I’d already tried. However, when I started messing with the parameters of the simulation, I noticed something interesting: increasing the number of possible witnesses to each interaction by one caused the rate of creativity to jump up rapidly. One way to say this is that by raising the number of people in a team who get to listen in on important interactions increases the rate at which new ideas can spread 

This provides some simulation evidence that the people in the Agile Software Development movement are on the right track. They’ve been saying for years that creating small, tightly-knit teams where information is regularly shared creates the right environment for creative work. Furthermore, the agents in this simulation are so simple and general that we can expect the same logic to hold for a large number of possible situations.

The important next step will be to see whether the same tricks work as I add in more of the complexity of real organizations. For instance, trust relationships, hierarchies, and differences in effective value between the habits being propagated all have yet to be added. If the splitting/merging effect still works when all these features are added in, it might be time to start rethinking how we grow companies. Organized packs of collaborative mammals could turn out to be a lot more adaptable than a few big dinosaurs.

Wednesday, June 8, 2011

Adventures in Game Theory, Part Four

For those of you freshly joining this adventure, the last three posts have led us on a strange, thrilling journey that has passed through the valleys of introductory game theory, the jungles of applied improv, and the mountains of software simulation. Now, at last we arrive at our thunderous finale on the shores of Lake Awesome. I highly recommend reading from the start of the sequence, otherwise what I have to say may be too extraordinary and wonderful for your mind to fully hold!

At the end of the last installment caught me teetering on the brink of a realization--that by adding just a little more functionality to my simulation, I could start exploring some more socially useful truths about how people behave. My insight was to add status.

What this meant in practice was splitting the population of agents in my model into two groups: bosses and workers, or in training community parlance: leaders and team-members. Then, in order to make the interactions between bosses and workers a little less benign, I added two extra constraints.

One: If bosses were aggressive (nose-thumbing) to workers, workers were not empowered to reciprocate and be aggressive back in their next encounter.

Two: Bosses were unable to remember the specifics of positive interactions they had with workers. So for instance, if a boss and a worker both chose paper in one round, the worker would remember the fact, but the boss would not.

Implementing these changes was easy, as it simply required that the two memory rules I’d already added to make the first simulation work were now dependent on status. (I also added a little extra logic around the movement of the agents to ensure that workers had to interact with bosses, and to make the movements of bosses dependent on other bosses but not workers. However, while necessary, that code is somewhat beside the point.)

What happened next was wonderfully clear. Within a few seconds, all the bosses were behaving aggressively while the workers normed on a set of social standards of their own. My simulation suddenly looked a lot like some of the more awful companies I’d worked for. Without having to say anything about the kinds of people who become leaders, or about the specifics of organizational culture, I’d captured a simple truth about leadership: that without the incentives to behave otherwise and the right skills to succeed, people with power slide towards bad behavior, even if they start off thinking like saints.

What was even more interesting was that as the simulation progressed, the bosses started to bump up against the corners of the virtual environment as if desperate to leave. Because aggressive behavior was so successful for bosses in their interactions with workers, they were applying the same behavior to each other, resulting in a rapid erosion of their ability to collaborate. The lesson: by letting leaders behave badly, we ensure that leaders have less pleasant interactions with each other, as well as with us.

My goal, though, was not to engage in rhetoric about leaders, but instead to see whether models like the one I was looking at could tell us something about how to help organizations do better. To do this, I looked at what happened when I turned each of the status dependencies off in isolation.

Turning off the status dependency for remembering positive interactions is rather like sending your managers on an employee recognition course. They learn to value the specific information they get from each person they work with, and to let their team members know that they’re seen and valued.

The result in the simulation is that the culture improves significantly. The workers integrate more tightly and the bosses take on the same cultural colors as the workers they lead. Interestingly, the bosses don’t all start cooperating at once. Many of them initially retain their aggressive behavior. Then, one by one, they figure out that collaboration is more effective.

The lesson here: that training leaders to listen can make a huge difference in their effectiveness, but that the change they take on depends on their willingness to implement what they learn.

If instead, we turn off the status dependency for worker retaliation to boss aggression, the effects are even more interesting. Making this change is rather like implementing a shared accountability system like the one that revolutionized the airline industry and transformed the safety standards in air travel. Under this system, the pilots of planes are no longer the unquestionable captains of the air that they once were. If copilots think that they’re witnessing a mistake, they’re duty-bound to call the pilot on it and to report it to air traffic control if necessary. In our simulated business, we can imagine that we’re instructing the worker agents to hold their bosses accountable if they don’t uphold the collaborative social standards of their organization.

What happens when we make this change is that the behaviors of the bosses have trouble settling onto any specific color. When we watch the ‘mood’ of the agents to see how many positive or negative interactions they’re having, we see that the tables have been turned. The workers are now having a pretty great time all round and the bosses are mostly miserable--the opposite of what we see if status dependence for retaliation is left on. This is because the workers now have an advantage that the bosses don’t--they can remember and repeat positive interactions whereas bosses cannot. Because aggression no longer secures automatic results, bosses don’t have an easy way of stabilizing on a successful behavior.

The lesson here is that enabling everyone in an organization to hold leaders accountable for their behavior is what creates the incentive for leaders to improve, but that without the right training and direction, the main result is leader unhappiness.

As you might expect, turning off both status-dependent features creates a benign, functional organization that settles rapidly onto a cooperative culture. If you want to play around yourself, and have Java installed, the simulation is the second applet on this page. (It has four buttons.)

As before, red, blue and green denote different positive interactions. Gray denotes aggressive behavior. Swapping to ‘mood view’ shows the success of the agents interactions, ranging from blue (unhappy agents) to yellow (cheerful ones).

Clearly there’s a lot more to do here. For a start, in order to turn this into a science result, the simulations will need to be a lot more rigorous, which will probably mean sacrificing the visual playfulness.  Furthermore, we’ve only looked at one memory model for agents and solid research would need to try out others. However, the results seem pretty clear. We’ve gone from a simple game played in a room full of people to a model that turns business intuition into something rather like unavoidable, mathematical fact.

Thus, in the wake of our adventure, we can say with real confidence that any society or organization that doesn’t empower its people hold its leaders accountable, and which doesn’t teach those leaders how to listen, can expect its leaders to turn out bad, regardless of how ‘good’ we believe them to be as people.

This is something most of us already believe but which we often fail to implement. For instance, we're all used to the idea of holding elected officials accountable, but explicit training in 'voter recognition'? We leave that to chance. Similarly, we're used to the idea that good managers are the ones who pay attention, but company-wide accountability systems? Those are pretty rare. I believe that simulations like this can make these points unavoidable, and also perhaps show us how to build measures that make our adherence to such standards quantifiable.

For any skeptics out there, my huge thanks for reading this far, and here’s a final thought to consider. Agent-based simulations of this sort have been used by biologists for years on the following basis: we can’t capture all the details of natural systems like cultures or the lives of organisms, so instead we capture only what we know is true. From that, we look to see what else must be true as a consequence. Thus we attempt to make the simplicity of the model a strength, not a weakness. In this instance, the agents are so simple that we can expect the same effects to arise regardless of the memory model we employ for our agents, so long as that memory model permits learning. Further work in this area will hopefully make that point even clearer.

That’s it. The adventure is finished. And while the ending perhaps isn’t unexpected, it feels like a step forwards to me. After all, if we can do this starting with Rock Paper Scissors, think what we can do with the game of Twister.

Tuesday, June 7, 2011

Adventures in Game Theory, Part Three

To those fresh to this sequence of postings, let me give you a little context. Two posts ago, I implied that some kind of wildly significant insight about how organizations and societies worked could be derived from looking at simple playground games like Rock Paper Scissors. Over the course of the last two posts, I’ve been building up the case for that statement. Now comes the next thrilling, life-changing installment--this time with some simulation results!

Before I can fully explain, though, first I have to give you a little more background.  Last week I had the good fortune to speak at the ASTD conference in Orlando, Florida, the world’s largest training and development business event. The topic of the session was the use of Tokenomics as a tool for organizational culture change. I delivered the talk with my good friend Cindy Ventrice, from, and to support the session we captured a large amount of material on the subject, which those interested can find on our collaboration website, The session went wonderfully and generated plenty of interest. However, what I’m most keen to talk about here doesn’t relate to that talk, exactly, but to the unexpected consequences of it.

In order demonstrate to the audience what the Tokenomics approach was capable of, I put together a short computer simulation based on Scissors Dilemma Party, a game which the readers of the last two posts will have already heard of. The simulation was designed to show how autonomous software agents, given nothing but a simple memory model and some behavioral rules based on token acquisition, would automatically aggregate into social groups defined by shared values.

To make the model more intuitively approachable for a conference audience, I chose to have the agents move around in a virtual environment rather like people in a workplace, interacting when they met. As well as making the simulation more visually appealing, it demonstrated how the agents’ behavior evolved over time as they learned more about their environment, much as players of the game do when they experience it at Behavior Lab.

Each agent had eight memory slots initially filled with random behaviors. With each interaction, an agent would pick a behavior from its memory and apply it. If the interaction resulted in a positive outcome for the agent (unreciprocated nose-thumbing, or a successful rock-paper-scissors match), that behavior was copied to another slot in memory. If the behavior resulted in any other outcome, that memory slot was overwritten with a new random behavior. Agents were designed to move towards other agents with whom they’d interacted positively, and away from those with whom interaction had failed.

At first, the simulation didn’t work very well. Aggressive behavior (nose-thumbing), was too seductive for the dim-witted agents and stable social groups never formed. In order to get the agents to behave a little more like people, I had to add a little extra subtlety. This came in the form of two new rules.

The first rule was that if an Agent A was aggressive to agent B, B would remember that fact and be aggressive back at the next opportunity. This captures the idea of ‘Tit for Tat’--a strategy that has proved very successful in Prisoner’s Dilemma tournaments.

The second rule was that if A and B had a successful match of rock, paper, or scissors, they’d both remember it and try for the same topic of conversation next time. This gave the agents a chance to reinforce positive relationships.

These two rules together did the trick and produced a somewhat mesmeric simulation. You can see it here, by just clicking on the first simulation button that appears. (Sadly, Blogger appears to become a trifle unstable when supporting applets, otherwise I would have included it in this blog. Also, note that you’ll need Java installed for this to work. If you don’t have Java, let me know. I’m thinking of writing an HTML5 version and am keen to know whether that would make life easier for people.) In this simulation, the colors red, green, and blue take the place of rock, paper and scissors. The color gray takes the place of nose-thumbing.

However, once I’d finished the simulation, it occurred to me that I’d only scratched the surface of what could be demonstrated with this approach. I could go further, do more, and start saying something really meaningful. Better still, the tools to achieve it were already in my hands! However, I’ve promised myself that each one of these postings will be short and readable by people with day jobs, so in order to discover what I did next, you’ll have to join me for Episode Four.

[Side note: my friend Cindy is awesome and so is her book. I can't recommend it highly enough.]

Sunday, June 5, 2011

Adventures in Game Theory, Part Two

In the previous installment of this adventure, I promised to reveal how the secrets to business effectiveness and social harmony could be achieved by playing games like Rock Paper Scissors. Will I be able to deliver on that outrageous promise? Only by reading on will you get to find out.

For the next part of our journey, let’s consider a new game which we’ll call Scissors Party. The rules are simple and very much like those of Rock Paper Scissors. Players bounce their fists as usual and then pick any one of the three gestures normally used in the game. However, the scoring system in this version is different. In Scissors Party, players get two points each if they successfully match their opponent’s choice and no points if they don’t match. So if two players both choose paper, they get two points each. If one player chooses scissors and the other chooses paper, nobody gets any points. As in Dilemma Party, players are free to stay with the same partner or mingle in the group as they like. Any guesses as to what happens?

You may have already guessed that players tend to form pairs and small clusters that make the same choice every time, eg: always rock or always paper. Even though lots of people will still mingle, they figure out fairly quickly that they’re not making as many points as the people who stay put. Just as in Dilemma Party, interpersonal dynamics add complexity to the game. Some people want to move around and take risks, while others just want to ace the game, so the results are never as perfectly consistent as we might imagine. However, the patterns are still pretty clear.

So far so good. But where it gets really interesting is when you put Dilemma Party and Scissors Party together. This gives you Scissors Dilemma Party: a game that gives players four options: rock, paper, scissors and nose-thumbing.  The scoring works as you’d expect:
  • Thumbing gets you three points against rock, paper, or scissors but only one point against another thumb. 
  • Successfully matching rock, paper, or scissors with your partner gets you two points.
  • Failing to match with rock, paper or scissors, or coming up against a thumb, gets you zero points.
Everyone confused yet?

What’s bizarre is what happens when you play this game with a room full of people who have just played Scissors Party moments before. Even though they know full well that they can form cliques and collaborate to get two points each turn, people will form little clusters that repeatedly thumb noses instead, getting one point each instead. This means that they’re being half as effective at playing as they were thirty seconds ago, simply because they’ve been given the option to play it safe at the cost of other players. This, to me, is a fascinating example of how being given the option to tune out and avoid cooperation produces instant defensiveness and a change in social cohesion.

Perhaps some of you will by now have figured out where I’m going with these games by now. Choosing different gestures in the game is very much like choosing tokens to collect in life. Pairwise interactions are rather like small versions of the conversations we have every day. Rock, paper and scissors equate to different forms of social value, such as sexiness, intelligence, or likability. Nose thumbing equates to extracting involuntary tokens from others for personal validation gain. Whereas our choice of gestures in the game is conscious and our choice of tokens in life is non-conscious, the same patterns of defensive behavior can be seen. In fact, in non-conscious group behavior, we tend toward more predictable responses. Thus, playing Scissors Dilemma Party gives us an interesting, lightweight model for looking at how social groups form and interact.

Intriguing, I hear you say, but still not yet a conclusive solution to the world’s ills. True. To see the awesome social significance of Scissors Dilemma Party in all its glory, you’ll have to read Adventures in Game Theory Part Three.

Saturday, June 4, 2011

Adventures in Game Theory, Part One

Question: Can playing simple games like Rock Paper Scissors teach us how to be better leaders, help us build effective, equitable organizations, and pave the way to a more harmonious world?
Answer: Yes! Undoubtedly!

If you want to know how, and why I would make such a ridiculous-sounding assertion, then I invite you to come with me on a journey into a dark and mysterious world of theoretical applied improv. The journey will be long and arduous (four blog posts), but for those who stick with me, there is treasure in store.

The starting point in this adventure is the Prisoner’s Dilemma--perhaps the best-known finding from Game Theory: a branch of math that studies how people or animals compete. Simply put, the Prisoner’s Dilemma is a formal description of a kind of situation we often face in life, in which cooperation between two parties comes with both risks and benefits, but where failing to cooperate is both safe and predictable.

People have studies Prisoner’s Dilemma very extensively. There have been research papers about it, world-spanning experiments, online tournaments between competing software programs, and dozens of books on the subject. Not satisfied by all this, I wanted to see what happened when I turned Prisoner’s Dilemma into an improv game and took it to Behavior Lab.

To this end, I created a game called Dilemma Party--a little like Rock Paper Scissors but with two  options per player instead of the traditional three. Here’s a slide I used at the ASTD conference in Orlando recently (more on that in later posts), that shows how to play, and how the scoring works.

As you can see, players have the option of thumbing their nose at their opponent or offering them an invisible gift. Offering a gift presents the best opportunity for mutual gain but comes with a risk. If the other player thumbs their nose at you, you get nothing and your opponent walks away with a nice stack of points. Thumbing your nose means that you always win something, regardless of what the other player does--it’s a safer bet but not a particularly friendly one.

Players of the game interact for an unspecified period of time, trying to rack up as many points as they can. They’re milling in a large group and can swap partners any time they like, or stay with their current partner if they prefer. What do you suppose happens if you put fifty random people in a room together and get them to play? Any guesses on what strategies they pick?

The answer is that it depends on the group. Put members of the general public together and the group norms to almost universally thumbing noses after a short time, with a few individuals doggedly giving gifts regardless of the losses they incur. However, put a room full of professional trainers together and the group norms to universal gift giving almost as fast. Perhaps unsurprisingly, pairs of players who settle on gift-giving tend to stay together. Pairs where one or more players thumb noses don’t stay together very long.

For the most part, people who aren’t already familiar with the Prisoner’s Dilemma do a very natural thing when reasoning about scores. They realize that by nose-thumbing, they can’t lose, so they keep doing it, even though they miss out on the chance to make more points by building stable relationships. No big surprises there.

Where the game gets interesting is when you look at how the rich, multi-layered nature of human interaction interferes with our stable assumptions about how the game should work. For instance, in one group, players repeatedly thumbed their opponents but then shared high-fives after each interaction. What this suggests is that the players knew they were making cautious, uncooperative choices, but still wanted to check in with each other to show that they were really friendly people at heart. Thumbing their noses felt awkward and antisocial but they didn’t want to change tactics and consequently lose! Giving high-fives was a way of subverting the game, and showing their opponents that they weren’t really in competition.

Also, those people who’ve spent a lot of time in a training, group therapy, or social workshop setting tend to repeatedly offer gifts, regardless of the consequences. I suspect that this has more to do with how those people are mentally parsing the game, rather than suggesting that they have fundamentally different personalities. These are people who’ve played similar games before and aware of the implications of cooperation. That makes them behave differently because perceiving themselves as cooperative affords them more validation than the points offered by the game. They’d rather feel positive and socially useful than win, even if that feeling comes with a very light dose of martyrdom.

Underpinning both of these reactions is the fascinating interplay between the choices made consciously in the game, and the very similar game of token exchange that the players are playing underneath. Because we load the game into the conscious awareness of the players, the acquisition of points can’t help but be held as an extrinsic goal. And because there aren’t cash prizes on offer, that goal comes with low priority. This means that the intrinsic motivations of the players guide their strategies. Thus, while we’re unlikely to get unbiased information about Prisoner’s Dilemma itself from the game, it shines a fascinating light on our motivations.

Interesting, I think, but not a recipe for social harmony just yet. There’s more we can do with these games. Much more. And for that, you’ll have to read my Adventures in Game Theory Part Two.

Monday, February 7, 2011

The Science Incubator Game

Science is fracturing. People from different fields don't really understand each other's work all that well. Specialty areas keep getting smaller and more focused. Furthermore, many scientists have to operate in a culture that discourages people from opening their mouths if they don't understand what's being said. This is because how 'brilliant' others imagine you to be often has immediate repercussions for the job you get next. This unwillingness to speak up only makes the fracturing happen faster.

I can't see this culture of caution ending any time soon without outside help because it's driven by two things:
  1. It's just harder to understand what people in other fields are doing these days because the amount of understanding that you have to invest to reach the coal-face of science is hugely more than it used to be. Consequently, people try less. 
  2. The pressure in the scientific job market is incredible and it's getting worse. Gone are the days when people walked straight from their PhDs to faculty jobs. The incentives for people to open their mouths and risk looking foolish have never been lower. 
This whole trend is unfortunate, because the research shows that interdisciplinary dialog accelerates progress. Groups with mixed skill sets consistently find solutions faster than teams of people who specialize in the same subfield. The act of having to articulate your ideas to those who may not understand is not only going to force you to bring order to your own ideas, but is also likely to lead to the creation of new ones. Creativity, as it turns out, is not driven by sudden sparks of spontaneous genius, but by a process of blending pre-existing notions. Whether this happens inside a single person's head or in a social context doesn't seem to matter.

Furthermore, social innovation is always best activated by play and many of today's scientific workplaces are still lamentably low on playfullness. Whereas software companies have incorporated all manner of tools for establishing a sense of fun into their offices, many scientific departments still imagine that it's somehow 'more professional' to have people sitting in silence in small offices.

So, is there something we can do to fix this? Can we use applied improv to make science healthier, smarter, and more playful? I think so, and here's my best guess so far as to how to do it.

The trick is to play the science incubator game. For this you need:
  • A cafe.
  • A volunteer to be master of ceremonies.
  • About six people who like to learn and think. 
The science incubator game takes the form of an open, flowing dialog. An ideal session is likely to last from one to two hours, depending on everyone's stamina, and works like this.
  • One person, the 'proposer' brings along an open question that they're trying to answer. This can be as abstruse and as deep into their work as they like. In fact, the more abstruse, the better. The proposer tries to explain their problem to the rest of the group.
  • The others ask questions every time what they're hearing isn't clear, and chuck in any ideas that seem relevant. Everyone else in the group to tries to learn, and be as supportive as possible.
  • The master of ceremonies acts as an adjudicator, making sure that everyone gets a voice and that the rules are followed.
  • The session should start with everyone in the group telling a deliberate lie, so as to activate the creative parts of their brains. The more confusing or elaborate the lies are, the better. 
The rules of the game are:
  • All questions and ideas are good. Nobody gets to pass judgement on anyone's question or idea, regardless of how flawed they think it is. The proposer should try to answer all questions that are asked. (Why: Uses the Yes-And principle to create a shared narrative.)
  • Interruption should be done politely, but is mandatory. If the dialog has gone on for five minutes without somebody chiming in with a thought or suggestion, the master of ceremonies asks a question of his own. (Why: Prevents grandstanding and encourages group ownership of the process.)
  • Silence is banned. If a silence lasts for more than five seconds, the master of ceremonies should chime in with a new question. (Why: To maintain the energy level.)
  • Everyone is equal. All work hierarchy is left at the door when the incubator game is in progress. Anyone who pulls rank, or attempts to refer to their depth of experience to validate a point gets an immediate reprimand from the master of ceremonies. (Why: To help the space feel safe to all, and removed from normal patterns of social cost.)
  • Negativity is banned. In academic settings, people often consider their value to be in filtering out the proposals that won't work. In the incubator game, this role is forbidden. The way to add value is to add more ideas. (Why: Prevents contributors from self-awarding value via 'critical rationality' and derailing the session at the same time.)
  • Everything is informal. The purpose of the incubator is to reduce the risks of making suggestions and asking questions. Everyone is there to learn. Humor is strongly encouraged. (Why: Laughter activates the signal for social learning.)
  • Everyone must contribute. If someone has been quiet, or has been left out of the process, it's the master of ceremonies's job to bring them back in and make sure they feel safe. (Why: To encourage acceptance of the 'price of entry' of the session, which is engagement.)
  • Everyone should try to insert at least one harebrained suggestion that they have just thought of without considering the implications properly. Anyone who confesses 'I have no idea what I'm talking about' should get an immediate cheer. (Why: To break the dangerous social habit of over-filtering ideas out of perceived risk.)
  • The dialog stays on topic until the session is over, and shouldn't deviate into gossip. (Why: To break the idea that 'shop talk' is somehow dull, and to create as much engagement in new ideas as possible.)
  • Nobody is 'on show'. If the master of ceremonies feels that the game is dissolving into a performance of sorts between a few people in the group, or if people are waiting to have a 'good idea' before chiming in, the master of ceremonies needs to fix the balance. If necessary, the master of ceremonies can ask a specific person in the group for 'a half-baked idea, please'. (Why: To prevent social dominance patterns from forming within the game.)
  • This game doesn't have to be played in a cafe, of course, but ground that feels neutral and safe is a good idea. At someone's house over a shared pizza would work equally well. A glass of wine might also assist the process. 
  • Also, it clearly doesn't need to be done by scientists, either. I suspect that any group of people who're looking to share ideas and brainstorm would probably benefit from something like this. 
  • Leaving out rules that aren't working, and letting the proces flex to reflect the needs of the group is a good idea, so long as the spirit of the rules is maintained. 
Ways to make the process stronger:
  • If everyone brings pencil and paper, and tries building pictures or mind-maps of what they're hearing while others are speaking, it's likely to broaden thinking and keep the energy level up. (Why: Using the visual parts of the brain is likely to increase engagement.) 
  • Courageous, committed scientists should wear unusual hats, or pin large, ridiculous flowers to their lapels, or some such thing. Any symbol that creates a sense of group inclusion and of willingness to be silly is commended. (Why: Anything which fosters laughter is likely to improve the quality of social learning, and the bond that's formed.) 
  • If a proposer focuses on recent work they've done which went badly wrong, and talks over what happened with positive input from others, this is likely to help boost the proposer's learning experience significantly. (Why: Research shows that negative reinforcement without emotional consequences maximizes learning rates.)
  • Wind down time after the game is probably a good idea. If people get to keep talking and laughing after the end of the session, it's likely to improve the sense of group connection. (Why: The game is likely to encourage intensity. People will feel the need to re-establish working norms afterward.)
I've no idea whether this game works or not because I haven't tried it yet. To some extent, this game simply outsources the difficulties of scientific collaboration to the master of ceremonies--a kind of magical mediator who requires the courage and wisdom to step in every time the dialog falls off track. However, finding such a mediator might not be so hard. The role that I've outlined here is something that every decent improv instructor knows how to do. The necessary skill sets are surely easy to find if we go looking for them in our communities or take the time to acquire them for ourselves.

I'll be running some experiments here in Berkeley if I can find enough brave scientists. If anyone has any suggestions on the format, or gets to try it out before I do, I'd be delighted to hear about it.

Tuesday, February 1, 2011

TED Talks I Like

I like TED talks. They’re a marvelous way of getting access to many fascinating ideas in a very short time. (They’re also a fascinating series of examples of what does and does not make for compelling public speaking, but that’s a whole other blog entry.) At the suggestion of my good friend at SFBehaviorLab, David Sals, I’ve put together a list of some of the talks I like.

First, here are some talks by authors I’ve already raved about in my recent reading list post. These people therefore need no introduction. 

And here are are a few that I’ve recently encountered, all of which I think deserve a viewing. 

This has to be a great place to start. Clearer evidence of the profound neurological effect of improvisation would be hard to find. 

This is a great one to watch after the Dan Ariely talk listed above. Laurie Santos clearly demonstrates that the kind of decision-making ‘mistakes’ we make aren’t specific to the human race. This suggests that these patterns of reasoning are old. To my mind, this has important implications.
The patterns of decision-making that behavioral economics has revealed don’t just tell us things about how people react. They’re very likely to be providing us with important insights about how effective reasoning works. These ‘mistakes’ have been selected for over the course of millions of years of evolution. If they cause us to make some choices ineffectively, there must be other advantages that we gain. Though it may not be clear yet exactly what those gains are, experiments in Machine Learning are likely to help us find out. 

What I like about this talk is the distinction that Prof. Kahneman makes between the way we experience things and the way we remember them. He points out that the connection between the two is far shakier than we’d like to imagine. For me, this says interesting things both about the nature of declarative memory, and how we can use it to make our interactions with each other better. For instance, it seems clear that following negative feedback to a person with something a little nicer is likely to cause that person to walk away with a far rosier impression of the experience than if only negative input is received. This suggests that fixing some toxic workplace interactions may be as simple as bolting positive rituals onto the end of them--a fascinating implication, if it’s right. 

This talk is on how we can use an understanding of social networks to gain insights about the spread of diseases, social trends, and even emotions. Most significantly, Prof. Christakis reveals a simple mechanism by which we can identify ‘hubs’ in social networks and use them to gain advance warning of changes sweeping through a population. However, he also shows that interacting with these hubs provides us with a way to intervene as well as to watch. For instance, it tells us how to best to deploy a vaccine into a population to save the maximum number of lives.
The implication for applied improv here is that the same tools enable us to find those members of a community most likely to A: reflect the values of a culture, and B: change them, if we can engage them and give them the right tools. 

Mr. Johnson’s research into the the kind of social environments that foster good ideas feel like a natural fit for applied improv. Lurking in here, I feel, are clues as to how to use the science of play and the study of behavioral games to create innovation incubators. This talk leads me to wonder what kinds of improv you can play sitting down with four molecular biologists in a Starbucks without having anyone raise their voice or leave their seat. My suspicion is that one can do quite a lot, and probably get some publication worthy material out of it at the same time. 

Monday, January 24, 2011

Books I Like

On my literary travels last year, I came across lots of books that helped me build a stronger understanding of how improv works in the brain. Not all of them look relevant to applied improv at first sight, so I decided to put together a short reading list of a few of my favorites for anyone interested in exploring the same topics. It's my hope to expand this list in future posts.

Influence: The Psychology of Persuasion (Collins Business Essentials)
Robert Cialdini
An astonishing book. In essence, it’s a guide to the operating system of human social behavior. Cialdini reveals ways that human beings run on automatic while trying to get along and shows how those behaviors are routinely exploited by the unscrupulous. This book is invaluable for anyone interested in not being manipulated by others, but is also incredibly useful from an improv standpoint. The chapter about the ‘authority principle’ is essentially a lesson on Status. However, there’s a lot more in here that improv hasn’t explored as deeply. The ‘reciprocity principle’, for instance, has a lot in common with ‘make your partner look good’ but seems to go deeper. The research that Cialdini recounts suggest a wealth of possible games that have yet to be explored.

Influencer: The Power to Change Anything
Patterson, Grenny, Maxfield, and McMillan
The best book I’ve read so far on enabling social change. Simultaneously readable and scholarly--this book encourages a data-driven approach to understanding organizational culture, and doesn’t pull punches about just how hard it can be to make a lasting difference. It outlines a clear plan of the steps that leaders need to take if they really want to mend the communities they work in.
There’s plenty here too for applied improvisers who don’t happen to be working directly with business. ‘Influencer’ provides a useful guide to the effect of human motivations on group behavior and draws on real life examples like drug rehabilitation programs and disease prevention projects in the developing world.

Predictably Irrational, Revised and Expanded Edition: The Hidden Forces That Shape Our Decisions
Dan Ariely
An approachable, clearly written book on the growing field of behavioral economics. While there are other books out there that also do a good job at introducing this material (eg: Sway by the Brafman brothers), Predictably Irrational lays out each important result in a clear and concise way.
For those out there who haven’t looked into behavioral economics, I highly recommend exploring it. It sheds a great deal of light on many quirks in the human decision making process--such as why we like free gifts so much, and how the credit crunch happened. This new field is rife with experiments that cry out for adaptation into improv games.

On Intelligence
Jeff Hawkins
This is a book on how the human neocortex works. Mr Hawkins wants to duplicate the learning system it employs and use it in software to create intelligent machines. His company, Numenta, is making great headway in this department, and has already developed software for motion detection and fraud analysis based on insights from biology.
At first sight, this might not seem like a book for applied improv enthusiasts, but in fact it was one of the most important books I read last year. It makes it very clear exactly what the brain does that’s so special, and how human learning actually works. Locked in here is the secret of why ‘I suck and I love to fail’ is such an important concept.

Drive: The Surprising Truth About What Motivates Us
Dan Pink
A friendly, highly digestible account of motivation theory research. While this book has some things in common with Influencer, its thrust is more inspirational in tone. Dan Pink shows how the principles of Autonomy, Mastery, and Purpose broadly define what people want out of life, and how some of the most enlightened business leaders in the world have been able to put those ideas to work.
In essence, the book is an appeal to managers to stop thinking in terms of cash incentives and old-fashioned economics, and to use modern psychology instead. Gratifyingly, his message lines up tightly with the kind of motivational wisdom that improvisers have been using for a long time. There are plenty of examples that trainers can grab hold of and apply directly with their clients.

Play: How It Shapes the Brain, Opens the Imagination, and Invigorates the Soul
Stuart Brown
This book is a must for applied improv enthusiasts. It lays out research that shows how the act of playing activates the oldest, most highly evolved system for learning that human beings have. The message is clear: training that doesn’t incorporate play isn’t really training. Sure, it might be informative, and even slightly useful, but nothing enables soft skill acquisition like the collaborative social experimentation that’s signaled by laughter.
This book very successfully explodes the myth that play is somehow trivial and that real business is serious, and reveals that the reverse is usually true. Effective businesses, communities and institutions leave room for play and laughter, while trying to be ‘serious’ tends to lead to impaired decision-making.

Thursday, January 20, 2011

Open Tokenomics Event--Everyone Invited

Tuesday, Feb. 15th, 6-8pm
1590 Bryant St., San Francisco
This is a free event

Good news for everyone out there interested in experimental applied improv, and particularly for those who happen to live in the Bay Area. The SF Behavior Lab project will be holding its first public open event in San Francisco next month--on Feb 15th to be precise. The event is open to all, and will happen from 6pm to 8pm. Afterward, we’ll be going to grab some dinner in the neighborhood and you’re welcome to join us for that too.

The evening will require no improv training and will hopefully have something to offer both interested laypeople and experienced improv trainers looking for new workshop ideas. There’ll be an opportunity for improv professionals to share experimental games in a safe setting, and also plenty of new games to try out. While we build up steam, we'll be keeping the Behavior Lab events free, so make sure you invite anyone you know who’s looking for a fun evening out.

The focus of this first session will be on Tokenomics--the topic I gave a talk on at the world applied improv conference in Amsterdam this year. When I spoke again at the Bay Area mini-conference in December, we didn’t have enough time to explore games that addressed this topic. Next month’s session will hopefully address that imbalance, and give everyone in the area a chance to try the material out for themselves.

If you don’t yet have a clear idea of what Tokenomics is or what I’m talking about, the good news is that you can now find out through the Behavior Lab website. This talk is a little rushed (I was trying to squash about 45 mins worth of slides into about 25 mins) but will give you a sense of what’s in store. It should be a very entertaining evening, and hopefully highly informative too.

The Feb. 15th event will be held at Sports Basement, at 1590 Bryant St. in San Francisco. Sports Basement has generously offered a 20% store-wide discount that evening for anyone connected with SF Behavior Lab, so be sure to come early and browse the aisles.

PS: For those of you who have already signed up at the Behavior Lab website, please take another look and check that your membership still exists. The site encountered some minor teething troubles at the end of last year, and some attempted sign-ups were lost. (We want everyone who wants to play with us to get to!)

Hope to see you all there!