Regular readers will know that I’ve been thinking a lot about evaluation for many years now. I am not an evaluator, but almost every project I am involved in contains some element of evaluation. Sometimes this evaluation is well done, well thought through and effective and other times (the worst of times, more often than you think) the well thought through evaluation plan crumbles in the face of the HIPPO – the Highest Paid Person’s Opinion. So how do we really know what is going on?
When I stumbled across Michael Quinn Patton’s work in Developmental Evaluation, a whole bunch of new doors opened up to me. I was able to see the crude boundaries of traditional evaluation methods very clearly and was able to see that most of the work I do in the world – facilitating strategic conversations – was actually a core practice of developmental evaluation. Crudely put, traditional “merit and worth” evaluation methods work well when you have a knowable and ordered system where the actual execution can be evaluated against a set of ideal causes that lead to an ideal state. Did we build the bridge? Does it work according to the specifications of the project? Was it a good use of money? All of that can be evaluated summatively.
In the unordered systems where complexity and emergence is at play, summative evaluation cannot work at all. The problem with complex systems is that you cannot know what set of actions will lead to the result you need to get to, so evaluating efforts against an ideal state is impossible. Well, it’s POSSIBLE, but what happens is that the evaluator brings her judgements to the situation. Complex problems (or more precisely, emergent problems generated from complex systems) cannot be solved, per se. While it is possible to build a bridge, it is not possible to create a violence free society. Violent societies are emergent.
So that’s the back story. Last December I went to London to do a deep dive into how the Cynefin framework and Cognitive Edge’s work in general can inform a more sophisticated practice of developmental evaluation. After a few months of thinking about it and being in conversation with several Cognitive Edge practitioners including Ray MacNeil in Nova Scotia, I think that my problem is that that term “evaluation” can’t actually make the jump to understanding action in complex systems. Ray and I agreed that Quinn Patton’s work on Developmental Evaluation is a great departure point to inviting people to leave behind what they usually think of as evaluation and to enter into the capacities that are needed in complexity. These capacities include addressing problems obliquely rather than head on, making small safe to fail experiments, undertaking action to better understand the system rather than to effect a change, practicing true adaptive leadership which means practicing anticipatory awareness and not predictive planning, working with patterns and sense-making as you go rather than rules and accountabilities, and so on.
Last night a little twitter exchange between myself, Viv McWaters and Dave Snowden based on Dave’s recent post compelled me to explore this a bit further. What grabbed me was especially this line: “The minute we evaluate, assess, judge, interpret or whatever we start to reduce what we scan. The more we can hold open a description the more we scan, the more possibility of seeing novel solutions or interesting features.”
What is needed in this practice is monitoring. You need to monitor the system in all kinds of different ways and monitor yourself, because in a complex system you are part of it. Monitoring is a fine art, and requires us to pay attention to story, patterns, finely grained events and simple numbers that are used to measure things rather than to be targets. Monitoring temperatures helps us to understand climate change, but we don’t use temperatures as targets. Nor should we equate large scale climate change with fine grained indicators like temperature.
Action in complex systems is a never ending art of responding to the changing context. This requires us to be adopting more sophisticated monitoring tools and using individual and distributed cognition to make enough sense of things to move, all the while watching what happens when you do move. It is possible to understand retrospectively what you have done, and that is fine as long as you don’t confuse what you learn by doing that with the urge to turn it into a strategic plan going forward.
What role can “evaluation” have when your learning about the past cannot be applied to the future?
For technical problems in ordered systems, evaluation is of course important and correct. Expert judgement is required to build safe bridges, to fix broken water mains, to do the books, audit banks and get food to those who need it. But in complex systems – economies, families, communities and democracies, I’m beginning to think that we need to stop using the word evaluation and really start adopting new language like monitoring and sense-making.
Share:
One has to be very careful attributing causes to things, or even attributing causality to things. in complex systems, causality is a trap. We can be, as Dave Snowden says “retrospectively coherent” but you can not know which causes will produce which effects going forward. That is the essence of emergent phenomena in the complex world.
But even in complicated problems, where causality should be straightforward, our thinking and view can confuse the situation. Consider this example.
Imagine someone, a man, who has never seen a cat. I know, highly implausible, but this is a hypothetical from Alan Watts’ book, On the Taboo Against Knowing Who You Are, which was written in the sixties; pre-YouTube. Watts uses this fictional fella to illustrate the unfelt influence of perspective and the dangers inherent in our strong inclination to seek cause-and-effect relationships.
“He is looking through a narrow slit in a fence, and, on the other side a cat walks by. [The man] sees first the head, then the less distinctly shaped furry trunk, and then the tail. Extraordinary! The cat turns round and walks back, and again he sees the head and a little later the tail. The sequence begins to look like something regular and reliable. Yet again the cat turns round and he witnesses the same regular sequence: first the head and later the tail. Thereupon he reasons that the event head is the invariable and necessary cause of the event tail which is the head’s effect. This absurd and confusing gobbledygook comes from his failure to see the head and tail go together; they are all one cat.”
We often create and embed the wrong patterns because we are looking through a slit. As Watts says, by paying very close attention to something, we are ignoring everything else. We try and infer simple cause-and-effect relationships much, much more often than is likely in a complex world. For example, making everyone in an organisation focus on hitting a few key performance indicators isn’t gong to mean that the organisation is going to get better at anything other than hitting those key performance indicators. All too often this will lead to damaging unintended consequences; absurd and confusing gobbledygook.
via abc ltd.
Share:
When I popped off to London last week to take a deep dive into Cognitive Edge’s work with complexity, one of the questions I held was about working with evaluation in the complex domain.
The context for this question stems from a couple of realities. First, evaluation of social programs, social innovation and other interventions in the human services is a huge industry and it holds great sway. And it is dominated by a world view of linear rationalism that says that we can learn something by determining whether or not you achieved the goals that you set out to achieve. Second, evaluation is an incredibly privileged part of many projects and initiatives and itself becomes a strange attractor for project planning and funding approval. In order for funders to show others that their funding is making a difference, they need a “merit and worth” evaluation of their funds. The only way to do that is to gauge progress against expected results. And no non-profit in its right mind will say “we failed to achieve the goals we set out to address” even though everyone knows that “creating safe communities” for example is an aspiration out of the control of any social institution and is subject to global economic trends as much as it is subject to discrete interventions undertaken by specific projects. The fact that folks working in human services are working in a complex domain means that we can all engage in a conspiracy of false causality in order to keep the money flowing (an observation Van Jones inspired in me a while ago.) Lots of folks are making change, because they know intuitively how to do this, but they way we learn about that change is so tied to an inappropriate knowledge system, that I’m not convinced we have much of an idea what works and what doesn’t. And I’m not talking about articulating “best practices.”
The evaluation methods that are used are great in the complicated domain, where causes and effects are easy to determine and where understanding critical pathways to solutions can have a positive influence on process. in other words, where you have replicable results, linear, summative evaluation works great. Where you have a system that is complex, where there are many dynamics working at many different scales to produce the problems you are facing, an entirely different way of knowing is needed. As Dave Snowden says, there is an intimate connection between ontology, epistemology and phenomenology. In plain terms, the kind of system we are in is connected to the ways of knowing about it and the ways of interpreting that knowledge.
I’m going to make this overly simplistic: If you are working with a machine, or a mechanistic process, that unfolds along a linear trajectory, than mechanistic knowledge (problems solving) and interpretive stratgies are fantastic. For complex systems, we need knowledge that is produced FROM the system and interpreted within the system. Evaluation that is done by people “outside” of the system and that reports finding filtered through “expert” or “disinterested” lenses is not useful for a system to understand itself.
Going into the Cynefin course I was interested to learn about how developmental evaluation fit into the complex domain. What I learned was the term “disintermediated sensemaking” which is actually the radical shift I was looking for. Here is an example of what it looks like in leadership practice.
Most evaluation uses processes employing a specialized evaluator undertaking the work. The problem with this is that it places a person between the data and experience and the use of the knowledge. And it also increases the time between an experience and the meaning making of that experience, which can be a fatal lag with strategy in emergent systems. The answer to this problem is to let people in the system have direct experience of the data, and make sense of it themselves.
There are many many ways to do this, depending on what you are doing. For example:
- When clustering ideas, have the group do it. When only a few people come forward, let them start and then break them up and let others continue. Avoid premature convergence.
- When people are creating data, let them tag what it means, for example, in the decision making process we used last weekend, participants tagged their thoughts with numbers, and tagged their numbers with thoughts, which meant that they ordered their own data.
- Produce knowledge at a scale you can do something about. A system needs to be able to produce knowledge at a scale that is usable, and only the system can determine this scale. I see many strategic plans for organizations that state things like “In order to create safe communities for children we must create a system of safe and nurturing foster homes.” The job of creating safe foster homes falls into the scope of the plan, but tying that to any bigger dynamics gets us into the problem of trying to focus our work on making an impact we have no ability to influence.
- Be really clear about the data you want people to produce and have a strategy for how they will make sense of it. World Cafe processes for example, often produce scads of data on table cloths at the centre of the table, but there is often so little context for this information that it is hard to make use of. My practice these days is to invite people to use the table cloths as scratch pads, and to collect important data on post it notes or forms that the group can work with. AND to do that in a way that allows people to be tagging and coding the data themselves, so that we don’t have to have someone else figure out what they meant.
- Have leaders and teams pour over the raw data and the signification frameworks that people have used and translate it into strategy.
These just begin to scratch the surface of this inquiry in practice. Over the next little while I’m going to be giving this approach a lot of thought and try it out in practice as often as I can, and where the context warrants it.
If you would like to try an exercise to see why this matters try this. the next time you are facilitating a brainstorm session, have the group record dozens of insights on post its and place them randomly on a wall. Take a break and look over the post its. Without touching the post its, start categorizing them and record your categorization scheme. Then invite the group to have a go at it. Make sure everyone gets a chance to participate. Compare your two categorization schemes and discuss the differences. Discuss what might happen if the group were to follow the strategy implicit in your scheme vs. the strategy implicit in their scheme.
Share:
This afternoon I’m coming home after a morning running a short process for a church in Victoria, BC. The brief was pretty straightforward: help us decide between four possible scenarios about our future. Lucky for me, it gave me an instant application for some of the stuff I was learning in London last week.
The scenarios themselves were designed through a series of meetings with people over a number of months and were intended to capture the church’s profile for its future, as a way of advertising themselves for new staff. What was smart about this exercise was the fact that the scenarios were left in very draft form so there was no way they could be confused for a “vision” of the future. It is quite common in the church world for people to engage in “visioning exercises” to deal with the complex problems that they face, but such visions are doomed evermore to failure as the bigger organization is beginning to enter into a period of massive transformation and churches are suffering from all kinds of influences over which they have no control.
Visioning therefore is not as useful as selecting a lens through which the organization can make some decisions.
Each scenario contained some possible activities and challenges that the church would be facing, and the committee overseeing the work was charged with refining these down to a report that would, to use my own terms, be a collection of heuristics for the way the organization would act as it addressed future challenges.
Our process was very informed by some thinking I have been doing with Dave Snowden’s “Simple rules for dealing with complexity.” Notably principles about avoiding premature convergence, distributing cognition and disrupting pattern entrainment. Furthermore, the follow up work will be informed by the heuristic of “disintermediation” meaning that the team working on the project will all be working with the raw data. There is no consultants report here. The meaning making is still very much located with the participants.
So here was our process.
- At small tables of four, participants were given 5 minutes to read over the scenarios silently.
- We then entered a period of three 15 minute small group conversations on the topic of “what do you think about these scenarios?” Cafe style, each conversation happened with three different groups of people. I was surprised how much introduction was going on as people met new folks. The question was deliberately chosen not to be too deep or powerful because with a simple question, the participants will provide their own depth and power. When you have a powerful need, you don’t need to contrive anything more powerful than what people are already up for.
- Following the cafe conversations, a round of silent reflection in which people were given the following direction. “Express your preference for each of the scenarios on a scale of 1-7. Seven means “Let’s do it” and one means “No Way.” For each scenario write your preference on your post it and write a short sentence about the one concrete thing that would make your vote one point higher.” So there is lots in this little exercise. First it’s a way of registering all of the objections to the scenarios without personalizing them. Secondly it gets at concrete things that the team can do to improve scenarios and third it harvest preferences and not simple yes/no decisions which are not appropriate for this kind of work.
- At each table someone gathered all the posts its of the same colour and by colour folks came to the front and placed them on the scale. Doing it this way meant that no one was sure whose preference was going where and it also meant that people couldn’t revise their post its once they saw how the preferences were being expressed.
The whole thing took about 75 minutes.
The result of this sense making was the chart you see above. Two hundred pieces of finely grained information ordered by the people themselves. The project team now has at least three things they can do with this material.
- They can recreate the scale, as each post it is colour and preference coded. That way they have a rough idea of the scenario with the greatest support, and they can show anyone who wants to see metrics where we stand on the proposals.
- They can cluster post its for each scenario according to “work that will make it better” which means they don’t have to pay attention to the scale. The scale is completely subjective, but each of these post-its contains one piece of concrete information to make the scenario better, so in some ways the numbers don’t really matter. They can cluster these ideas by each scenario AND they can re-cluster them by each topic to give an idea of overall issues that are happening within the organization.
- If we wanted to go a step further, we could use these post it notes to do a number of Cognitive Edge exercises including a Cynefin contextualization (which would tell us which things were Obvious, Complicated and Complex (and maybe Chaotic) and we could also do some archetype extraction which might be very useful indeed for constructing the final report, which would stand as an invitation to thier new personal and an invitation to the congregation.
Share:
I had the great pleasure of coaching a team of folks last night who were running their first World Cafe. I’ve been working with this crew for a while – a core team looking at the future of the Victoria Presbytery of the United Church of Canada – and this was the first time they’ve stepped up to run their own conversational process as part of our work. Last night it was a Cafe to sense the future of what the Presbytery could be and do. And they did great.
One of the advantages of coaching is that one gets to reflect on the little bits and pieces of practice that make things work. Last night a number of them came up, so I thought I’d share them here.
Give instructions one at a time. Don’t give a long list of instructions. At the beginning of the Cafe let people know how the time will flow, but when it comes time to invite people to do certain things (move between tables, change questions, reflect, summarize…whatever) just give one instruction at a time. It is important that people know WHY we are doing a thing, but not important that they have the whole flow. And especially if your instruction involves them moving, then don’t give any more instructions until they have stopped.
Invite people to mark the paper early. The paper in the middle of the table is for all to use. “Typical” facilitated sessions imprint people with the pattern that someone will take notes while everyone else talks. It’s important that before the conversation begins, you invite people to pick up a marker, write something and draw something on the paper in front of them. That way, before the conversation begins, folks know that the paper is for everyone to use, there is no top or bottom, and images and words are equally welcome.
Have one more marker and one fewer post it note than people. If you have tables of four, give them five markers. This means that people can trade colours without prying a marker from someone’s hand. And if you are summarizing key findings, have three post-its for a table of four, to encourage people to pick three things together rather than just having everyone put their best thought down. World Cafe is about tapping and making visible collective intelligence. You lose that if you just have individual thoughts.
Build in silence. At the conclusion of a round, have a minute or two of silence. It calms the room down, allows people to reflect and integrate what they are hearing and makes it easier to give directions. This is especially important if you are wanting people to raise their level of awareness from what is important personally to what patterns are emerging. It requires a shift in awareness to see that.
Collect post its before having a summary conversation. Last night we used post its at the conclusion of the third round to capture the patterns that people were hearing consistently in all three rounds. Collecting the post-its before we had a summary conversation meant that people couldn’t “report out” and instead we hosted a “conversation with the whole” whereby we roved around asking people what stood out for them. What emerged was indeed a conversation and not a boring reporting out of things that everyone knew anyway.
Avoid the temptation to use a different question for each round. This is important. Having a different question for all three rounds creates three shallow conversations and inhibits pattern finding. It can also leave people feeling like they are being led down a garden path and it doesn’t leave a lot of space for emergent conversation. For all Cafe beginners, I always suggest they do their first Cafe with a single question for all three rounds. This gives you a clear picture of how the process can work to surface COLLECTIVE intelligence.
Keep the question simple and broad and make sure you can answer it on your own. Trust the group. They want to have a conversation, not guess at answers that you are trying to get them to. Last night our question was simple; given a context in which the structures of the Church are becoming increasingly unsustainable and in which congregations still need to be connected on a local level “What should Presbytery be and what should it do?” That was it. Three rich rounds on that, with lots of great insight and some amazingly courageous admissions (“Time to finally admit that this structure is dead.” etc.)
Invitation matters. Even though the 50 people we had out last night are used to being together every few months, the core team mworked on their invitation for a month. They held the purpose of the event close (discovering what the new shape and function of the Presbytery could be) and they shared the question with participants, even before we had decided on what the final question was. The team made sure people RSVP’d on the invitation which helped us to know the logistics of food and space, and also gave a chance for the conversation to begin as folks started sharing what they were thinking right away. This primed the conversation and meant that people were really ready for the work. Ninety minutes was not enough.
Know what you will do with the harvest and tell people. People learned in the invitation what our plans were for the harvest. This even was about helping the core team design some experiments over the next year for new ways that the Presbytery could meet and be useful to the two dozen United Church congregations on southern Vancouver Island. We summarized the patterns that people found (above photo) and began right away writing a report. But the bigger piece of work will be engaging in design over the next couple of months to create new and interesting gatherings in line with what the Presbytery members actually want.