twitterFacebooklinkedin

How can a rethink of lessons from field experiments inform future research in transparency, participation and accountability?

Date: May 29, 2019
Author(s): Jonathan Fox
Publication type: Blog
Published by: International Initiative for Impact Evaluation (3ie)

Strategists in the field of transparency, participation and accountability are rethinking overly optimistic theories of change. How can a reassessment of existing studies help to inform this needed reboot? What do ‘mixed results’ from field experiments with transparency and accountability interventions tell us about where the field should go next? Classic field experiments to improve governance and service delivery often rely on administering modest doses of information to stimulate small-scale citizen action (in contrast to studies based on natural experiments, which can unfold at larger scales). Growing recognition that information is not a magic bullet is leading to efforts to understand why. This blog builds on a recent academic review of 48 studies involving information-led inventions, which concludes that the power to act on information matters. This suggests bringing power to the center of the analysis, rather than leaving it to be discovered as a finding. To offer a reality check to inform the ‘what works’ question in the field of transparency, participation and accountability (TPA), this blog offers three propositions for discussion:

  1. We should not be surprised when ‘low dose’ interventions lead to uninspiring impacts. Let’s keep in mind that when a medicine is administered in a very small dose, to little effect, we still don’t know whether the medicine could work. Many field experiments in the TPA field test tool-led, ‘light touch’ interventions. That is, they are brief, information-led, and localized – sometimes involving a small number of community ‘interface’ meetings. They tend to rely on information that is not clearly actionable, in contrast to a ‘targeted transparency’ approach.The ‘low dose’, exclusively local approach risks addressing symptoms rather than underlying causes of governance failures, which may also be ‘upstream’, at higher levels of governance (e.g. medicine stock outs, frontline worker absenteeism, or missing textbooks). Moreover, few field experiments address the ‘fear factor’, which for good reason inhibits collective action that might question authority.
  2. Some field experiments are based on ‘fuzzy proxies’, which leads to fuzzy results. Experiments that use very indirect indicators of participation or accountability may fall short of actually measuring what they say they measure. This leads to slippage between links in the causal chain that we postulate connects transparency, participation and accountability. To mention just a couple of examples of fuzzy proxies from influential studies, official informational posters on the walls of clinics are not indicators of community monitoring. Attendance at meetings says little about accountability, in the absence of assessments of deliberative processes. An influential study even assumed that the ostensibly perceived potential for collective action was functionally equivalent to the actual exercise of voice.Plus, some field experiments are fuzzy because they blur the difference between a failure to generate participatory oversight actions – and the failure of actual participatory actions to improve service delivery. The difference between these two problems is elided when study titles implicitly conflate nominal formal structures for participatory citizen oversight with actual accountability efforts. Moreover, indirect survey questions about hypothetical scenarios have been used as substitutes for empirical evidence that bottom-up accountability really happened – to then suggest that accountability efforts were ineffective at improving service delivery. Few experiments draw on the substantial existing literature on how to measure participation and empowerment. The T4D study stands out, insofar as it attends to the nuances by documenting the nature of participation.
  3. Large-scale evaluations have assessed only a narrow range of theories of change in the TPA field. The dominant experimental approach in the field of transparency, participation and accountability addresses only one of many possible theories of change – the ‘low dose’ inventions mentioned above. If one applies a political economy approach, this suggests a ‘crowding out’ effect that neglects the broader, more diverse mix of innovations that ‘action strategists’ are experimenting with in practice. For example, multi-pronged, multi-level strategies combine multiple tactics, which allow the whole to become greater than the sum of the parts, may have more potential to encourage pro-accountability power shifts. Plus, most field experiments are one-off, which excludes change strategies based on adaptive learning.

Examples of practitioner-led change strategies that are under-researched include:

These diverse change strategies, informed by power analysis and the search for state-society synergy, are certainly hard to measure. But they may be more impactful than ‘low dose’ experiments. These three propositions drawn from rethinking the experimental literature suggest the following takeaways. Broad-based, constituency-led change initiatives may be more promising than locally-bounded, tool-led approaches. Grounded campaigns may have more leverage than externally-designed interventions. Take the fear factor into account. This suggests proactive measures to bolster the enabling environment for collective action by reducing both actual and perceived risks associated with voice. Being strategic involves listening to strategists. This suggests co-designing change initiatives with actual change agents, who may well know more about how power is distributed and exercised than most scholars. To sum up, this is a good time to reassess both theories of change and how they are tested. Building in the power analysis up front is a good place to start. With input from Joy Aceron, Angela Bailey, Daniel Berliner, Julia Fischer-Mackey and Brendan Halloran.

Jonathan Fox Jonathan Fox is a professor in the School of International Service at American University. He studies the relationships between accountability, transparency and citizen participation. He is the director of the Accountability Research Center. His most recent publications include articles in World Development and the IDS Bulletin, and reports published by Making All Voices Count, U4: Anti-Corruption Resource Center and the Transparency and Accountability Initiative. His books include Accountability Politics: Power and Voice in Rural Mexico (Oxford 2007) and Mexico’s Right-to-Know Reforms: Civil Society Perspectives (co-editor, Fundar/Wilson Center 2007). He was a founding member of the International Expert Panel of the Independent Reporting Mechanism of the Open Government Partnership and currently serves on the boards of directors of Fundar (Mexico) and the Bank Information Center (DC). For online publications, see http://jonathan-fox.org/.