They discuss the problems with current representations of disability, recent examples of policy that has failed disabled people and the changes that could be made so people with disabilities can be better supported and allowed to participate in policy making.
Elizabeth C. Long, Rebecca L. Smith, Jennifer T. Scott, Brittany Gay, Cagla Giray, Shannon Guillot-Wright and Daniel M. Crowley
Want to conduct surveys with national-level policymakers about their research use, but not sure how? We at the Research-to-Policy Collaboration offer a new measurement protocol to understand the role of science in national-level policymaking and provide lessons we learned based on our experiences surveying congressional staff in the US.
The implementation of public policies is a process that is as complex as its formulation, especially when we set out to solve issues in communities whose codes and languages are not shared by policymakers.
For example, Brazil is a country of continental proportions. Its population and socio-political contexts are very diverse. In the 2010 census, 305 indigenous ethnicities were identified, ranging from peoples living under voluntary isolation to groups living in major cities, spread across the 27 Brazilian states. Such diversity engenders a multitude of viewpoints about socially relevant problems, which universal public policies cannot cover without proper adaptations.
After a period in which the onward march of evidence-informed decision-making appeared to be faltering in countries such as the US and UK, the acute uncertainties of the COVID-19 pandemic have triggered a fresh explosion of engagement with evidence and policy interactions – from diverse disciplinary, sectoral and institutional perspectives.
Does scientifically-backed information capture the attention of policymakers? To test this, I conducted a field experiment embedded in a real-life advocacy initiative targeted to members of the European Parliament in April 2018. As described in my Evidence & Policy article, ‘The silver bullet reversed: the impact of empirical evidence on policymaker attention’, results indicate that ideas-based information, rather than empirical information, gathers more attention from policymakers. More precisely, it is the announcement of ideas rather the actual information which manages to capture policymaker interest. Crucially, these findings hold across political groups, policy support and gender.
What makes experts legitimate in the eyes of policymakers? Even though this is one of the foundational questions of the interdisciplinary scholarship on evidence and policy, the answer is neither straightforward nor simple. Expert legitimacy is driven by seeming contradictions – experts have to be responsive to policymakers’ needs but, at the same time, they cannot be too close to politics. They have to provide advice which is strongly grounded in science but if their advice is too complex it risks being ignored or being perceived too ‘detached’ and ‘academic’. Experts are legitimate when they are insiders and outsiders at the same time. This dynamic has become particularly evident in the ongoing pandemic, where government advisors have had to represent (and at times defend) science whilst at the same time accounting for what policy directions are ‘doable’ – publicly and politically acceptable and economically feasible.
One of the major trends within the contemporary policy scene is ‘the use of behavioural insights (BI)’ to improve policymaking. All around the world, from Qatar to England and Japan, ‘Behavioural Insights Teams’ (or ‘BITs’), ‘Nudge advisers’ and ‘Chief Behavioural Officers’ now inhabit government, seeking to infuse it with state-of-the-art knowledge and methods from the behavioural sciences. The more specific signature traits of this BI agenda appear to be its focus on new behavioural economics, nudge techniques and Randomized Controlled Trials (RCTs). The COVID-19 crisis hasn’t hampered the behavioural momentum – quite the contrary: in the absence of a distributed vaccine, halting the spread of the coronavirus has very much been a behaviour change challenge, with BI being in great demand. The recent launch of dedicated ‘COVID-19 Teams’ and ‘Corona Behavioural Units’ within the UK’s and Dutch policy scene didn’t come as a surprise, and only confirmed that behavioural government is here to stay.
Intriguingly enough, though, one question about the new institutional praxis of ‘using BI’ remains not yet convincingly answered: What is it, really?
Many of us place our hopes on innovative breakthroughs and groundbreaking discoveries, believing them to be our best bet to achieve a better world. And indeed, science has produced extraordinary breakthroughs. Vaccines radically reduced the risk of death from communicable diseases. Nitrogen-based fertilisers vastly increased the production of food. Computers completely transformed how modern humans learn, work and communicate. Surely, it would seem that investing in scientific breakthroughs is the key to progress. In this spirit, social scientists develop ‘evidence-based’ practices and policies and create hierarchies of evidence to determine ‘what works’. Many believe that if only science can produce enough evidence, discoveries will follow that can change the world – if only we can effectively compel others to accept them.
The UK Parliament performs key democratic functions holding the government to account by scrutinising policy, debating legislation and providing a venue for the public to air their views through elected representatives. Despite the key role of the UK Parliament in shaping government policy, for example in recent times on Brexit and COVID-19 (though many argue Parliament should have a greater role on the latter), scholars of science-policy interfaces have rarely explored how evidence is sourced and used in legislatures.