homethe consortiumabout cspoour libraryprojectseducation & outreachsearchcontact us

 

Projects


 

Distributional Impacts of Science and Technology Policies

Recently, a good deal of attention has been devoted to the “digital divide" - the influence of socioeconomic status on access to computers and information technology. Given the impact that information technology has on access to learning and many occupations, this focus is most welcome, but it is insufficient. An array of equally apparent income-based disparities in access to and impacts of S&T have not received equal attention. There is certainly a digital divide, but there is also an MRI divide, a consumer electronics divide, a reproductive technology divide and an injection molded polymer divide, to name a few.

The purpose of this research project is to develop knowledge about the factors affecting the distributional impacts of S&T. The focus is on factors internal to science and technology knowledge production processes rather than the social factors that mitigate distributional impacts (e.g., income inequities; lack of universal health care). Amazingly, little or no research of this nature has been done. Using a combination of survey research, historiography and intensive case studies, the research seeks to isolate those attributes of S&T associated with access and benefit distribution. The results of the study will support public policy designs that can promote more equitable outcomes from S&T. This would represent a different set of priorities for science policy, a field that has been dominated by a focus on enhancing research budgets and economic growth, with insufficient attention paid to distributional concerns.

 

Publications and Conference Presentations

 

Bozeman, Barry. 2002. Public-Value Failure: When Efficient Markets May Not Do. Public Administration Review 62, 2: 145-161.  

The familiar market-failure model remains quite useful for issues of price efficiency and traditional utilitarianism, but it has many shortcomings as a standard for public-value aspects of public policy and management. In a public-value-failure model, I present criteria for diagnosing values problems that are not easily addressed by market-failure models. Public value-failure occurs when: (1) mechanisms for values articulation and aggregation have broken down; (2) “imperfect monopolies” occur; (3) benefit hoarding occurs; (4) there is a scarcity of providers of public values; (5) a short time horizon threatens public value; (6) a focus on substitutability of assets threatens conservation of public resources; and (7) market transaction threaten fundamental human subsistence. After providing examples for diagnosis of public-values failure, including an extended example concerning the market for human organs, I introduce a “public-failure grid” to facilitate values choices in policy and public management.

 

Bozeman, Barry, 2003. Science, Technology and Culpability: Some Hypotheses about Why the Disadvantaged Benefit Less from Discovery and Innovation. Presentation prepared for the Annual Research Conference of the Association for Public Policy Analysis and Management, Washington, D.C., November, 2003.

 

Bozeman, Barry and Daniel Sarewitz, 2005. Public Values and Public Failure in US Science Policy. Science and Public Policy, 32(2), 119-136.

Domestic science policy in the United States is linked inextricably to economic thinking. We seek to develop a practical analytical framework that confronts the manifest problems of economic valuing for science and technology activities. We argue that pervasive use of market valuation, market-failure assumptions and economic metaphors shapes the structure of science policy in undesirable ways. In particular, reliance on economic reasoning tends to shift the discourse about science policy away from political questions of “why?” and “to what end?” to economic questions of “how much?” Borrowing from the “public values failure framework”, we examine public values criteria for science policy, illustrated with case vignettes on such topics as genetically modified crops and the market for human organs.

 

Bozeman, Barry and Paul Hirsch, 2003. Science, Technology and the Distribution of Impacts: Alternative Theories of the ‘Handicapper General’. Presentation to Annual Meeting of the Society for Social Studies of Science, Atlanta, Georgia, October 15, 2003.

 

Barry Bozeman and Paul Hirsch. 2004. Science and Institutionalized Ethics: Prospects for Protecting Human Subjects. Paper prepared for presentation at the European Association for the Study of Science and Technology, August 25-28, Paris, France.

Scientists whose work has potential relevance for vulnerable and disadvantaged populations are in a sort of catch-22:  if they ignore the needs of minorities, the poor, or the disabled, scientists are accused of being a part of an inequitable capitalist structure where the rich get the benefits of the best science, and the rest get the leftovers.  But if they focus their scientific lens directly on the needs of the most vulnerable, even the most well-intentioned scientists are at risk of being accused of treating people as human guinea pigs. Sometimes such accusations are accurate, sometimes they’re questionable.  Either way, every accusation serves as a wedge between scientists and members of at-risk populations.  The purpose of our paper is to use a case study approach to shed light on some of the limitations of “institutionalized science ethics” for protecting human subjects and for avoiding the kind of negativity generated by many of these cases, and to argue for a supplementary approach that might, in cases where ethical edge is the sharpest, offer a method of adjudicating proposed research that would better protect both subjects and researchers.

 

Bozeman, Barry, and Paul Hirsch. 2006. Science ethics as a bureaucratic problem: IRBs, rules, and failures of control. Policy Science.

“Institutionalized science ethics” refers to the statutory, professional and institution-based ethical standards that guide and constrain scientists’ research work. The primary institution responsible for implementing institutionalized science ethics is the Institutional Review Board. We examine the limitations of IRBs and institutionalized science ethics, using bureaucratic theory and, especially, theory related to the development and enactment of rules. We suggest that due to the very character of rules-based systems, improvements in IRB outcomes are unlikely to be achieved through either more or better rules or even by bureaucratic reform. Instead, we suggest that improvements in human subject protection can best be advanced through increased participation. Ours is not a call for more participation by the general public but participation, via “Participant Review Boards” of persons who are eligible, by the protocols of the research in question, to serve as subjects. This provides a level of legitimacy and face validity that cannot be obtained by IRB affiliates, even by “external representatives.” In making these points, we review a recent science ethics controversy, the KKI/Johns Hopkins lead paint study. In spite of being approved by IRBs, the study resulted in a civil lawsuit that reached the Maryland Court of Appeals. The case illustrates the limits of institutionalized science ethics and the bureaucracies created for their enactment. The case also underscores the complex and equivocal nature of the ethical guidelines established under the National Research Act.

 

   Feeney, Mary Kathleen and Barry Bozeman, 2006. Public Values and Public Failure:          

Implications of the 2004-2005 Flu Vaccine Case. Public Integrity.

Public interest concerns within policy analysis have been marginalized by the popularity of economic individualism and the growing reliance on market failure models to analyze public policy issues. The public interest should be the center of policy analytical frameworks. This case study applies a public values framework to the 2004-2005 influenza vaccine shortage to illustrate the potential for infusing public values in analysis of public policy controversies. The public values framework leads to a quite different, and we think a preferable, focus on policy analysis and forecasting compared to predominant economic thinking, especially market failure criteria. The case of flu vaccine shortage is especially apt inasmuch relatively few of the determinants pertained to economic forces or to market criteria. Public health is a classic instance of public goods, perhaps even more so than defense and national security, because the level of one’s threat is in large measure a function of the threat to others. 

 

   Hirsch, Paul, 2003. Shattering the Glass: Scientific Research, Societal Inequality, and Formalized Review. 2003 Research Value Mapping Conference. Atlanta, GA.

Scientific progress and technological development are often viewed as primary forces in the growth and development of the American economy (Kleinman 1995, Brooks 1996, Guston 2000).  Technology boosts productivity and creates new markets; science cures diseases and expands the range of human possibilities. The jury is still out, however, on the distributional implications of scientific and technological progress: whether the benefits generated by science and technology will serve to increase or decrease the gap between the haves and the have-nots.  Given a pre-existing state of inequality, the interaction between vulnerable populations and scientific researchers can take one of several forms: researchers can ignore vulnerable populations; they can exploit them; or they can attempt to ameliorate some of the disadvantages accrued to vulnerable populations.  These three possibilities provide the structure for this paper. After drawing on previous research and historical examples to explain each of the possible interactions between research and inequality, we use them as the organizational principle for subsequent analysis.  We argue that researchers’ ideas regarding the underlying causes of inequality – whether stated or unstated, reasoned or unreasoned – play a major role in determining whether scientific research ends up ignoring, exploiting, or ameliorating pre-existing inequalities, regardless of the researchers’ original intent.

 

   Hirsch, Paul. 2004. Science and Institutionalized Ethics: The Johns Hopkins Lead Paint Study and its Implications for Protecting Human Subjects. Georgia Tech Faculty Seminar, Atlanta, GA.

 

   Sosinska, Olga, and Mary K. Feeney. 2006. Working Paper. Public Value Mapping: The Cooperative Research and Development Agreement between Yellowstone National Park and Diversa Inc.

 

 

 

Science, Policy & Social Inequity Workshop sponsored by the W.K. Kellogg Foundation and the Consortium for Science, Policy & Outcomes.

Read more



Privacy Policy . Copyright 2007 . Arizona State University
Consortium for Science, Policy & Outcomes
College of Liberal Arts and Sciences
P.O. Box 874401, Tempe AZ 85287-4401, Phone: 480-727-8787, Fax: 480-727-8791
cspo@asu.edu