Start Submission

Reading: Philosophical Foundations for Citizen Science

Download

A- A+
Alt. Display

Essays

Philosophical Foundations for Citizen Science

Authors:

Kevin C. Elliott ,

Michigan State University, US
About Kevin C.
Kevin Elliott is an Associate Professor in Lyman Briggs College, the Department of Fisheries and Wildlife, and the Department of Philosophy at Michigan State University. His research focuses on the role of values in science and ethical issues related to scientific research. He is the author of A Tapestry of Values: An Introduction to Values in Science (Oxford University Press, 2017).
X close

Jon Rosenberg

University of Washington, US
X close

Abstract

Citizen science is increasingly being recognized as an important approach for gathering data, addressing community needs, and creating fruitful engagement between citizens and professional scientists. Nevertheless, the implementation of citizen science projects can be hampered by a variety of barriers. Some of these are practical (e.g., lack of funding or lack of training for both professional scientists and volunteers), but others are theoretical barriers having to do with concerns about whether citizen science lives up to standards of good scientific practice. These concerns about the overall quality of citizen science are ethically significant, because it is ethically problematic to waste resources on low-quality research, and it is also problematic to denigrate or dismiss research that is of high quality. Scholarship from the philosophy of science is well-placed to address these theoretical barriers, insofar as it is fundamentally concerned about the nature of good scientific inquiry. This paper examines three important concerns: (1) the worry that citizen science is not appropriately hypothesis-driven; (2) the worry that citizen science does not generate sufficiently high-quality data or use sufficiently rigorous methods; and (3) the worry that citizen science is tainted by advocacy and is therefore not sufficiently disinterested. We show that even though some of these concerns may be relevant to specific instances of citizen science, none of these three concerns provides a compelling reason to challenge the overall quality of citizen science in principle.

How to Cite: Elliott, K.C. and Rosenberg, J., 2019. Philosophical Foundations for Citizen Science. Citizen Science: Theory and Practice, 4(1), p.9. DOI: http://doi.org/10.5334/cstp.155
216
Views
120
Downloads
2
Twitter
  Published on 08 Mar 2019
 Accepted on 30 Mar 2018            Submitted on 10 Feb 2018

Introduction

On July 17–18, 2017, the National Science Foundation funded a workshop in Raleigh, North Carolina, on the topic of “Filling the ‘Ethics Gap’ in Citizen Science.” The workshop was motivated by the recognition that citizen science has become increasingly prevalent in a wide variety of different fields, but relatively little work has engaged with the ethical dimensions of citizen science. The workshop explored a wide variety of ethical issues related to citizen science, including concerns about treating citizen research participants with appropriate respect; avoiding unintended harms; creating appropriate oversight mechanisms; navigating concerns about data quality, privacy, and ownership; and creating equitable opportunities for a wide range of people to participate.

In addition to these narrowly defined ethical issues, however, some recurring issues raised at the workshop involved broader conceptual concerns about whether citizen science has features that make it unlikely to meet standards of sound scientific practice. Specifically, academic scientists and community members involved in citizen science projects felt that their work was hampered by widespread perceptions that it could be lacking in rigor or quality. These concerns are significant because it is ethically problematic when the quality of scientific work is severely misjudged. On the one hand, when low-quality research is inappropriately treated with too much respect, one result can be poor-quality decision making and a waste of scarce resources that could have been used to support better studies. On the other hand, when high-quality research is inappropriately dismissed or prevented from taking place, decision makers can be deprived of valuable information that could benefit society.

The philosophy of science has long been concerned with questions about what distinguishes science from non-science or pseudoscience and what qualities are essential for good scientific practice. Thus, scholarship in this field has the potential to shed light on these concerns about citizen science. In this essay, we examine three prominent concerns that were raised at the workshop, and we show how the philosophy of science can provide resources for addressing them:

  1. The worry that, because citizen science often focuses on collecting large quantities of data, it can become an uninteresting exercise in “stamp collecting” rather than a properly scientific, hypothesis-driven enterprise;
  2. The worry that the data collected by citizens and the methods used in citizen science may not be of high quality;
  3. The worry that citizen science participants may fall into advocacy rather than maintaining a properly disinterested approach to their work.

While one or more of these concerns may pose a genuine problem in particular cases or situations, this paper argues that none of them provide in-principle reasons for questioning the quality of citizen science in general.

Concern #1: Data-Intensive versus Hypothesis-Driven Science

One worry is that citizen science tends to focus on the collection of data, and in many cases, these data are not associated with a particular scientific hypothesis. In some cases citizen science does incorporate specific hypotheses, such as when professional scientists collaborate with citizens to perform epidemiological studies of the effects of pollutants in their communities (e.g., Wing et al. 2008). In other cases, however, citizen scientists collect information without a specific hypothesis (or even a clear research question) in mind. Instead, they collect information about specific phenomena of interest to them or to regulators or monitoring agencies. For example, they might collect data about the presence of particular species in specific locations at particular points of time (Cohn 2008; Silvertown 2009), or they might collect information about the environmental quality of the air or water near them (Corburn 2005; Ottinger 2010), or they might monitor local temperatures or precipitation (Roy et al. 2012). Critics sometimes dismiss these kinds of studies, whether they are performed by citizens or professional scientists, as uninteresting “fishing expeditions” or exercises in “stamp collecting” (Boyd and Crawford 2012; Johnson 2007).

A recent essay by two program officers at the National Institute for Environmental Health Sciences (NIEHS), Liam O’Fallon and Symma Finn (2015), helps to sharpen these concerns. They identify important differences that they perceive between community-engaged research (which they view as being driven by academic scientists in most cases) as opposed to citizen science (which they view as being driven by citizens). They argue that community-engaged research is typically driven by a scientific question associated with the principal investigator’s expertise and interests. In contrast, they observe that citizen science is more likely to be driven by practical concerns, such as the desire to collect information about harmful environmental exposures. While O’Fallon and Finn do not use this observation to denigrate the importance of citizen science, it is easy to see how one could come to the conclusion that citizen science fails to meet standards of good scientific practice. Given that many accounts of scientific method revolve around the testing and proposal of hypotheses, the potential for citizen science to focus on collecting information without a clear hypothesis or research question seems to relegate it to a lesser status.

Despite the initial plausibility of this perspective, recent literature in both the history and philosophy of science challenges the idea that hypothesis-driven science is always of higher value than other approaches to science.1 First, the history of science illustrates that very important scientific advances were made by investigators who did not see themselves as functioning in a straightforward mode of proposing and testing hypotheses. For example, Isaac Newton insisted that his methodology was not based on the proposal of hypotheses but rather on inductive generalizations from the available evidence (Laudan 1981; Glass and Hall 2008). Newton’s suspicion of proposing and testing bold hypotheses was typical of most scientists throughout the eighteenth century. Natural philosopher Émilie du Châtelet stood out as an exception to most eighteenth-century thinkers when she defended the use of hypotheses and contended that Newton himself actually formulated hypotheses (Detlefsen 2014). It was not until the late nineteenth century that the “method of hypothesis” gained dominance, and then it received further support from the ideas of twentieth-century thinkers like Karl Popper (Elliott et al. 2016; Laudan 1981; Popper 1963).

A further problem with the notion that hypothesis-driven science is of more value than other approaches to science is that, according to a number of contemporary philosophers, science is best seen as an iterative endeavor that moves back and forth between multiple activities or “modes” (e.g., Chang 2004; Elliott 2012; O’Malley et al. 2009; O’Malley et al. 2010). On this view, scientists engage in multiple activities that are typically not arranged in a linear fashion (Kell and Oliver 2004). According to one philosophical account, scientists engage in at least four different “modes” of research: (1) exploratory inquiry; (2) technology-oriented research; (3) question-driven research; and (4) hypothesis-driven research (O’Malley et al. 2010). Thus scientists move iteratively between different practices as they seek to understand phenomena. Initially, they may have general questions that result in exploratory efforts to look for important patterns, regularities, or structures. To engage in this exploratory inquiry, they may have to develop new technologies (e.g., tools or methods). Over time, they are often able to narrow their questions, propose underlying mechanisms responsible for the patterns that they observe, and develop precise hypotheses that can be tested. As they study these mechanisms and test these hypotheses, they are often launched into new exploratory inquiries and succeeding rounds of questioning, technology-development, and hypothesis generation (O’Malley et al. 2010).

The upshot of these philosophical accounts of scientific practice is that one cannot evaluate the quality or appropriateness of specific scientific practices without considering how they are situated within a broader network of research activities (Elliott et al. 2016). Thus, even if most citizen science is not explicitly designed for testing hypotheses, it can generate questions and facilitate exploratory inquiries that turn out to be very valuable scientifically. For example, ecologists are becoming increasingly interested in studying the changes that occur at large spatial and temporal scales in response to development, pollution, habitat loss, and climate change (Heffernan et al. 2014; Soranno et al. 2014). Unfortunately, professional scientists often have difficulty finding the time and money to engage in the amount of data collection needed to answer these questions. Faced with this challenge, ecologists are finding monitoring data collected partly or wholly by citizen scientists to be an extremely valuable source of information (Poisson et al. in preparation).

Concern #2: Quality of Data and Methods

Another recurring concern about citizen science projects is that the quality or integrity of their data and the sufficiency of their methodologies could be compromised by the involvement of non-professional scientists (Riesch and Potter 2014). From a certain perspective, concerns about data quality and methodology are reasonable. Professionalization and long-term education in particular disciplinary frameworks are important processes through which scientists develop expertise for producing high-quality knowledge. Moreover, institutional incentives and disincentives (such as pressures to maintain one’s funding sources, employment, and reputation) provide important checks on the behavior of professional scientists. Relying on the data and work of non-professional scientists bypasses a large part of our institutional systems for quality assurance. Thus, a certain amount of care and attention to the quality of citizen science is appropriate.

Nevertheless, an important body of philosophical literature suggests that the quality of scientific work (including citizen science) cannot be adequately evaluated without considering the aims for which it is produced (e.g., Bhakthavatsalam and Cartwright 2017; Brigandt 2015; Elliott and McKaughan 2014; Intemann 2015). For example, when deciding how to model complex phenomena such as climate change, researchers must consider what questions they are trying to answer, at what levels of spatial and temporal resolution, and with what levels of accuracy (Intemann 2015). Similarly, when risk assessments are produced for the purposes of making regulatory or policy decisions, scientists have to balance the goal of producing reliable results against the goal of obtaining results in a timely fashion. In some cases, social costs could be minimized if scientists used less accurate assessment methods that generated results much more quickly (and thus facilitated faster regulatory decision making) but with a modest increase in false positives or false negatives (Cranor 1995). In the same way, when research is geared toward identifying potential public-health threats faced by communities, it might be appropriate to accept data or methods that are more prone to false positive errors in order to minimize the potential for false negatives (Barrett and Raffensperger 1999; Elliott 2014). The approach of using data and methods that are appropriate for the specific purposes for which they will be used is especially important when scientists or citizens are engaging in exploratory studies that can provide impetus for subsequent investigations.

Given this philosophical perspective (i.e., that data and methods must be evaluated based on the purposes for which they will be used), empirical evidence suggests that the quality of citizen science data has often been sufficient for the projects being pursued. For example, although some evidence suggests that volunteers engaged in reef monitoring consistently overestimate fish density, their data appear to accurately track long-term changes in coral reef ecology (Done et al. 2017; Forrester et al. 2015). Similarly, although volunteer contributions to European Union biodiversity monitoring appear to be less complete than expert-collected data, they also appear to be fairly accurate in general (Kallimanis et al. 2017). Citizen volunteers have also been found to provide data of equivalent quality to professional scientists and experts in the identification of marine debris, as well as equivalent data when identifying human impacts on the environment and land cover (See et al. 2013; Velde et al. 2017). In other studies, volunteers with appropriate training were able to accurately identify 94% or more of the insects that visited specific flowers (Ratnieks et al. 2016), and aggregated citizen-generated classifications of animals in photos agreed with expert classifications 98% of the time (Swanson et al. 2016).

When weaknesses in citizen science data have been identified, researchers have sometimes been able to tailor their investigations and hypotheses so that they take account of these limitations. For example, volunteers taking part in urban tree inventories are at least 90% consistent with expert identification across a number of areas, including site-type and genus-identification (Roman et al. 2017). However, they do occasionally miss or count extra trees (~1%), and they lag behind experts in the correct identification of tree species, agreeing with expert identification only 84% of the time. Yet, as Kallimanis et al. (2017) point out, that fact need only discourage researchers from using citizen data for projects that require fine-grained identification of trees; their data can be fruitfully used for a range of other projects.

In other cases, steps can be taken to strengthen the quality of data so that they meet the needs of particular scientific investigations (Riesch and Potter 2014). For example, the quality of data provided by volunteers can often be improved by increasing or modifying the training given to volunteers (Crall et al. 2011; Fore et al. 2001; Foster-Smith and Evans 2003; Ratnieks et al. 2016). If training is either too costly or too difficult, a variety of other methods can be used to increase the quality of data. Given that new volunteers often improve over time even without training, one can simply discard results from new or inexperienced volunteers to improve the overall quality of data. If available data are so scarce that results cannot be discarded, or if discarding the data would be unethical for some reason, then unlikely results can be flagged for investigation by an expert (Bonter and Cooper 2012). Further, even when the reliability of particular volunteers is low, the data can often be aggregated in ways that result in expert quality data (Kosmala et al. 2016). In the absence of particular data validation or quality assurance procedures, if enough data can be generated by volunteers – which is one of the primary selling points of citizen science – the absolute volume of data sometimes can limit the influence of low quality contributions by volunteers (Bonter and Cooper 2012). In some cases, this alone may be sufficient to allow confidence in citizen science results.

It would be a mistake, however, to focus solely on the extent to which citizen-generated data are compatible with professional scientific practices. To do so would be to ignore important distinctions between different kinds of citizen science. For example, Ottinger (2017b) compares scientific authority-driven citizen science, which strives to emulate the protocols and practices of professional scientists, with social movement-based citizen science, which is sometimes designed to challenge traditional scientific approaches (see also Brown 1992; Cooper and Lewenstein 2016; Irwin 1995; Riesch and Potter 2014). Social movement-based citizen science is frequently driven by communities that are striving to address public-health concerns related to environmental pollution. In some cases, these communities argue that professional scientists demand too much evidence before they are willing to draw linkages between environmental pollution and health effects (Elliott 2014; Ottinger 2017b). They also sometimes challenge assumptions involved in expert risk assessments (Elliott 2017; Wynne 1989). Moreover, in some cases they are more concerned about using the data they collect for the purposes of creating effective narratives to facilitate political activism rather than using the data for more traditional forms of scientific analysis (Ottinger 2017a). As discussed in the following section, this does not imply that when citizens engage in political advocacy the quality of their scientific work is threatened; rather, it illustrates that different kinds of data are needed for different sorts of projects and analyses.

When evaluating the legitimacy of these activities, it is once again helpful to appeal to the philosophical literature that emphasizes the importance of evaluating whether particular scientific activities serve the aims of inquiry in specific contexts. For example, if citizens are trying to identify potential health threats for the purposes of motivating further scientific, legal, or political action, it might be entirely appropriate for them to demand lower standards of evidence than scientists would typically require before drawing causal inferences in scholarly publications (Barrett and Raffensperger 1999; Cranor 1993; Douglas 2009; Elliott and Richards 2017). Similarly, if they are striving to challenge the assumptions involved in expert risk assessments, performing alternative assessments in an effort to determine whether they yield different results might make sense. For example, Ottinger (2010) has reported on efforts by citizen groups in Louisiana to collect information about air quality near industrial facilities using low-cost sampling equipment colloquially known as “buckets.” Regulators have pointed out that the buckets provide very short-term measurements of air quality, whereas Louisiana air-quality standards focus on long-term (24-hour or annual) average exposures to air pollutants. However, this departure from standard approaches makes sense, given that one of the goals of the citizens is to challenge current regulatory policy by exploring connections between their health problems and short-term spikes in air pollution that are not addressed within the current regulatory framework (Ottinger 2010).

It is noteworthy that the citizen science community is already sensitive to the fact that quality standards for scientific data may appropriately vary depending on the nature of the question being asked. Rather than trying to address general questions about the quality of citizen science data, practitioners argue that it is more fruitful to understand and define quality in terms of data’s “fitness for intended use” or “fitness for intended purpose” (Wiggins et al. 2011; Wiggins and He 2016). For example, in response to a survey of professional scientists who had been involved in citizen science projects, one respondent noted, “there is no such thing as quality of data, it’s what you use the data for. Different uses will require different quality” (Riesch and Potter 2014, p. 112). Fitness for use has even been recommended as a guiding principle for U.S. federal agencies that incorporate citizen science. John Holdren, former director of the U.S. Office of Science and Technology Policy, warned that quality assurance for citizen science cannot be thought of as “one-size-fits-all”; rather, oversight and quality assurance must ensure that “data have the appropriate level of quality for the purposes of a particular project” (Holdren 2015).

More recently, the citizen science community also has recognized that the concept of fitness for use requires further elaboration and specification. Clarifying this concept would be an important step in standardizing citizen science projects, as well as providing practitioners with the vocabulary necessary to articulate the particular data quality needs of their project. It would also facilitate further collaboration and data sharing between projects and help generate a framework for discussing questions about data quality (Bowser 2017; Bowser et al. 2017).

Thus, the philosophical literature on the aims of science, the empirical literature on data quality in citizen science, and the citizen science community’s own reflections on how to define data quality all converge on the conclusion that it is unfruitful to pose concerns about the quality of data and methodology as a universal critique of citizen science. Rather than voicing general concerns about the quality of citizen-generated data, we should direct our attention towards particular projects aimed at particular results. In those specific contexts, we can ask whether the data provided by particular citizen groups and the methods that they have employed are sufficient for addressing the epistemic or practical task at hand. We can also explore whether there are ways of altering the questions being asked or improving the data being collected so that they are compatible. In some cases, citizen-science projects may even challenge professional scientists to reconsider their assumptions about what kinds of data or methods are most helpful and what kinds of questions ought to be investigated.

Concern #3: Inappropriate Advocacy

Participants at the July 2017 workshop also raised the persistent concern that their work was challenged for being too focused on advocacy. This concern is particularly relevant in the context of the “social movement-based” citizen science projects discussed in the last section (Ottinger 2017b). Worries about inappropriate advocacy can take a number of different forms, some of which have already been addressed in the preceding sections of this paper. For example, one might worry that social movement-based projects tend to be focused on monitoring potential hazards rather than on testing scientific hypotheses. One might also worry that these projects tend to employ methods that are not sufficiently rigorous. A different worry, which we have not explored in the preceding sections, is that good science ought to be “value free,” in the sense that those conducting it should be motivated by the disinterested search for truth rather than by a social or political agenda (Nelson and Vucetich 2009). Those who promote “value free” science are likely to worry that citizens or academic researchers working on social movement-based research projects might not maintain an appropriate level of objectivity when analyzing results, and those collecting data for these projects could be inappropriately influenced by ideological conflicts of interest (Resnik et al. 2015).

In recent decades, philosophers of science have written a great deal about whether good science can or should be value free (Douglas 2009; Elliott and Richards 2017; Lacey 1999; Longino 2002). One important contribution of this literature has been to clarify a variety of different senses in which science can be value-laden. For example, it is now widely recognized that scientists unavoidably appeal to epistemic (i.e., truth-oriented) values in their work, including values such as explanatory power, empirical accuracy, scope, and consistency (Douglas 2013). Moreover, ethical values play important roles in guiding the kinds of inquiries that are allowed to proceed (Shamoo and Resnik 2015). For example, prohibitions against subjecting human or animal research subjects to undue suffering are based on ethical values. In addition, social and ethical values often have legitimate roles to play in determining what areas of research deserve the most funding and support (Kitcher 2001). Thus, it seems clear that values have many important roles to play in scientific research.

Nevertheless, one could acknowledge all these legitimate roles for values in science and still worry that the kind of advocacy associated with many citizen science projects goes too far. For example, one might worry that both professional scientists and citizens involved in these projects could be so concerned about promoting environmental health that they are no longer objective when designing, interpreting, or communicating their work. This is an important concern, but recent work in the philosophy of science suggests that scientists can be influenced by values like the concern for environmental and public health without inappropriately jeopardizing scientific objectivity (Elliott and Resnik 2014).

Although philosophers do not speak with a unified voice on this issue, most argue that scientists can legitimately be influenced by values under at least some circumstances (e.g., Douglas 2009; Elliott 2017; Lacey 1999). Indeed, it is psychologically unrealistic to think that scientists who perform research on policy-relevant topics could prevent their personal, social, and political values from influencing their work in a variety of subtle ways (Elliott and Resnik 2014). Some philosophers argue that the key to maintaining objectivity is for scientists to be as transparent as possible about the roles that values might have played in their reasoning (Elliott 2017). Others argue that scientists should allow values to influence some aspects of their reasoning (such as the standards of evidence they demand for drawing conclusions) while doing their best to prevent values from playing other roles in their reasoning (e.g., supplanting empirical evidence; Douglas 2009). Still others argue that the secret to maintaining scientific objectivity is to focus on the scientific community rather than on individual scientists (Longino 1990; Longino 2002). On this view, we should be striving to promote scientific communities with diverse perspectives and with procedures in place to uncover and critically evaluate value influences.

This philosophical literature suggests that it is not inherently problematic for those engaged in citizen science to bring a perspective of advocacy to their work. What is more important is that the values they bring are made sufficiently transparent and scrutinized appropriately. Consider the case of the Louisiana Bucket Brigade and their use of special “buckets” to collect air samples near polluting facilities (Ottinger 2010). As mentioned in the previous section, these buckets meet some scientific and regulatory standards, but they depart from typical standards in other respects; for example, they collect data over short periods of time rather than extended time periods. This created controversies among community members, regulators, and scientists about the propriety of departing from typical approaches for measuring air pollution (Ottinger 2010). From a philosophical perspective, what is important in a case like this one is to be explicit about employing methods or assumptions that depart from standard scientific norms (DeWinter 2016). By doing so, all parties to the dispute can decide whether they feel comfortable with the findings.

Admittedly, it is unrealistic to think that those engaged in citizen science can be fully aware of (and transparent about) all the subtle ways in which their values might have influenced their work. But concerns about the failure to clarify important methodological and interpretive choices are not unique to citizen science. A growing body of empirical evidence suggests that some areas of industry-funded research are biased in favor of the funders (Doucet and Sismondo 2008; Sismondo 2008). There are also growing concerns that a great deal of academic research is not reproducible (Begley and Ioannidis 2015; Open Science Collaboration 2015; “Reality Check on Reproducibility” 2016). Of course, one might worry that citizen science can be prone to particularly severe conflicts of interest and that citizens may be less likely than professional scientists to have training or institutional constraints that could mitigate these conflicts. But the seriousness of these concerns varies dramatically from case to case, depending on the kinds of conflicts of interest that are present and the steps taken to alleviate those conflicts. In citizen science as well as industrial and academic science, procedural steps can be put in place to help bring implicit methodological choices and value judgments to light. These steps include systematically publishing scientific results, providing open access to data and methods, using quality-control procedures, following reporting guidelines, and creating venues for critically evaluating reported findings (Elliott 2018; Longino 2002; Moermond et al. 2016; Molloy 2011; Royal Society 2012; Stodden et al. 2016).

Furthermore, citizen science projects can potentially play an important role in bringing implicit methodological choices and value judgments in scientific research to light, thereby increasing objectivity and reproducibility. Philosophers of science have pointed out that when values are shared across a research community they can easily go unrecognized; thus, one of the best ways to draw attention to implicit values and subject them to critical scrutiny is to incorporate investigators or reviewers with diverse perspectives (Harding 2015; Longino 2002). Non-professional scientists who participate in research projects are often able to bring unique perspectives because of their social situation or life experiences. For example, citizens affected by environmental pollution might value precaution to a greater extent than many professional scientists and therefore challenge the standards of proof typically required for drawing conclusions about public-health hazards (Elliott 2017; Ottinger 2017b). Citizens might also recognize important knowledge gaps or questions that have received inadequate attention from professional scientists (Wynne 1989). For example, two scientists involved in a citizen group called the Madison Environmental Justice Organization (MEJO) in Madison, Wisconsin, have reported how citizens helped to challenge a variety of questionable assumptions held by professional risk assessors, including assumptions about the level of pollution in nearby lakes, the number of fish being consumed by subsistence anglers, and the parts of the fish being consumed (Powell and Powell 2011). Thus, not only is it possible for those engaged in citizen science to take a perspective of advocacy without destroying scientific objectivity, but sometimes their value-laden perspective can actually increase scientific objectivity by uncovering values or assumptions in traditional scientific work that would otherwise have escaped critical scrutiny.

Conclusion

As citizen science has become more widespread, questions have arisen about its bona fides as a legitimate form of scientific inquiry. These questions are ethically significant, because wasting scarce resources on low-quality research is problematic, as is dismissing valuable research that can contribute to better social decision making. We have shown that the philosophy of science provides helpful resources for addressing three concerns about research quality: (1) that citizen science is not sufficiently hypothesis-driven; (2) that citizen science does not generate sufficiently high-quality data or use sufficiently rigorous methods; and (3) that citizen science is tainted by advocacy and therefore is not sufficiently disinterested. The message of this paper is that these broad concerns are something of a red herring. From a philosophical perspective, one cannot evaluate the overall quality of citizen science—or any form of science—without considering the details of specific research contexts. In many cases, citizen science provides one of the best avenues for achieving scientific goals and moving scientific research forward.

Notes

1Definitions for hypotheses vary from one discipline to another (Donovan et al. 2015), but we define hypothesis-driven science as a linear process of making observations, proposing a hypothesis, and performing an experiment to test the hypothesis (Elliott et al. 2016). 

Acknowledgements

We would especially like to thank Lisa Rasmussen for her guidance and advice on this paper, as well as her invitation to participate in the 2017 “Filling the ‘Ethics Gap’ in Citizen Science” workshop. We are grateful to all workshop participants for their insights and concerns, which helped inspire this paper. We also thank Anne Bowser, Gwen Ottinger, and two anonymous reviewers for valuable input that strengthened the paper.

Competing Interests

The authors have no competing interests to declare.

References

  1. Barrett, K and Raffensperger, C. 1999. Precautionary science. In: Raffensperger, C and Tickner, J (eds.), Protecting Public Health and the Environment, 106–122. Washington, DC: Island Press. 

  2. Begley, CG and Ioannidis, JP. 2015. Reproducibility in science: Improving the standard for basic and preclinical research. Circulation Research, 116: 116–126. DOI: https://doi.org/10.1161/CIRCRESAHA.114.303819 

  3. Bhakthavatsalam, S and Cartwright, N. 2017. What’s so special about empirical adequacy? European Journal for Philosophy of Science, 7: 445–465. DOI: https://doi.org/10.1007/s13194-017-0171-7 

  4. Bonter, DN and Cooper, CB. 2012. Data validation in citizen science: A case study from Project FeederWatch. Frontiers in Ecology and the Environment, 10(6): 305–307. DOI: https://doi.org/10.1890/110273 

  5. Bowser, A. 2017. Standardizing Citizen Science? Proceedings of TDWG, 1: e21123. DOI: https://doi.org/10.3897/tdwgproceedings.1.21123 

  6. Bowser, A, Brenton, P, Stevenson, R, Newman, G, Schade, S, Bastin, L, Parker, A and Oliver, J. 2017. Citizen Science Association Data & Metadata Working Group: Report from CSA 2017 and Future Outlook. https://www.wilsoncenter.org/article/citizen-science-association-data-metadata-working-group-report-csa-2017-and-future-outlook. 

  7. Boyd, D and Crawford, K. 2012. Critical questions for big data: Provocations for a cultural, technological, and scholarly phenomenon. Information, Communication & Society, 15: 662–679. DOI: https://doi.org/10.1080/1369118X.2012.678878 

  8. Brigandt, I. 2015. Social values influence the adequacy conditions of scientific theories: Beyond inductive risk. Canadian Journal of Philosophy, 45(3): 326–356. DOI: https://doi.org/10.1080/00455091.2015.1079004 

  9. Brown, P. 1992. Popular Epidemiology and Toxic Waste Contamination: Lay and Professional Ways of Knowing. Journal of Health and Social Behavior, 33(3): 267–281. DOI: https://doi.org/10.2307/2137356 

  10. Chang, H. 2004. Inventing Temperature: Measurement and Scientific Progress. New York: Oxford University Press. DOI: https://doi.org/10.1093/0195171276.001.0001 

  11. Cohn, JP. 2008. Citizen science: Can volunteers do real research? AIBS Bulletin, 58(3): 192–197. DOI: https://doi.org/10.1641/B580303 

  12. Cooper, CB and Lewenstein, BV. 2016. Two meanings of Citizen Science. In: Cavalier, D and Kennedy, E (eds.), The Rightful Place of Science: Citizen Science, 51–62. Tempe, AZ: Arizona State University Press. 

  13. Corburn, J. 2005. Street Science: Community Knowledge and Environmental Health Justice. Cambridge, MA: MIT Press. 

  14. Crall, AW, Newman, GJ, Stohlgren, TJ, Holfelder, KA, Graham, J and Waller, DM. 2011. Assessing citizen science data quality: An invasive species case study. Conservation Letters, 4(6): 433–442. DOI: https://doi.org/10.1111/j.1755-263X.2011.00196.x 

  15. Cranor, C. 1993. Regulating Toxic Substances: A Philosophy of Science and the Law. New York: Oxford University Press. DOI: https://doi.org/10.1093/acprof:oso/9780195074369.001.0001 

  16. Cranor, C. 1995. The social benefits of expedited risk assessments. Risk Analysis, 15: 353–358. DOI: https://doi.org/10.1111/j.1539-6924.1995.tb00328.x 

  17. Detlefsen, K. 2014. Émilie du Châtelet. Stanford Encyclopedia of Philosophy. Available at: https://plato.stanford.edu/entries/emilie-du-chatelet/. Accessed March 27, 2018. 

  18. DeWinter, J. 2016. Interests and Epistemic Integrity in Science. Lanham, MD: Lexington Books. 

  19. Done, T, Roelfsema, C, Harvey, A, Schuller, L, Hil, J, Schläppy, M, Lea, A, Bauer-Civiello, A and Loder, J. 2017. Reliability and utility of citizen science reef monitoring data collected by Reef Check Australia, 2002–2015. Marine Pollution Bulletin, 117(1–2): 148–155. DOI: https://doi.org/10.1016/j.marpolbul.2017.01.054 

  20. Donovan, S, O’Rourke, M and Looney, C. 2015. Your hypothesis or mine? Terminological and conceptual variation across disciplines. SAGE Open, 1–13. April–June. DOI: https://doi.org/10.1177/2158244015586237 

  21. Doucet, M and Sismondo, S. 2008. Evaluating Solutions to Sponsorship Bias. Journal of Medical Ethics, 34: 627–630. DOI: https://doi.org/10.1136/jme.2007.022467 

  22. Douglas, H. 2009. Science, Policy, and the Value-Free Ideal. Pittsburgh: University of Pittsburgh Press. DOI: https://doi.org/10.2307/j.ctt6wrc78 

  23. Douglas, H. 2013. The value of cognitive values. Philosophy of Science, 80(5): 796–806. DOI: https://doi.org/10.1086/673716 

  24. Elliott, KC. 2012. Epistemic and methodological iteration in scientific research. Studies in History and Philosophy of Science, 43: 376–382. DOI: https://doi.org/10.1016/j.shpsa.2011.12.034 

  25. Elliott, KC. 2014. Ethical and Societal Values in Nanotoxicology. In: Gordijn, B and Cutter, AM (eds.), In Pursuit of Nanoethics, 147–163. Dordrecht: Springer. DOI: https://doi.org/10.1007/978-1-4020-6817-1_10 

  26. Elliott, KC. 2017. A Tapestry of Values: An Introduction to Values in Science. New York: Oxford University Press. DOI: https://doi.org/10.1093/acprof:oso/9780190260804.001.0001 

  27. Elliott, KC. 2018. Addressing industry-funded research with criteria for objectivity. Philosophy of Science, 85 (Proceedings): Forthcoming. 

  28. Elliott, KC, Cheruvelil, KS, Montgomery, GM and Soranno, PA. 2016. Conceptions of Good Science in Our Data-Rich World. BioScience, 66: 880–889. DOI: https://doi.org/10.1093/biosci/biw115 

  29. Elliott, KC and McKaughan, DJ. 2014. Nonepistemic values and the multiple goals of science. Philosophy of Science, 81(1): 1–21. DOI: https://doi.org/10.1086/674345 

  30. Elliott, KC and Resnik, DB. 2014. Science, Policy, and the Transparency of Values. Environmental Health Perspectives, 122: 647–650. DOI: https://doi.org/10.1289/ehp.1408107 

  31. Elliott, KC and Richards, T. (eds.) 2017. Exploring Inductive Risk: Case Studies of Values in Science. New York: Oxford University Press. DOI: https://doi.org/10.1093/acprof:oso/9780190467715.001.0001 

  32. Fore, LS, Paulsen, K and O’Laughlin, K. 2001. Assessing the performance of volunteers in monitoring streams. Freshwater Biology, 46(1): 109–123. DOI: https://doi.org/10.1111/j.1365-2427.2001.00640.x 

  33. Forrester, G, Baily, P, Conetta, D, Forrester, L, Kintzing, E and Jarecki, L. 2015. Comparing monitoring data collected by volunteers and professionals shows that citizen scientists can detect long-term change on coral reefs. Journal for Nature Conservation, 24: 1–9. DOI: https://doi.org/10.1016/j.jnc.2015.01.002 

  34. Foster-Smith, J and Evans, SM. 2003. The value of marine ecological data collected by volunteers. Biological Conservation, 113(2): 199–213. DOI: https://doi.org/10.1016/S0006-3207(02)00373-7 

  35. Glass, DJ and Hall, N. 2008. A brief history of the hypothesis. Cell, 134: 378–381. DOI: https://doi.org/10.1016/j.cell.2008.07.033 

  36. Harding, S. 2015. Objectivity and Diversity: Another Logic of Scientific Research. Chicago: University of Chicago Press. DOI: https://doi.org/10.7208/chicago/9780226241531.001.0001 

  37. Heffernan, JB, Soranno, PA, Angilletta, MJ, Buckley, LB, Gruner, DS, Keitt, TH, Kellner, JR, Kominoski, JS, Rocha, AV, Xiao, J, Harms, T, Goring, SJ, Koenig, LE, McDowell, WH, Powell, H, Richardson, AD, Stow, CA, Vargas, R and Weathers, KC. 2014. Macrosystems ecology: Understanding ecological patterns and processes at continental scales. Frontiers in Ecology and the Environment, 12(1): 5–14. DOI: https://doi.org/10.1890/130017 

  38. Holdren, J. 2015. Memorandum to the heads of executive departments and agencies. Executive Office of the President, Office of Science and Technology Policy. September 30. Available at: https://obamawhitehouse.archives.gov/sites/default/files/microsites/ostp/holdren_citizen_science_memo_092915_0.pdf. Accessed February 6, 2018. 

  39. Intemann, K. 2015. Distinguishing between legitimate and illegitimate values in climate modeling. European Journal for Philosophy of Science, 5(2): 217–232. DOI: https://doi.org/10.1007/s13194-014-0105-6 

  40. Irwin, A. 1995. Citizen Science: A Study of People, Expertise, and Sustainable Development. London: Routledge. 

  41. Johnson, K. 2007. Natural history as stamp collecting: A brief history. Archives of Natural History, 34: 244–258. DOI: https://doi.org/10.3366/anh.2007.34.2.244 

  42. Kallimanis, AS, Panitsa, M and Dimopoulos, P. 2017. Quality of non-expert citizen science data collected for habitat type conservation status assessment in Natura 2000 protected areas. Scientific Reports, 1–10. September 2016. DOI: https://doi.org/10.1038/s41598-017-09316-9 

  43. Kell, D and Oliver, S. 2004. Here is the evidence, now what is the hypothesis? The complementary roles of inductive and hypothesis-driven science in the post-genomic era. Bioessays, 26: 99–105. DOI: https://doi.org/10.1002/bies.10385 

  44. Kitcher, P. 2001. Science, Truth, and Democracy. New York: Oxford University Press. DOI: https://doi.org/10.1093/0195145836.001.0001 

  45. Kosmala, M, Wiggins, A, Swanson, A and Simmons, B. 2016. Assessing data quality in citizen science. Frontiers in Ecology and the Environment, 14(10): 551–560. DOI: https://doi.org/10.1002/fee.1436 

  46. Lacey, H. 1999. Is Science Value Free? London: Routledge. 

  47. Laudan, L. 1981. Science and Hypothesis: Historical essays on scientific methodology. Dordrecht: D. Reidel. DOI: https://doi.org/10.1007/978-94-015-7288-0 

  48. Longino, H. 1990. Science as Social Knowledge. Princeton: Princeton University Press. 

  49. Longino, H. 2002. The Fate of Knowledge. Princeton: Princeton University Press. 

  50. Moermond, C, Kase, R, Korkaric, M and Agerstrand, M. 2016. CRED: Criteria for Reporting and Evaluating Ecotoxicity Data. Environmental Toxicology and Chemistry, 35: 1297–1309. DOI: https://doi.org/10.1002/etc.3259 

  51. Molloy, JC. 2011. The Open Knowledge Foundation: Open Data Means Better Science. PLOS Biology, 9: e1001195. DOI: https://doi.org/10.1371/journal.pbio.1001195 

  52. Nelson, M and Vucetich, J. 2009. On Advocacy by Environmental Scientists: What, Whether, Why, and How. Conservation Biology, 23: 1090–1101. DOI: https://doi.org/10.1111/j.1523-1739.2009.01250.x 

  53. O’Fallon, L and Finn, S. 2015. Citizen Science and Community-Engaged Research in Public Health. Lab Matters, Fall: 5. 

  54. O’Malley, M, Elliott, KC and Burian, R. 2010. From genetic to genomic regulation: Iterative methods in miRNA research. Studies in the History and Philosophy of the Biological and Biomedical Sciences, 41: 407–417. DOI: https://doi.org/10.1016/j.shpsc.2010.10.011 

  55. O’Malley, M, Elliott, KC, Haufe, C and Burian, R. 2009. Philosophies of funding. Cell, 138: 611–615. DOI: https://doi.org/10.1016/j.cell.2009.08.008 

  56. Open Science Collaboration. 2015. Estimating the Reproducibility of Psychological Science. Science, 349: 943. DOI: https://doi.org/10.1126/science.aac4716 

  57. Ottinger, G. 2010. Buckets of Resistance: Standards and the Effectiveness of Citizen Science. Science, Technology & Human Values, 35: 244–270. DOI: https://doi.org/10.1177/0162243909337121 

  58. Ottinger, G. 2017a. Making Sense of Citizen Science: Stories as a Hermeneutic Resource. Energy Research and Social Science, 31: 41–49. DOI: https://doi.org/10.1016/j.erss.2017.06.014 

  59. Ottinger, G. 2017b. Reconstructing or Reproducing? Scientific Authority and Models of Change in Two Traditions of Citizen Science. In: Tyfield, D, Lave, R, Randalls, S and Thorpe, C (eds.), The Routledge Handbook of the Political Economy of Science, 351–364. London: Routledge. 

  60. Poisson, A, McCullough, I, Cheruvelil, KS, Elliott, KC, Latimore, J and Soranno, P. In preparation. Quantifying the Contribution of Citizen Science to Broad-Scale Ecological Databases. 

  61. Popper, K. 1963. Conjectures and Refutations. London: Routledge and Kegan Paul. 

  62. Powell, M and Powell, J. 2011. Invisible People, Invisible Risks: How Scientific Assessments of Environmental Health Risks Overlook Minorities—and How Community Participation Can Make Them Visible. In: Ottinger, G and Cohen, B (eds.), Technoscience and Environmental Justice: Expert Cultures in a Grassroots Movement, 149–178. Cambridge, MA: MIT Press. DOI: https://doi.org/10.7551/mitpress/9780262015790.003.0007 

  63. Ratnieks, FLW, Schrell, F, Shappard, RC, Brown, E, Bristow, OW and Garbuzov, M. 2016. Data reliability in citizen science: Learning curve and the effects of training method, volunteer background and experience on identification accuracy of insects visiting ivy flowers. Methods in Ecology and Evolution, 7(10): 1226–1235. DOI: https://doi.org/10.1111/2041-210X.12581 

  64. Reality Check on Reproducibility. 2016. Nature, 533: 437. DOI: https://doi.org/10.1038/533437a 

  65. Resnik, DB, Elliott, KC and Miller, A. 2015. A Framework for Addressing Ethical Issues in Citizen Science. Environmental Science and Policy, 54: 475–481. DOI: https://doi.org/10.1016/j.envsci.2015.05.008 

  66. Riesch, H and Potter, C. 2014. Citizen Science as Seen by Scientists: Methodological, Epistemological, and Ethical Dimensions. Public Understanding of Science, 23: 107–120. DOI: https://doi.org/10.1177/0963662513497324 

  67. Roman, LA, Scharenbroch, BC, Östberg, JPA, Mueller, LS, Henning, JG, Koeser, AK, Sanders, JR, Betz, DR and Jordan, RC. 2017. Data quality in citizen science urban tree inventories. Urban Forestry and Urban Greening, 22: 124–135. DOI: https://doi.org/10.1016/j.ufug.2017.02.001 

  68. Roy, HE, Pocock, MJ, Preston, CD, Roy, DB, Savage, J, Tweddle, JC and Robinson, LD. 2012. Understanding citizen science and environmental monitoring: Final report on behalf of UK Environmental Observation Framework. Wallingford, UK: NERC/Centre for Ecology & Hydrology. 

  69. Royal Society. 2012. Science as an Open Enterprise. London: Royal Society. 

  70. See, L, Comber, A, Salk, C, Fritz, S, van der Velde, M, Perger, C, Schill, C, McCallum, I, Kraxner, F and Obersteiner, M. 2013. Comparing the Quality of Crowdsourced Data Contributed by Expert and Non-Experts. PLoS ONE, 8(7): 1–11. DOI: https://doi.org/10.1371/journal.pone.0069958 

  71. Shamoo, A and Resnik, DB. 2015. Responsible Conduct of Research. New York: Oxford University Press. 

  72. Silvertown, J. 2009. A new dawn for citizen science. Trends in ecology & evolution, 24(9): 467–471. DOI: https://doi.org/10.1016/j.tree.2009.03.017 

  73. Sismondo, S. 2008. Pharmaceutical Company Funding and Its Consequences: A Qualitative Systematic Review. Contemporary Clinical Trials, 29: 109–113. DOI: https://doi.org/10.1016/j.cct.2007.08.001 

  74. Soranno, PA, Cheruvelil, KS, Bissell, EG, Bremigan, MT, Downing, JA, Fergus, CE and Webster, KE. 2014. Cross-scale interactions: Quantifying multi-scaled cause–effect relationships in macrosystems. Frontiers in Ecology and the Environment, 12(1): 65–73. DOI: https://doi.org/10.1890/120366 

  75. Stodden, V, McNutt, M, Bailey, D, Deelman, E, Gil, Y, Hanson, B, Heroux, M, Ioannidis, JP and Taufer, M. 2016. Enhancing reproducibility for computational methods. Science, 354: 1240–1241. DOI: https://doi.org/10.1126/science.aah6168 

  76. Swanson, A, Kosmala, M, Lintott, C and Packer, C. 2016. A generalized approach for producing, quantifying, and validating citizen science data from wildlife images. Conservation Biology, 30(3): 520–531. DOI: https://doi.org/10.1111/cobi.12695 

  77. van der Velde, T, Milton, DA, Lawton, TJ, Wilcox, C, Lansdell, M, Davis, G, Perkins, G and Hardesty, BD. 2017. Comparison of marine debris data collected by researchers and citizen scientists: Is citizen science data worth the effort? Biological Conservation, 208: 127–138. DOI: https://doi.org/10.1016/j.biocon.2016.05.025 

  78. Wiggins, A and He, Y. 2016. Community-based Data Validation Practices in Citizen Science. Proceedings of the 19th ACM Conference on computer-supported cooperative work & social computing, 1548–1559. DOI: https://doi.org/10.1145/2818048.2820063 

  79. Wiggins, A, Newman, G, Stevenson, RD and Crowston, K. 2011. Mechanisms for Data Quality and Validation in Citizen Science. e-Science Workshops (eScienceW), 2011 IEEE Seventh International Conference on, 14–19. DOI: https://doi.org/10.1109/eScienceW.2011.27 

  80. Wing, S, Horton, RA, Muhammad, N, Grant, GR, Tajik, M and Thu, K. 2008. Integrating Epidemiology, Education, and Organizing for Environmental Justice: Community Health Effects of Industrial Hog Operations. American Journal of Public Health, 98(8): 1390–1397. DOI: https://doi.org/10.2105/AJPH.2007.110486 

  81. Wynne, B. 1989. Sheepfarming after Chernobyl: A case study in communicating scientific information. Environment: Science and Policy for Sustainable Development, 31(2): 10–39. DOI: https://doi.org/10.1080/00139157.1989.9928930 

comments powered by Disqus